학술논문

Multi-Agent Deep Reinforcement Learning for Coordinated Multipoint in Mobile Networks
Document Type
Periodical
Source
IEEE Transactions on Network and Service Management IEEE Trans. Netw. Serv. Manage. Network and Service Management, IEEE Transactions on. 21(1):908-924 Feb, 2024
Subject
Communication, Networking and Broadcast Technologies
Computing and Processing
Resource management
Training
Quality of experience
Power control
Adaptation models
Transfer learning
Surveys
Mobility management
coordinated multipoint
CoMP
cell selection
resource management
reinforcement learning
multi agent
self-learning
self-adaptation
QoE
Language
ISSN
1932-4537
2373-7379
Abstract
Macrodiversity is a key technique to increase the capacity of mobile networks. It can be realized using coordinated multipoint (CoMP), simultaneously connecting users to multiple overlapping cells. Selecting which users to serve by how many and which cells is NP-hard but needs to happen continuously in real time as users move and channel state changes. Existing approaches often require strict assumptions about or perfect knowledge of the underlying radio system, its resource allocation scheme, or user movements, none of which is readily available in practice. Instead, we propose three novel self-learning and self-adapting approaches using model-free deep reinforcement learning (DRL): DeepCoMP, DD-CoMP, and D3-CoMP. DeepCoMP leverages central control and observations of all users to select cells almost optimally. DD-CoMP and D3-CoMP use multi-agent DRL, which allows distributed, robust, and highly scalable coordination. All three approaches learn from experience and self-adapt to varying scenarios, reaching 2x higher Quality of Experience than other approaches. They have very few built-in assumptions and do not need prior system knowledge, making them more robust to change and better applicable in practice than existing approaches.