Deep Reinforcement Learning for Multi-Agent Power Control in Heterogeneous Networks

被引:38
作者
Zhang, Lin [1 ]
Liang, Ying-Chang [2 ]
机构
[1] Univ Elect Sci & Technol China, Key Lab Commun, Chengdu 611731, Peoples R China
[2] Univ Elect Sci & Technol China, Ctr Intelligent Networking & Commun, Chengdu 611731, Peoples R China
基金
中国国家自然科学基金;
关键词
Power control; Wireless communication; Resource management; Interference; Heuristic algorithms; Rayleigh channels; Reinforcement learning; DRL; multi-agent; power control; MASC; HetNet; RESOURCE-ALLOCATION; FEEDBACK; ACCESS;
D O I
10.1109/TWC.2020.3043009
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
We consider a typical heterogeneous network (HetNet), in which multiple access points (APs) are deployed to serve users by reusing the same spectrum band. Since different APs and users may cause severe interference to each other, advanced power control techniques are needed to manage the interference and enhance the sum-rate of the whole network. Conventional power control techniques first collect instantaneous global channel state information (CSI) and then calculate sub-optimal solutions. Nevertheless, it is challenging to collect instantaneous global CSI in the HetNet, in which global CSI typically changes fast. In this article, we exploit deep reinforcement learning (DRL) to design a multi-agent power control algorithm, which has a centralized-training-distributed-execution framework. To be specific, each AP acts as an agent with a local deep neural network (DNN) and we propose a multiple-actor-shared-critic (MASC) method to train the local DNNs separately in an online trial-and-error manner. With the proposed algorithm, each AP can independently use the local DNN to control the transmit power with only local observations. Simulations results show that the proposed algorithm outperforms the conventional power control algorithms in terms of both the converged average sum-rate and the computational complexity.
引用
收藏
页码:2551 / 2564
页数:14
相关论文
共 40 条
[11]   SON Coordination in Heterogeneous Networks: A Reinforcement Learning Framework [J].
Iacoboaiea, Ovidiu-Constantin ;
Sayrac, Berna ;
Ben Jemaa, Sana ;
Bianchi, Pascal .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2016, 15 (09) :5835-5847
[12]   Does Frequent Low Resolution Feedback Outperform Infrequent High Resolution Feedback for Multiple Antenna Beamforming Systems? [J].
Kim, Taejoon ;
Love, David J. ;
Clerckx, Bruno .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2011, 59 (04) :1654-1669
[13]   Resource Allocation for Spectrum Underlay in Cognitive Radio Networks [J].
Le, Long Bao ;
Hossain, Ekram .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2008, 7 (12) :5306-5315
[14]  
Lillicrap T. P., 2016, P INT C LEARN REPR S, P1
[15]   Dynamic Spectrum Management: Complexity and Duality [J].
Luo, Zhi-Quan ;
Zhang, Shuzhong .
IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2008, 2 (01) :57-73
[16]   Applications of Deep Reinforcement Learning in Communications and Networking: A Survey [J].
Luong, Nguyen Cong ;
Hoang, Dinh Thai ;
Gong, Shimin ;
Niyato, Dusit ;
Wang, Ping ;
Liang, Ying-Chang ;
Kim, Dong In .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2019, 21 (04) :3133-3174
[17]   Multi-Agent Deep Learning for Multi-Channel Access in Slotted Wireless Networks [J].
Mennes, Ruben ;
De Figueiredo, Felipe A. P. ;
Latre, Steven .
IEEE ACCESS, 2020, 8 (08) :95032-95045
[18]   Deep Reinforcement Learning for 5G Networks: Joint Beamforming, Power Control, and Interference Coordination [J].
Mismar, Faris B. ;
Evans, Brian L. ;
Alkhateeb, Ahmed .
IEEE TRANSACTIONS ON COMMUNICATIONS, 2020, 68 (03) :1581-1592
[19]   Human-level control through deep reinforcement learning [J].
Mnih, Volodymyr ;
Kavukcuoglu, Koray ;
Silver, David ;
Rusu, Andrei A. ;
Veness, Joel ;
Bellemare, Marc G. ;
Graves, Alex ;
Riedmiller, Martin ;
Fidjeland, Andreas K. ;
Ostrovski, Georg ;
Petersen, Stig ;
Beattie, Charles ;
Sadik, Amir ;
Antonoglou, Ioannis ;
King, Helen ;
Kumaran, Dharshan ;
Wierstra, Daan ;
Legg, Shane ;
Hassabis, Demis .
NATURE, 2015, 518 (7540) :529-533
[20]   Heuristically Accelerated Reinforcement Learning for Dynamic Secondary Spectrum Sharing [J].
Morozs, Nils ;
Clarke, Tim ;
Grace, David .
IEEE ACCESS, 2015, 3 :2771-2783