Deep Reinforcement Learning for User Association and Resource Allocation in Heterogeneous Cellular Networks

被引:326
作者
Zhao, Nan [1 ,2 ]
Liang, Ying-Chang [2 ]
Niyato, Dusit [3 ]
Pei, Yiyang [4 ]
Wu, Minghu [5 ]
Jiang, Yunhao [5 ]
机构
[1] Hubei Univ Technol, Hubei Collaborat Innovat Ctr High Efficiency Util, Wuhan 430068, Hubei, Peoples R China
[2] Univ Elect Sci & Technol China, CINC, Chengdu 611731, Sichuan, Peoples R China
[3] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore, Singapore
[4] Singapore Inst Technol, Infocomm Technol Cluster, Singapore, Singapore
[5] Hubei Univ Technol, Hubei Key Lab High Efficiency Utilizat Solar Ener, Wuhan 430068, Hubei, Peoples R China
基金
中国国家自然科学基金;
关键词
Heterogeneous cellular networks; user association; resource allocation; multi-agent deep reinforcement learning; ACCESS; MANAGEMENT; SELECTION; HETNETS;
D O I
10.1109/TWC.2019.2933417
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Heterogeneous cellular networks can offload the mobile traffic and reduce the deployment costs, which have been considered to be a promising technique in the next-generation wireless network. Due to the non-convex and combinatorial characteristics, it is challenging to obtain an optimal strategy for the joint user association and resource allocation issue. In this paper, a reinforcement learning (RL) approach is proposed to achieve the maximum long-term overall network utility while guaranteeing the quality of service requirements of user equipments (UEs) in the downlink of heterogeneous cellular networks. A distributed optimization method based on multi-agent RL is developed. Moreover, to solve the computationally expensive problem with the large action space, multi-agent deep RL method is proposed. Specifically, the state, action and reward function are defined for UEs, and dueling double deep Q-network (D3QN) strategy is introduced to obtain the nearly optimal policy. Through message passing, the distributed UEs can obtain the global state space with a small communication overhead. With the double-Q strategy and dueling architecture, D3QN can rapidly converge to a subgame perfect Nash equilibrium. Simulation results demonstrate that D3QN achieves the better performance than other RL approaches in solving large-scale learning problems.
引用
收藏
页码:5141 / 5152
页数:12
相关论文
共 46 条
  • [1] [Anonymous], 2015, ARXIV150906461
  • [2] [Anonymous], 2018, ARXIV180800490
  • [3] [Anonymous], 2017, J. Commun. Inf. Netw
  • [4] [Anonymous], 2004, An introduction to game theory
  • [5] An Autonomous Learning-Based Algorithm for Joint Channel and Power Level Selection by D2D Pairs in Heterogeneous Cellular Networks
    Asheralieva, Alia
    Miyanaga, Yoshikazu
    [J]. IEEE TRANSACTIONS ON COMMUNICATIONS, 2016, 64 (09) : 3996 - 4012
  • [6] Distributed User Association and Femtocell Allocation in Heterogeneous Wireless Networks
    Bayat, Siavash
    Louie, Raymond H. Y.
    Han, Zhu
    Vucetic, Branka
    Li, Yonghui
    [J]. IEEE TRANSACTIONS ON COMMUNICATIONS, 2014, 62 (08) : 3027 - 3043
  • [7] A comprehensive survey of multiagent reinforcement learning
    Busoniu, Lucian
    Babuska, Robert
    De Schutter, Bart
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2008, 38 (02): : 156 - 172
  • [8] Proactive Resource Management for LTE in Unlicensed Spectrum: A Deep Learning Perspective
    Challita, Ursula
    Dong, Li
    Saad, Walid
    [J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2018, 17 (07) : 4674 - 4689
  • [9] Markov Approximation for Combinatorial Network Optimization
    Chen, Minghua
    Liew, Soung Chang
    Shao, Ziyu
    Kai, Caihong
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2013, 59 (10) : 6301 - 6327
  • [10] Joint User Association and Resource Allocation in the Downlink of Heterogeneous Networks
    Chen, Youjia
    Li, Jun
    Chen, Wen
    Lin, Zihuai
    Vucetic, Branka
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2016, 65 (07) : 5701 - 5706