Double Deep Q-Network-Based Energy-Efficient Resource Allocation in Cloud Radio Access Network

被引:65
作者
Iqbal, Amjad [1 ]
Tham, Mau-Luen [1 ]
Chang, Yoong Choon [1 ]
机构
[1] Univ Tunku Abdul Rahman, Lee Kong Chian Fac Engn & Sci, Dept Elect & Elect Engn, Sungai Long Campus, Kajang 43000, Selangor, Malaysia
关键词
Energy efficiency; Resource management; Minimization; Reinforcement learning; Quality of service; Optimization; Heuristic algorithms; Cloud RAN; double deep Q-network (DDQN); energy efficiency (EE); Markov decision process (MDP); power allocation; USER ASSOCIATION; CELLULAR NETWORKS; DOWNLINK; OPTIMIZATION; MAXIMIZATION; MINIMIZATION;
D O I
10.1109/ACCESS.2021.3054909
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Cloud radio access network (CRAN) has been shown as an effective means to boost network performance. Such gain stems from the intelligent management of remote radio heads (RRHs) in terms of on/off operation mode and power consumption. Most conventional resource allocation (RA) methods, however, optimize the network utility without considering the switching overhead of RRHs in adjacent time intervals. When the network environment becomes time-correlated, mathematical optimization is not directly applicable. In this paper, we aim to optimize the energy efficiency (EE) subject to the constraints on per-RRH transmission power and user data rates. To this end, we formulate the EE problem as a Markov decision process (MDP) and subsequently adopt deep reinforcement learning (DRL) technique to reap the cumulative EE rewards. Our starting point is the deep Q network (DQN), which is a combination of deep learning and Q-learning. In each time slot, DQN configures the status of RRHs yielding the largest Q-value (known as state-action value) prior to solving a power minimization problem for active RRHs. To overcome the Q-value overestimation issue of DQN, we propose a Double DQN (DDQN) framework that obtains optimal reward better than DQN by separating the selected action from the target Q-value generator. Simulation results validate that the DDQN-based RA method is more energy-efficient than the DQN-based RA algorithm and a baseline solution.
引用
收藏
页码:20440 / 20449
页数:10
相关论文
共 39 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]   Joint User Association, Power Allocation, and Throughput Maximization in 5G H-CRAN Networks [J].
Ali, Mudassar ;
Rabbani, Quratulain ;
Naeem, Muhammad ;
Qaisar, Saad ;
Qamar, Farhan .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2017, 66 (10) :9254-9262
[3]  
Alimov S., 2017, IEEE T GREEN COMMUN, V1, P145
[4]  
[Anonymous], 2017, 2017 IEEE International Conference on Communications ICC
[5]  
[Anonymous], 2019, Ericsson mobility report June 2019
[6]   Energy Efficient Resource Allocation for Energy Harvesting Aided H-CRAN [J].
Chughtai, Naveed Ahmad ;
Ali, Mudassar ;
Qaisar, Saad ;
Imran, Muhammad ;
Naeem, Muhammad ;
Qamar, Farhan .
IEEE ACCESS, 2018, 6 :43990-44001
[7]   Energy Efficiency of Downlink Transmission Strategies for Cloud Radio Access Networks [J].
Dai, Binbin ;
Yu, Wei .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2016, 34 (04) :1037-1050
[8]   Distributed Robust Power Minimization for the Downlink of Multi-Cloud Radio Access Networks [J].
Dhif-Allah, Oussama ;
Dahrouj, Hayssam ;
Al-Naffouri, Tareq Y. ;
Alouini, Mohamed-Slim .
IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, 2018, 2 (02) :327-335
[9]  
Etoh Minoru, 2008, 2008 International Symposium on Applications and the Internet, P365, DOI 10.1109/SAINT.2008.84
[10]   Green Communication in Next Generation Cellular Networks: A Survey [J].
Gandotra, Pimmy ;
Jha, Rakesh Kumar ;
Jain, Sanjeev .
IEEE ACCESS, 2017, 5 :11727-11758