A Deep Reinforcement Learning-Based Power Control Scheme for the 5G Wireless Systems

被引:4
作者
Liang, Renjie [1 ]
Lyu, Haiyang [2 ]
Fan, Jiancun [1 ]
机构
[1] Xi An Jiao Tong Univ, Sch Elect & Informat Engn, Xian 710049, Shaanxi, Peoples R China
[2] Xi An Jiao Tong Univ, Sch Cyber Sci & Engn, Xian 710049, Shaanxi, Peoples R China
关键词
Interference; Signal to noise ratio; Wireless communication; Fluctuations; Fading channels; Training; 5G mobile communication; reinforcement learning; closed-loop power control (CLPC); signal-to-interference-plus-noise ratio (SINR); LTE;
D O I
10.23919/JCC.ea.2021-0523.202302
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
In the fifth generation (5G) wireless system, a closed-loop power control (CLPC) scheme based on deep Q learning network (DQN) is introduced to intelligently adjust the transmit power of the base station (BS), which can improve the user equipment (UE) received signal to interference plus noise ratio (SINR) to a target threshold range. However, the selected power control (PC) action in DQN is not accurately matched the fluctuations of the wireless environment. Since the experience replay characteristic of the conventional DQN scheme leads to a possibility of insufficient training in the target deep neural network (DNN). As a result, the Q-value of the sub-optimal PC action exceed the optimal one. To solve this problem, we propose the improved DQN scheme. In the proposed scheme, we add an additional DNN to the conventional DQN, and set a shorter training interval to speed up the training of the DNN in order to fully train it. Finally, the proposed scheme can ensure that the Q value of the optimal action remains maximum. After multiple episodes of training, the proposed scheme can generate more accurate PC actions to match the fluctuations of the wireless environment. As a result, the UE received SINR can achieve the target threshold range faster and keep more stable. The simulation results prove that the proposed scheme outperforms the conventional schemes.
引用
收藏
页码:109 / 119
页数:11
相关论文
共 25 条
[1]  
[Anonymous], 2010, TR, V36, P814
[2]  
[Anonymous], 2013, PROC INT C NEURAL IN
[3]  
[Anonymous], 2009, 36213 3GPP TS
[4]   Deep Reinforcement Learning A brief survey [J].
Arulkumaran, Kai ;
Deisenroth, Marc Peter ;
Brundage, Miles ;
Bharath, Anil Anthony .
IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (06) :26-38
[5]   Dynamic Range-Aware Uplink Transmit Power Control in LTE Networks: Establishing an Operational Range for LTE's Open-Loop Transmit Power Control Parameters (α, P0) [J].
Berger, Sascha ;
Almeroth, Bjoern ;
Suryaprakash, Vinay ;
Zanier, Paolo ;
Viering, Ingo ;
Fettweis, Gerhard .
IEEE WIRELESS COMMUNICATIONS LETTERS, 2014, 3 (05) :521-524
[6]  
Boussif Malek, 2010, 2010 European Wireless Conference (EW), P288, DOI 10.1109/EW.2010.5483432
[7]   Smart Mode Selection Using Online Reinforcement Learning for VR Broadband Broadcasting in D2D Assisted 5G HetNets [J].
Feng, Lei ;
Yang, Zhixiang ;
Yang, Yang ;
Que, Xiaoyu ;
Zhang, Kai .
IEEE TRANSACTIONS ON BROADCASTING, 2020, 66 (02) :600-611
[8]   Large Intelligent Surface-Assisted Wireless Communication Exploiting Statistical CSI [J].
Han, Yu ;
Tang, Wankai ;
Jin, Shi ;
Wen, Chao-Kai ;
Ma, Xiaoli .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2019, 68 (08) :8238-8242
[9]   Robust Transmission Design for Intelligent Reflecting Surface-Aided Secure Communication Systems With Imperfect Cascaded CSI [J].
Hong, Sheng ;
Pan, Cunhua ;
Ren, Hong ;
Wang, Kezhi ;
Chai, Kok Keong ;
Nallanathan, Arumugam .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2021, 20 (04) :2487-2501
[10]  
Hu X, 2021, CHINA COMMUN, V18, P25, DOI 10.23919/JCC.2021.07.003