Deep Reinforcement Learning for RAN Optimization and Control

被引:0
作者
Chen, Yu [1 ,2 ,3 ]
Chen, Jie [4 ]
Krishnamurthi, Ganesh [4 ]
Yang, Huijing [4 ]
Wang, Huahui [4 ]
Zhao, Wenjie [4 ]
机构
[1] Carnegie Mellon Univ, Neurosci Inst, 5000 Forbes Ave, Pittsburgh, PA 15213 USA
[2] Carnegie Mellon Univ, Machine Learning Dept, 5000 Forbes Ave, Pittsburgh, PA 15213 USA
[3] AT&T Labs, Summer Internship Program, Atlanta, GA USA
[4] AT&T Labs, 1 AT&T Way, Bedminster, NJ USA
来源
2021 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC) | 2021年
关键词
self-organizing network; radio access networks; deep reinforcement learning; testbed;
D O I
10.1109/WCNC49053.2021.9417275
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Due to the high variability of the traffic in the radio access network (RAN), fixed network configurations are not flexible enough to achieve optimal performance. Our vendors provide several settings of the eNodeB to optimize the RAN performance, such as media access control scheduler, loading balance, etc. But the detailed mechanisms of the eNodeB configurations are usually very complicated and not disclosed, not to mention the large key performance indicators (KPIs) space needed to be considered. These make constructing a simulator, offline tuning, or rule-based solutions difficult. We aim to build an intelligent controller without strong assumption or domain knowledge about the RAN and can run 24/7 without supervision. To achieve this goal, we first build a closed-loop control testbed RAN in a lab environment with one eNodeB provided by one of the largest wireless vendors and four smartphones. Next, we build a double Q network agent trained with the live feedback of the key performance indicators from the RAN. Our work proved the effectiveness of applying deep reinforcement learning to improve network performance in a real RAN network environment.
引用
收藏
页数:6
相关论文
共 25 条
[1]   Learn to Schedule (LEASCH): A Deep Reinforcement Learning Approach for Radio Resource Scheduling in the 5G MAC Layer [J].
Al-Tam, Faroq ;
Correia, Noelia ;
Rodriguez, Jonathan .
IEEE ACCESS, 2020, 8 :108088-108101
[2]  
[Anonymous], 2012, Int. J. Inf. Electron. Eng.
[3]  
[Anonymous], 2011, 32450 3GPP TS
[4]  
Challita U., 2017, ARXIV PREPRINT ARXIV
[5]  
Chen MH, 2017, PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P1511
[6]  
Chinchali S, 2018, AAAI CONF ARTIF INTE, P766
[7]   D-DASH: A Deep Q-Learning Framework for DASH Video Streaming [J].
Gadaleta, Matteo ;
Chiariotti, Federico ;
Rossi, Michele ;
Zanella, Andrea .
IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2017, 3 (04) :703-718
[8]  
García J, 2015, J MACH LEARN RES, V16, P1437
[9]  
He Y., 2017, PROC 6 ACM S DEV ANA, P47
[10]  
Hessel M, 2018, AAAI CONF ARTIF INTE, P3215