Reinforcement Learning-Based Intelligent Decision-Making for Communication Parameters

被引:4
作者
Xie, Xia [1 ]
Dou, Zheng [1 ]
Zhang, Yabin [1 ]
机构
[1] Harbin Engn Univ, Dept Informat & Commun Engn, Harbin, Heilongjiang, Peoples R China
来源
KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS | 2022年 / 16卷 / 09期
基金
中国国家自然科学基金;
关键词
reinforcement learning; decision-making; Q-learning; cognitive radio; adaptive modulation and coding; COGNITIVE RADIO; ACCESS;
D O I
10.3837/tiis.2022.09.006
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The core of cognitive radio is the problem concerning intelligent decision-making for communication parameters, the objective of which is to find the most appropriate parameter configuration to optimize transmission performance. The current algorithms have the disadvantages of high dependence on prior knowledge, large amount of calculation, and high complexity. We propose a new decision-making model by making full use of the interactivity of reinforcement learning (RL) and applying the Q-learning algorithm. By simplifying the decision-making process, we avoid large-scale RL, reduce complexity and improve timeliness. The proposed model is able to find the optimal waveform parameter configuration for the communication system in complex channels without prior knowledge. Moreover, this model is more flexible than previous decision-making models. The simulation results demonstrate the effectiveness of our model. The model not only exhibits better decision-making performance in the AWGN channels than the traditional method, but also make reasonable decisions in the fading channels.
引用
收藏
页码:2942 / 2960
页数:19
相关论文
共 31 条
[21]  
O'Shea TJ, 2019, INT CONF COMPUT NETW, P681, DOI [10.1109/ICCNC.2019.8685573, 10.1109/iccnc.2019.8685573]
[22]   Bacterial Foraging Optimization [J].
Passino, Kevin M. .
INTERNATIONAL JOURNAL OF SWARM INTELLIGENCE RESEARCH, 2010, 1 (01) :1-16
[23]   Comparative performance analysis of evolutionary algorithm based parameter optimization in cognitive radio engine: A survey [J].
Pradhan, Pyari Mohan ;
Panda, Ganapati .
AD HOC NETWORKS, 2014, 17 :129-146
[24]  
Sutton RS, 2018, ADAPT COMPUT MACH LE, P1
[25]  
Tubachi S, 2017, 2017 IEEE INTERNATIONAL CONFERENCE ON POWER, CONTROL, SIGNALS AND INSTRUMENTATION ENGINEERING (ICPCSI), P564, DOI 10.1109/ICPCSI.2017.8391775
[26]   Deep Reinforcement Learning for Dynamic Multichannel Access in Wireless Networks [J].
Wang, Shangxing ;
Liu, Hanpeng ;
Gomes, Pedro Henrique ;
Krishnamachari, Bhaskar .
IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2018, 4 (02) :257-265
[27]  
Xu Dong, 2010, 2010 12th IEEE International Conference on Communication Technology (ICCT 2010), P893, DOI 10.1109/ICCT.2010.5688723
[28]  
Xu HY, 2013, 2013 15TH IEEE INTERNATIONAL CONFERENCE ON COMMUNICATION TECHNOLOGY (ICCT), P115, DOI 10.1109/ICCT.2013.6820357
[29]  
Xu HY, 2013, 2013 13TH INTERNATIONAL SYMPOSIUM ON COMMUNICATIONS AND INFORMATION TECHNOLOGIES (ISCIT): COMMUNICATION AND INFORMATION TECHNOLOGY FOR NEW LIFE STYLE BEYOND THE CLOUD, P143, DOI 10.1109/ISCIT.2013.6645839
[30]  
Yang Y., 2012, PROC 2012 SPRING C E, DOI 10.1109/SCET.2012.6342027