A Novel Deep Q-learning Method for Dynamic Spectrum Access

被引:4
作者
Tomovic, S. [1 ]
Radusinovic, I [1 ]
机构
[1] Univ Montenegro, Fac Elect Engn, Dzordza Vasingtona Bb, Podgorica 81000, Montenegro
来源
2020 28TH TELECOMMUNICATIONS FORUM (TELFOR) | 2020年
关键词
Cognitive radio; Reinforcement learning; OPTIMALITY;
D O I
10.1109/telfor51502.2020.9306591
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
In this paper, we propose a new Dynamic Spectrum Access (DSA) method for multi-channel wireless networks. We assume that DSA nodes, as secondary users, do not have prior knowledge of the system dynamics. Since DSA nodes have only partial observability of the channel states, the problem is formulated as a Partially Observable Markov Decision Process (POMDP) with exponential time complexity. We have developed a novel Deep Reinforcement Learning (DRL) based DSA method which combines a double deep Q-learning architecture with a recurrent neural network and takes advantage of a prioritized experience buffer. The simulation analysis shows that the proposed method accurately predicts the channels state based on the fixed-length history of partial observations. Compared with other DRL methods, the proposed solution is able to find a near-optimal policy in a smaller number of iterations and suits a wide range of communication environment conditions. The performance improvement increases with the number of channels and a channel state transition uncertainty.
引用
收藏
页码:9 / 12
页数:4
相关论文
共 50 条
[31]   A META-PRECONDITIONING APPROACH FOR DEEP Q-LEARNING [J].
Evmorfos, Spilios ;
Petropulu, Athina P. .
2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, :6485-6489
[32]   Autonomous Warehouse Robot using Deep Q-Learning [J].
Peyas, Ismot Sadik ;
Hasan, Zahid ;
Tushar, Md Rafat Rahman ;
Al Musabbir ;
Azni, Raisa Mehjabin ;
Siddique, Shahnewaz .
2021 IEEE REGION 10 CONFERENCE (TENCON 2021), 2021, :857-862
[33]   Deep spectral Q-learning with application to mobile health [J].
Gao, Yuhe ;
Shi, Chengchun ;
Song, Rui .
STAT, 2023, 12 (01)
[34]   Deep Spatial Q-Learning for Infectious Disease Control [J].
Liu, Zhishuai ;
Clifton, Jesse ;
Laber, Eric B. ;
Drake, John ;
Fang, Ethan X. .
JOURNAL OF AGRICULTURAL BIOLOGICAL AND ENVIRONMENTAL STATISTICS, 2023, 28 (04) :749-773
[35]   Autonomous Overtaking Decision Making of Driverless Bus Based on Deep Q-learning Method [J].
Yu, Lingli ;
Shao, Xuanya ;
Yan, Xiaoxin .
2017 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (IEEE ROBIO 2017), 2017, :2267-2272
[36]   A novel dynamic integration approach for multiple load forecasts based on Q-learning algorithm [J].
Ma, Minhua ;
Jin, Bingjie ;
Luo, Shuxin ;
Guo, Shaoqing ;
Huang, Hongwei .
INTERNATIONAL TRANSACTIONS ON ELECTRICAL ENERGY SYSTEMS, 2020, 30 (07)
[37]   Backward Q-learning: The combination of Sarsa algorithm and Q-learning [J].
Wang, Yin-Hao ;
Li, Tzuu-Hseng S. ;
Lin, Chih-Jui .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2013, 26 (09) :2184-2193
[38]   The Application of Deep Reinforcement Learning to Distributed Spectrum Access in Dynamic Heterogeneous Environments With Partial Observations [J].
Xu, Yue ;
Yu, Jianyuan ;
Buehrer, R. Michael .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2020, 19 (07) :4494-4506
[39]   A scalable Deep Q-Learning approach for hot stamping process under dynamic control environment [J].
Nievas, Nuria ;
Pages-Bernaus, Adela ;
Abio, Albert ;
Lange, Danillo ;
Garcia-Llamas, Eduard ;
Grane, Marc ;
Pujante, Jaume ;
Echeverria, Lluis ;
Bonada, Francesc .
INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2025, 63 (08) :2831-2852
[40]   Online adaptive Q-learning method for fully cooperative linear quadratic dynamic games [J].
Xinxing Li ;
Zhihong Peng ;
Lei Jiao ;
Lele Xi ;
Junqi Cai .
Science China Information Sciences, 2019, 62