A Novel Deep Q-learning Method for Dynamic Spectrum Access

被引:4
作者
Tomovic, S. [1 ]
Radusinovic, I [1 ]
机构
[1] Univ Montenegro, Fac Elect Engn, Dzordza Vasingtona Bb, Podgorica 81000, Montenegro
来源
2020 28TH TELECOMMUNICATIONS FORUM (TELFOR) | 2020年
关键词
Cognitive radio; Reinforcement learning; OPTIMALITY;
D O I
10.1109/telfor51502.2020.9306591
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
In this paper, we propose a new Dynamic Spectrum Access (DSA) method for multi-channel wireless networks. We assume that DSA nodes, as secondary users, do not have prior knowledge of the system dynamics. Since DSA nodes have only partial observability of the channel states, the problem is formulated as a Partially Observable Markov Decision Process (POMDP) with exponential time complexity. We have developed a novel Deep Reinforcement Learning (DRL) based DSA method which combines a double deep Q-learning architecture with a recurrent neural network and takes advantage of a prioritized experience buffer. The simulation analysis shows that the proposed method accurately predicts the channels state based on the fixed-length history of partial observations. Compared with other DRL methods, the proposed solution is able to find a near-optimal policy in a smaller number of iterations and suits a wide range of communication environment conditions. The performance improvement increases with the number of channels and a channel state transition uncertainty.
引用
收藏
页码:9 / 12
页数:4
相关论文
共 50 条
[41]   Online adaptive Q-learning method for fully cooperative linear quadratic dynamic games [J].
Xinxing LI ;
Zhihong PENG ;
Lei JIAO ;
Lele XI ;
Junqi CAI .
ScienceChina(InformationSciences), 2019, 62 (12) :164-177
[42]   Online adaptive Q-learning method for fully cooperative linear quadratic dynamic games [J].
Li, Xinxing ;
Peng, Zhihong ;
Jiao, Lei ;
Xi, Lele ;
Cai, Junqi .
SCIENCE CHINA-INFORMATION SCIENCES, 2019, 62 (12)
[43]   Algorithms for Dynamic Spectrum Access With Learning for Cognitive Radio [J].
Unnikrishnan, Jayakrishnan ;
Veeravalli, Venugopal V. .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2010, 58 (02) :750-760
[44]   Enhanced Machine Learning Algorithms: Deep Learning, Reinforcement Learning, ana Q-Learning [J].
Park, Ji Su ;
Park, Jong Hyuk .
JOURNAL OF INFORMATION PROCESSING SYSTEMS, 2020, 16 (05) :1001-1007
[45]   A novel double-mGBDT-based Q-learning [J].
Fu, Qiming ;
Ma, Shuai ;
Tian, Dawei ;
Chen, JianPing ;
Gao, Zhen ;
Zhong, Shan .
INTERNATIONAL JOURNAL OF MODELLING IDENTIFICATION AND CONTROL, 2021, 37 (3-4) :232-239
[46]   Deep Multi-User Reinforcement Learning for Distributed Dynamic Spectrum Access [J].
Naparstek, Oshri ;
Cohen, Kobi .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2019, 18 (01) :310-323
[47]   Learning rates for Q-learning [J].
Even-Dar, E ;
Mansour, Y .
JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 5 :1-25
[48]   ADAPTIVE CONTENTION WINDOW DESIGN USING DEEP Q-LEARNING [J].
Kumar, Abhishek ;
Verma, Gunjan ;
Rao, Chirag ;
Swami, Ananthram ;
Segarra, Santiago .
2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, :4950-4954
[49]   Entropy-Based Prioritized Sampling in Deep Q-Learning [J].
Ramicic, Mirza ;
Bonarini, Andrea .
2017 2ND INTERNATIONAL CONFERENCE ON IMAGE, VISION AND COMPUTING (ICIVC 2017), 2017, :1068-1072
[50]   Performing Deep Recurrent Double Q-Learning for Atari Games [J].
Moreno-Vera, Felipe .
2019 IEEE LATIN AMERICAN CONFERENCE ON COMPUTATIONAL INTELLIGENCE (LA-CCI), 2019, :125-128