Deep Reinforcement Learning-Based Anti-Jamming Algorithm Using Dual Action Network

被引:8
作者
Li, Xiangchen [1 ]
Chen, Jienan [1 ]
Ling, Xiang [1 ]
Wu, Tingyong [1 ]
机构
[1] Univ Elect Sci & Technol China, Natl Key Lab Sci & Technol Commun, Chengdu 611731, Sichuan, Peoples R China
基金
中国国家自然科学基金;
关键词
Jamming; Wireless communication; Games; Receivers; Electromagnetics; Testing; Switches; Anti-jamming; Markov decision process; Index Terms; deep reinforcement learning; dual action network; action feedback mechanism; power efficiency; frequency switching overhead; field testing; STACKELBERG GAME; SYSTEMS;
D O I
10.1109/TWC.2022.3227575
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Due to the open nature of wireless communication, malicious electromagnetic jamming has long been a severe threat to the establishment and stability of communication links. To address this anti-jamming problem, a Markov decision process (MDP) with a two-dimensional action space consisting of transmit frequency and power is proposed in this paper, modeling the interaction between a normal communication link and the presence of malicious jammers in a frequency hopping (FH) communication system. Furthermore, we also prove the existence of the deterministic optimal policy of the proposed model theoretically. To obtain a policy for the communication link to avoid being jammed, the Dual Action Network-Based Deep Reinforcement Learning Algorithm, and Action Feedback Mechanism are proposed. The energy consumption and frequency switching overhead are considered and evaluated in both the proposed model and the algorithm. Finally, the proposed model and algorithm are verified not only in a virtual simulation environment but also in the field testing environment. The result suggests that the proposed algorithm is of great practical value for solving anti-jamming problems.
引用
收藏
页码:4625 / 4637
页数:13
相关论文
共 39 条
[1]  
Aref MA, 2017, 2017 COGNITIVE COMMUNICATIONS FOR AEROSPACE APPLICATIONS WORKSHOP (CCAA)
[2]  
Bi Y., 2019, PROC IEEE INT C COMM, P1
[3]  
Bozic M., 2011, 2011 19th Telecommunications Forum Telfor (TELFOR), P425, DOI 10.1109/TELFOR.2011.6143578
[4]  
Chen JC, 2019, IEEE CONF COMPU INTE
[5]   AuTO: Scaling Deep Reinforcement Learning for Datacenter-Scale Automatic Traffic Optimization [J].
Chen, Li ;
Lingys, Justinas ;
Chen, Kai ;
Liu, Feng .
PROCEEDINGS OF THE 2018 CONFERENCE OF THE ACM SPECIAL INTEREST GROUP ON DATA COMMUNICATION (SIGCOMM '18), 2018, :191-205
[6]  
Feinberg E. A., 2011, WILEY ENCY OPERATION
[7]   Deep Reinforcement Learning for Distributed Dynamic MISO Downlink-Beamforming Coordination [J].
Ge, Jungang ;
Liang, Ying-Chang ;
Joung, Jingon ;
Sun, Sumei .
IEEE TRANSACTIONS ON COMMUNICATIONS, 2020, 68 (10) :6070-6085
[8]  
Gleave A., 2019, arXiv
[9]  
Gouissem A, 2020, INT WIREL COMMUN, P770, DOI 10.1109/IWCMC48107.2020.9148376
[10]   Multi-Hop RIS-Empowered Terahertz Communications: A DRL-Based Hybrid Beamforming Design [J].
Huang, Chongwen ;
Yang, Zhaohui ;
Alexandropoulos, George C. ;
Xiong, Kai ;
Wei, Li ;
Yuen, Chau ;
Zhang, Zhaoyang ;
Debbah, Merouane .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2021, 39 (06) :1663-1677