A deep reinforcement learning framework and its implementation for UAV-aided covert communication

被引:0
作者
Fu, Shu [1 ]
Su, Yi [1 ]
Zhang, Zhi [2 ]
Yin, Liuguo [3 ]
机构
[1] Chongqing Univ, Coll Microelect & Commun Engn, Chongqing 400044, Peoples R China
[2] Tsinghua Univ, Dept Elect Engn, Beijing 100084, Peoples R China
[3] Tsinghua Univ, Beijing Natl Res Ctr Informat Sci & Technol, Beijing 100084, Peoples R China
基金
中国国家自然科学基金;
关键词
Covert communication; Unmanned aerial vehicle; Deep reinforcement learning; Trajectory planning; Power allocation; Communication systems; INTELLIGENT REFLECTING SURFACE; RESOURCE-ALLOCATION; POWER-CONTROL; OPTIMIZATION; INFORMATION; LIMITS;
D O I
10.1016/j.cja.2024.09.033
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
In this work, we consider an Unmanned Aerial Vehicle (UAV)-aided covert transmission network, which adopts the uplink transmission of Communication Nodes (CNs) as a cover to facilitate covert transmission to a Primary Communication Node (PCN). Specifically, all nodes transmit to the UAV exploiting uplink non-Orthogonal Multiple Access (NOMA), while the UAV performs covert transmission to the PCN at the same frequency. To minimize the average age of covert information, we formulate a joint optimization problem of UAV trajectory and power allocation designing subject to multi-dimensional constraints including covertness demand, communication quality requirement, maximum flying speed, and the maximum available resources. To address this problem, we embed Signomial Programming (SP) into Deep Reinforcement Learning (DRL) and propose a DRL framework capable of handling the constrained Markov decision processes, named SP embedded Soft Actor-Critic (SSAC). By adopting SSAC, we achieve the joint optimization of UAV trajectory and power allocation. Our simulations show the optimized UAV trajectory and verify the superiority of SSAC compared with various existing baseline schemes. The results of this study suggest that by maintaining appropriate distances from both the PCN and CNs, one can effectively enhance the performance of covert communication by reducing the detection probability of the CNs. (c) 2024 Chinese Society of Aeronautics and Astronautics. Published by Elsevier Ltd. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
引用
收藏
页数:15
相关论文
共 51 条
[1]   Optimal LAP Altitude for Maximum Coverage [J].
Al-Hourani, Akram ;
Kandeepan, Sithamparanathan ;
Lardner, Simon .
IEEE WIRELESS COMMUNICATIONS LETTERS, 2014, 3 (06) :569-572
[2]   Deep Reinforcement Learning A brief survey [J].
Arulkumaran, Kai ;
Deisenroth, Marc Peter ;
Brundage, Miles ;
Bharath, Anil Anthony .
IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (06) :26-38
[3]   Hiding Information in Noise: Fundamental Limits of Covert Wireless Communication [J].
Bash, Boulat A. ;
Goeckel, Dennis ;
Towsley, Don ;
Guha, Saikat .
IEEE COMMUNICATIONS MAGAZINE, 2015, 53 (12) :26-+
[4]   Limits of Reliable Communication with Low Probability of Detection on AWGN Channels [J].
Bash, Boulat A. ;
Goeckel, Dennis ;
Towsley, Don .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2013, 31 (09) :1921-1930
[5]   A tutorial on geometric programming [J].
Boyd, Stephen ;
Kim, Seung-Jean ;
Vandenberghe, Lieven ;
Hassibi, Arash .
OPTIMIZATION AND ENGINEERING, 2007, 8 (01) :67-127
[6]   Deep Reinforcement Learning For Multi-User Access Control in Non-Terrestrial Networks [J].
Cao, Yang ;
Lien, Shao-Yu ;
Liang, Ying-Chang .
IEEE TRANSACTIONS ON COMMUNICATIONS, 2021, 69 (03) :1605-1619
[7]   Multi-Antenna Covert Communication via Full-Duplex Jamming Against a Warden With Uncertain Locations [J].
Chen, Xinying ;
Sun, Wen ;
Xing, Chengwen ;
Zhao, Nan ;
Chen, Yunfei ;
Yu, F. Richard ;
Nallanathan, Arumugam .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2021, 20 (08) :5467-5480
[8]   NOMA-Based Multi-User Mobile Edge Computation Offloading via Cooperative Multi-Agent Deep Reinforcement Learning [J].
Chen, Zhao ;
Zhang, Lei ;
Pei, Yukui ;
Jiang, Chunxiao ;
Yin, Liuguo .
IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2022, 8 (01) :350-364
[9]   Power control by geometric programming [J].
Chiang, Mung ;
Tan, Chee Wei ;
Palomar, Daniel P. ;
O'Neill, Daniel ;
Julian, David .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2007, 6 (07) :2640-2651
[10]  
Dan Wang, 2018, IEEE Communications Magazine, V56, P114, DOI 10.1109/MCOM.2018.1701310