Deep Reinforcement Learning Based Trajectory Design for Customized UAV-Aided NOMA Data Collection

被引:2
作者
Zhang, Lei [1 ]
Zhang, Yuandi [1 ]
Lu, Jiawangnan [1 ]
Xiao, Yunfa [1 ]
Zhang, Guanglin [1 ]
机构
[1] Donghua Univ, Coll Informat Sci & Technol, Shanghai 201620, Peoples R China
基金
中国国家自然科学基金;
关键词
Autonomous aerial vehicles; Channel models; Loss measurement; NOMA; Data collection; Trajectory; Three-dimensional displays; Customized channel model; data collection; optimized PPO; UAV communication;
D O I
10.1109/LWC.2024.3465509
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this letter, we design a customized channel model based on ray tracing (RT) and machine learning (ML). RT is used to generate path loss for selected areas. The generated path loss is trained through the Deep Neural Network (DNN). The channel model can output the path loss by inputting the transceiver's three-dimensional (3D) coordinates. We investigated the task of collecting data by unmanned aerial vehicle (UAV) based on the customized channel model. The time it takes for the UAV to finish collecting data generated by ground user equipment (UE) is minimized. We combine non-orthogonal multiple access (NOMA) to analyze UAVs' optimal 3D and 2D flight trajectories and demonstrate that 3D outperforms 2D. The optimized proximal policy optimization (optimized PPO) based deep reinforcement learning (DRL) algorithm is proposed to address this issue. The UAV can adjust its speed and direction. Simulation results demonstrate the effectiveness of the proposed customized channel model and algorithm.
引用
收藏
页码:3365 / 3369
页数:5
相关论文
共 14 条
[1]   Radio Channel Modeling for UAV Communication Over Cellular Networks [J].
Amorim, Rafhael ;
Huan Nguyen ;
Mogensen, Preben ;
Kovacs, Istvan Z. ;
Wigard, Jeroen ;
Sorensen, Troels B. .
IEEE WIRELESS COMMUNICATIONS LETTERS, 2017, 6 (04) :514-517
[2]  
Engstrom L, 2020, Arxiv, DOI arXiv:2005.12729
[3]   What Will the Future of UAV Cellular Communications Be? A Flight From 5G to 6G [J].
Geraci, Giovanni ;
Garcia-Rodriguez, Adrian ;
Azari, M. Mahdi ;
Lozano, Angel ;
Mezzavilla, Marco ;
Chatzinotas, Symeon ;
Chen, Yun ;
Rangan, Sundeep ;
Di Renzo, Marco .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2022, 24 (03) :1304-1335
[4]  
Huang S., The 37 implementation details of proximal policy optimization
[5]   Air-to-ground path loss prediction using ray tracing and measurement data jointly driven DNN [J].
Li, Hanpeng ;
Chen, Xiaomin ;
Mao, Kai ;
Zhu, Qiuming ;
Qiu, Yanheng ;
Ye, Xijuan ;
Zhong, Weizhi ;
Lin, Zhipeng .
COMPUTER COMMUNICATIONS, 2022, 196 :268-276
[6]  
Schulman J, 2017, Arxiv, DOI arXiv:1707.06347
[7]   Deep Reinforcement Learning Based Dynamic Trajectory Control for UAV-Assisted Mobile Edge Computing [J].
Wang, Liang ;
Wang, Kezhi ;
Pan, Cunhua ;
Xu, Wei ;
Aslam, Nauman ;
Nallanathan, Arumugam .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2022, 21 (10) :3536-3550
[8]   Reinforcement Learning-Based NOMA Power Allocation in the Presence of Smart Jamming [J].
Xiao, Liang ;
Li, Yanda ;
Dai, Canhuang ;
Dai, Huaiyu ;
Poor, H. Vincent .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2018, 67 (04) :3377-3389
[9]   Deep Reinforcement Learning Based Trajectory Design and Resource Allocation for UAV-Assisted Communications [J].
Zhang, Chiya ;
Li, Zhukun ;
He, Chunlong ;
Wang, Kezhi ;
Pan, Cunhua .
IEEE COMMUNICATIONS LETTERS, 2023, 27 (09) :2398-2402
[10]   Load Forecasting-Based Learning System for Energy Management With Battery Degradation Estimation: A Deep Reinforcement Learning Approach [J].
Zhang, Hongtao ;
Zhang, Guanglin ;
Zhao, Mingbo ;
Liu, Yuping .
IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, 2024, 70 (01) :2342-2352