Deep Reinforcement Learning Based Trajectory Design and Resource Allocation for UAV-Assisted Communications

被引:17
作者
Zhang, Chiya [1 ]
Li, Zhukun [1 ]
He, Chunlong [2 ]
Wang, Kezhi [3 ]
Pan, Cunhua [4 ]
机构
[1] Harbin Inst Technol, Sch Elect & Informat Engn, Shenzhen 518055, Peoples R China
[2] Shenzhen Univ, Guangdong Key Lab Intelligent Informat Proc, Shenzhen 518060, Peoples R China
[3] Brunel Univ London, Dept Comp Sci, London UB8 3PH, England
[4] Southeast Univ, Natl Mobile Commun Res Lab, Nanjing 211189, Peoples R China
基金
中国国家自然科学基金;
关键词
Unmanned aerial vehicles; deep reinforcement learning; 3-D trajectory design; uncertain flight time;
D O I
10.1109/LCOMM.2023.3292816
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
In this letter, we investigate the Unmanned Aerial Vehicles (UAVs)-assisted communications in three dimensional (3-D) environment, where one UAV is deployed to serve multiple user equipments (UEs). The locations and quality of service (QoS) requirement of the UEs are varying and the flying time of the UAV is unknown which depends on the battery of the UAVs. To address the issue, a proximal policy optimization 2 (PPO2)-based deep reinforcement learning (DRL) algorithm is proposed, which can control the UAV in an online manner. Specifically, it can allow the UAV to adjust its speed, direction and altitude so as to minimize the serving time of the UAV while satisfying the QoS requirement of the UEs. Simulation results are provided to demonstrate the effectiveness of the proposed framework.
引用
收藏
页码:2398 / 2402
页数:5
相关论文
共 10 条
[1]   Optimal LAP Altitude for Maximum Coverage [J].
Al-Hourani, Akram ;
Kandeepan, Sithamparanathan ;
Lardner, Simon .
IEEE WIRELESS COMMUNICATIONS LETTERS, 2014, 3 (06) :569-572
[2]   3-D Placement of an Unmanned Aerial Vehicle Base Station for Maximum Coverage of Users With Different QoS Requirements [J].
Alzenad, Mohamed ;
El-Keyi, Amr ;
Yanikomeroglu, Halim .
IEEE WIRELESS COMMUNICATIONS LETTERS, 2018, 7 (01) :38-41
[3]   3D UAV Trajectory Design and Frequency Band Allocation for Energy-Efficient and Fair Communication: A Deep Reinforcement Learning Approach [J].
Ding, Ruijin ;
Gao, Feifei ;
Shen, Xuemin .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2020, 19 (12) :7796-7809
[4]   Minimizing Mission Completion Time of UAVs by Jointly Optimizing the Flight and Data Collection Trajectory in UAV-Enabled WSNs [J].
Li, Min ;
He, Shuangshuang ;
Li, Hao .
IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (15) :13498-13510
[5]   Energy-Effective Offloading Scheme in UAV-Assisted C-RAN System [J].
Li, Xingquan ;
Zhang, Chiya ;
Zhao, Rujun ;
He, Chunlong ;
Zheng, Hongxia ;
Wang, Kezhi .
IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (13) :10821-10832
[6]  
Schulman J, 2017, Arxiv, DOI arXiv:1707.06347
[7]   Deep Reinforcement Learning Based Dynamic Trajectory Control for UAV-Assisted Mobile Edge Computing [J].
Wang, Liang ;
Wang, Kezhi ;
Pan, Cunhua ;
Xu, Wei ;
Aslam, Nauman ;
Nallanathan, Arumugam .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2022, 21 (10) :3536-3550
[8]   Completion Time Minimization in Wireless-Powered UAV-Assisted Data Collection System [J].
Wang, Zhen ;
Zhang, Guopeng ;
Wang, Qiu ;
Wang, Kezhi ;
Yang, Kun .
IEEE COMMUNICATIONS LETTERS, 2021, 25 (06) :1954-1958
[9]   Joint Trajectory and Communication Design for Multi-UAV Enabled Wireless Networks [J].
Wu, Qingqing ;
Zeng, Yong ;
Zhang, Rui .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2018, 17 (03) :2109-2121
[10]   Accessing From the Sky: A Tutorial on UAV Communications for 5G and Beyond [J].
Zeng, Yong ;
Wu, Qingqing ;
Zhang, Rui .
PROCEEDINGS OF THE IEEE, 2019, 107 (12) :2327-2375