Joint path planning and power allocation of a cellular-connected UAV using apprenticeship learning via deep inverse reinforcement learning

被引:3
|
作者
Shamsoshoara, Alireza [1 ,2 ]
Lotfi, Fatemeh [2 ]
Mousavi, Sajad [2 ,3 ]
Afghah, Fatemeh [2 ]
Guevenc, Ismail [2 ,4 ]
机构
[1] No Arizona Univ, Sch Informat Comp & Cyber Syst, Flagstaff, AZ 86011 USA
[2] Clemson Univ, Dept Elect & Comp Engn, Clemson, SC 29634 USA
[3] Harvard Med Sch, Boston, MA USA
[4] North Carolina State Univ, Raleigh, NC USA
基金
美国国家科学基金会;
关键词
Apprenticeship learning; Cellular-connected drones; Inverse reinforcement learning; Path planning; UAV communication; COMMUNICATION; NETWORKS; ALTITUDE;
D O I
10.1016/j.comnet.2024.110789
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
This paper investigates an interference-aware joint path planning and power allocation mechanism for a cellular-connected unmanned aerial vehicle (UAV) in a sparse suburban environment. The UAV's goal is to fly from an initial point and reach a destination point by moving along the cells to guarantee the required quality of service (QoS). In particular, the UAV aims to maximize its uplink throughput and minimize interference to the ground user equipment (UEs) connected to neighboring cellular base stations (BSs), considering both the shortest path and limitations on flight resources. Expert knowledge is used to experience the scenario and define the desired behavior for the sake of the agent (i.e., UAV) training. To solve the problem, an apprenticeship learning method is utilized via inverse reinforcement learning (IRL) based on both Q-learning and deep reinforcement learning (DRL). The performance of this method is compared to learning from a demonstration technique called behavioral cloning (BC) using a supervised learning approach. Simulation and numerical results show that the proposed approach can achieve expert-level performance. We also demonstrate that, unlike the BC technique, the performance of our proposed approach does not degrade in unseen situations.
引用
收藏
页数:20
相关论文
共 50 条
  • [21] Multi-UAV Path Planning for Wireless Data Harvesting With Deep Reinforcement Learning
    Bayerlein, Harald
    Theile, Mirco
    Caccamo, Marco
    Gesbert, David
    IEEE OPEN JOURNAL OF THE COMMUNICATIONS SOCIETY, 2021, 2 : 1171 - 1187
  • [22] Joint Power Control and UAV Trajectory Design for Information Freshness via Deep Reinforcement Learning
    Li, Xinmin
    Yin, Baolin
    Yan, Jiaxin
    Zhang, Xiaoqiang
    Wei, Ran
    2022 IEEE 95TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2022-SPRING), 2022,
  • [23] Modeling and Analysis of Intermittent Federated Learning Over Cellular-Connected UAV Networks
    Liu, Chun-Hung
    Liang, Di-Chun
    Gau, Rung-Hung
    Wei, Lu
    2022 IEEE 95TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2022-SPRING), 2022,
  • [24] Joint 3D Deployment and Power Allocation for UAV-BS: A Deep Reinforcement Learning Approach
    Zhang, Meng
    Fu, Shu
    Fan, Qilin
    IEEE WIRELESS COMMUNICATIONS LETTERS, 2021, 10 (10) : 2309 - 2312
  • [25] Layerwise Quantum Deep Reinforcement Learning for Joint Optimization of UAV Trajectory and Resource Allocation
    Silvirianti
    Narottama, Bhaskara
    Shin, Soo Young
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (01) : 430 - 443
  • [26] Joint UAV Deployment and Resource Allocation: A Personalized Federated Deep Reinforcement Learning Approach
    Xu, Xinyi
    Feng, Gang
    Qin, Shuang
    Liu, Yijing
    Sun, Yao
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2024, 73 (03) : 4005 - 4018
  • [27] Benchmarking Off-Policy Deep Reinforcement Learning Algorithms for UAV Path Planning
    Garg, Shaswat
    Masnavi, Houman
    Fidan, Baris
    Janabi-Sharifi, Farrokh
    Mantegh, Iraj
    2024 INTERNATIONAL CONFERENCE ON UNMANNED AIRCRAFT SYSTEMS, ICUAS, 2024, : 317 - 323
  • [28] Jamming-Resilient Path Planning for Multiple UAVs via Deep Reinforcement Learning
    Wang, Xueyuan
    Gursoy, M. Cenk
    Erpek, Tugba
    Sagduyu, Yalin E.
    2021 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS (ICC WORKSHOPS), 2021,
  • [29] Deep Reinforcement Learning Assisted UAV Path Planning Relying on Cumulative Reward Mode and Region Segmentation
    Wang, Zhipeng
    Ng, Soon Xin
    EI-Hajjar, Mohammed
    IEEE OPEN JOURNAL OF VEHICULAR TECHNOLOGY, 2024, 5 : 737 - 751
  • [30] UAV swarm path planning with reinforcement learning for field prospecting
    Puente-Castro, Alejandro
    Rivero, Daniel
    Pazos, Alejandro
    Fernandez-Blanco, Enrique
    APPLIED INTELLIGENCE, 2022, 52 (12) : 14101 - 14118