Joint path planning and power allocation of a cellular-connected UAV using apprenticeship learning via deep inverse reinforcement learning

被引:3
作者
Shamsoshoara, Alireza [1 ,2 ]
Lotfi, Fatemeh [2 ]
Mousavi, Sajad [2 ,3 ]
Afghah, Fatemeh [2 ]
Guevenc, Ismail [2 ,4 ]
机构
[1] No Arizona Univ, Sch Informat Comp & Cyber Syst, Flagstaff, AZ 86011 USA
[2] Clemson Univ, Dept Elect & Comp Engn, Clemson, SC 29634 USA
[3] Harvard Med Sch, Boston, MA USA
[4] North Carolina State Univ, Raleigh, NC USA
基金
美国国家科学基金会;
关键词
Apprenticeship learning; Cellular-connected drones; Inverse reinforcement learning; Path planning; UAV communication; COMMUNICATION; NETWORKS; ALTITUDE;
D O I
10.1016/j.comnet.2024.110789
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
This paper investigates an interference-aware joint path planning and power allocation mechanism for a cellular-connected unmanned aerial vehicle (UAV) in a sparse suburban environment. The UAV's goal is to fly from an initial point and reach a destination point by moving along the cells to guarantee the required quality of service (QoS). In particular, the UAV aims to maximize its uplink throughput and minimize interference to the ground user equipment (UEs) connected to neighboring cellular base stations (BSs), considering both the shortest path and limitations on flight resources. Expert knowledge is used to experience the scenario and define the desired behavior for the sake of the agent (i.e., UAV) training. To solve the problem, an apprenticeship learning method is utilized via inverse reinforcement learning (IRL) based on both Q-learning and deep reinforcement learning (DRL). The performance of this method is compared to learning from a demonstration technique called behavioral cloning (BC) using a supervised learning approach. Simulation and numerical results show that the proposed approach can achieve expert-level performance. We also demonstrate that, unlike the BC technique, the performance of our proposed approach does not degrade in unseen situations.
引用
收藏
页数:20
相关论文
共 50 条
  • [31] UAV swarm path planning with reinforcement learning for field prospecting
    Alejandro Puente-Castro
    Daniel Rivero
    Alejandro Pazos
    Enrique Fernandez-Blanco
    Applied Intelligence, 2022, 52 : 14101 - 14118
  • [32] Deep reinforcement learning path planning and task allocation for multi-robot collaboration
    Li, Zhixian
    Shi, Nianfeng
    Zhao, Liguo
    Zhang, Mengxia
    ALEXANDRIA ENGINEERING JOURNAL, 2024, 109 : 408 - 423
  • [33] Apprenticeship Bootstrapping: Inverse Reinforcement Learning in a Multi-Skill UAV-UGV Coordination Task
    Hung The Nguyen
    Garratt, Matthew
    Lam Thu Bui
    Abbass, Hussein
    PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS (AAMAS' 18), 2018, : 2204 - 2206
  • [34] Inverse Reinforcement Learning Meets Power Allocation in Multi-user Cellular Networks
    Zhang, Ruichen
    Xiong, Ke
    Tian, Xingcong
    Lu, Yang
    Fan, Pingyi
    Ben Letaief, Khaled
    IEEE INFOCOM 2022 - IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS (INFOCOM WKSHPS), 2022,
  • [35] Mobile Service Robot Path Planning Using Deep Reinforcement Learning
    Kumaar, A. A. Nippun
    Kochuvila, Sreeja
    IEEE ACCESS, 2023, 11 : 100083 - 100096
  • [36] Socially Adaptive Path Planning in Human Environments Using Inverse Reinforcement Learning
    Beomjoon Kim
    Joelle Pineau
    International Journal of Social Robotics, 2016, 8 : 51 - 66
  • [37] Socially Adaptive Path Planning in Human Environments Using Inverse Reinforcement Learning
    Kim, Beomjoon
    Pineau, Joelle
    INTERNATIONAL JOURNAL OF SOCIAL ROBOTICS, 2016, 8 (01) : 51 - 66
  • [38] A Deep Reinforcement Learning Approach for UAV Path Planning Incorporating Vehicle Dynamics with Acceleration Control
    Sabzekar, Sina
    Samadzad, Mahdi
    Mehditabrizi, Asal
    Tak, Ala Nekouvaght
    UNMANNED SYSTEMS, 2024, 12 (03) : 477 - 498
  • [39] Robot Path Planning Based on Deep Reinforcement Learning
    Zhang, Rui
    Jiang, Yuhao
    Wu Fenghua
    2022 34TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2022, : 1697 - 1701
  • [40] Connectivity-Aware 3D UAV Path Design With Deep Reinforcement Learning
    Xie, Hao
    Yang, Dingcheng
    Xiao, Lin
    Lyu, Jiangbin
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2021, 70 (12) : 13022 - 13034