Quadrotor Path Following and Reactive Obstacle Avoidance with Deep Reinforcement Learning

被引:14
|
作者
Rubi, Bartomeu [1 ]
Morcego, Bernardo [1 ]
Perez, Ramon [1 ]
机构
[1] Univ Politecn Catalunya UPC, Res Ctr Supervis Safety & Automat Control CS2AC, Rbla St Nebridi 22, Terrassa, Spain
关键词
Unmanned aerial vehicles; Obstacle avoidance; Path following; Deep reinforcement learning; LIDAR; Deep deterministic policy gradient; UNMANNED AERIAL VEHICLES; COLLISION-AVOIDANCE; TRACKING; UAVS;
D O I
10.1007/s10846-021-01491-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A deep reinforcement learning approach for solving the quadrotor path following and obstacle avoidance problem is proposed in this paper. The problem is solved with two agents: one for the path following task and another one for the obstacle avoidance task. A novel structure is proposed, where the action computed by the obstacle avoidance agent becomes the state of the path following agent. Compared to traditional deep reinforcement learning approaches, the proposed method allows to interpret the training process outcomes, is faster and can be safely trained on the real quadrotor. Both agents implement the Deep Deterministic Policy Gradient algorithm. The path following agent was developed in a previous work. The obstacle avoidance agent uses the information provided by a low-cost LIDAR to detect obstacles around the vehicle. Since LIDAR has a narrow field-of-view, an approach for providing the agent with a memory of the previously seen obstacles is developed. A detailed description of the process of defining the state vector, the reward function and the action of this agent is given. The agents are programmed in python/tensorflow and are trained and tested in the RotorS/gazebo platform. Simulations results prove the validity of the proposed approach.
引用
收藏
页数:17
相关论文
共 50 条
  • [1] Quadrotor Path Following and Reactive Obstacle Avoidance with Deep Reinforcement Learning
    Bartomeu Rubí
    Bernardo Morcego
    Ramon Pérez
    Journal of Intelligent & Robotic Systems, 2021, 103
  • [2] A Deep Reinforcement Learning Approach for Path Following on a Quadrotor
    Rubi, Bartomeu
    Morcego, Bernardo
    Perez, Ramon
    2020 EUROPEAN CONTROL CONFERENCE (ECC 2020), 2020, : 1092 - 1098
  • [3] Deep reinforcement learning for quadrotor path following with adaptive velocity
    Rubi, Bartomeu
    Morcego, Bernardo
    Perez, Ramon
    AUTONOMOUS ROBOTS, 2021, 45 (01) : 119 - 134
  • [4] Deep reinforcement learning for quadrotor path following with adaptive velocity
    Bartomeu Rubí
    Bernardo Morcego
    Ramon Pérez
    Autonomous Robots, 2021, 45 : 119 - 134
  • [5] Path-Following and Obstacle Avoidance Control of Nonholonomic Wheeled Mobile Robot Based on Deep Reinforcement Learning
    Cheng, Xiuquan
    Zhang, Shaobo
    Cheng, Sizhu
    Xia, Qinxiang
    Zhang, Junhao
    APPLIED SCIENCES-BASEL, 2022, 12 (14):
  • [6] Path Following Control of Unmanned Quadrotor Helicopter with Obstacle Avoidance Capability
    Liu, Zhixiang
    Ciarletta, Laurent
    Yuan, Chi
    Zhang, Youmin
    Theilliol, Didier
    2017 INTERNATIONAL CONFERENCE ON UNMANNED AIRCRAFT SYSTEMS (ICUAS'17), 2017, : 304 - 309
  • [7] Obstacle Avoidance Planning of Virtual Robot Picking Path Based on Deep Reinforcement Learning
    Xiong J.
    Li Z.
    Chen S.
    Zheng Z.
    Nongye Jixie Xuebao/Transactions of the Chinese Society for Agricultural Machinery, 2020, 51 : 1 - 10
  • [8] Dynamic Obstacle Avoidance and Path Planning through Reinforcement Learning
    Almazrouei, Khawla
    Kamel, Ibrahim
    Rabie, Tamer
    APPLIED SCIENCES-BASEL, 2023, 13 (14):
  • [9] Self-Configuring Robot Path Planning With Obstacle Avoidance via Deep Reinforcement Learning
    Sangiovanni, Bianca
    Incremona, Gian Paolo
    Piastra, Marco
    Ferrara, Antonella
    IEEE CONTROL SYSTEMS LETTERS, 2021, 5 (02): : 397 - 402
  • [10] Path Planning of Mobile Robot in Dynamic Obstacle Avoidance Environment Based on Deep Reinforcement Learning
    Zhang, Qingfeng
    Ma, Wenpeng
    Zheng, Qingchun
    Zhai, Xiaofan
    Zhang, Wenqian
    Zhang, Tianchang
    Wang, Shuo
    IEEE ACCESS, 2024, 12 : 189136 - 189152