UAV autonomous obstacle avoidance via causal reinforcement learning

被引:1
作者
Sun, Tao [1 ]
Gu, Jiaojiao [1 ]
Mou, Junjie [1 ]
机构
[1] Naval Aeronaut Univ, Yantai 264001, Peoples R China
关键词
Unmanned aerial vehicles (UAVs); Obstacle avoidance; Navigation; Causal inference; Reinforcement learning; SCALE ESTIMATION;
D O I
10.1016/j.displa.2025.102966
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The role of unmanned aerial vehicles (UAVs) in everyday life is becoming increasingly important, and there is a growing demand for UAVs to autonomously perform obstacle avoidance and navigation tasks. Traditional UAV navigation methods typically divide the navigation problem into three stages: perception, mapping, and path planning. However, this approach significantly increases processing delays, causing UAVs to lose their agility advantage. In this paper, we propose a causal reinforcement learning-based end-to-end navigation strategy that directly learns from data, bypassing the explicit mapping and planning steps, thus enhancing responsiveness. To address the issue where using a continuous action space prevents the agent from learning effective experiences from past actions, we introduce an Actor-Critic method with a fixed horizontal plane and a discretized action space. This approach enhances the efficiency of sampling from the experience replay buffer and stabilizes the optimization process, ultimately improving the success rate of the reinforcement learning algorithm in UAV obstacle avoidance and navigation tasks. Furthermore, to overcome the limited generalization capability of end-to-end methods, we incorporate causal inference into the reinforcement learning training process. This step mitigates overfitting caused by insufficient interaction with the environment during training, thereby increasing the success rate of UAVs in performing obstacle avoidance and navigation tasks in unfamiliar environments. We validate the effectiveness of causal inference in improving the generalization capability of the reinforcement learning algorithm by using convergence steps in the training environment and navigation success rates of random targets in the testing environment as quantitative metrics. The results demonstrate that causal inference can effectively reduce overfitting of the policy network to the training environment.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] An Obstacle-Avoidance Motion Planning Method for Redundant Space Robot via Reinforcement Learning
    Huang, Zeyuan
    Chen, Gang
    Shen, Yue
    Wang, Ruiquan
    Liu, Chuankai
    Zhang, Long
    [J]. ACTUATORS, 2023, 12 (02)
  • [32] Memory-Based Deep Reinforcement Learning for Obstacle Avoidance in UAV With Limited Environment Knowledge
    Singla, Abhik
    Padakandla, Sindhu
    Bhatnagar, Shalabh
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (01) : 107 - 118
  • [33] An obstacle avoidance method for robotic arm based on reinforcement learning
    Wu, Peng
    Su, Heng
    Dong, Hao
    Liu, Tengfei
    Li, Min
    Chen, Zhihao
    [J]. INDUSTRIAL ROBOT-THE INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH AND APPLICATION, 2025, 52 (01): : 9 - 17
  • [34] Robot obstacle avoidance system using deep reinforcement learning
    Zhu, Xiaojun
    Liang, Yinghao
    Sun, Hanxu
    Wang, Xueqian
    Ren, Bin
    [J]. INDUSTRIAL ROBOT-THE INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH AND APPLICATION, 2022, 49 (02): : 301 - 310
  • [35] Q-Learning for Autonomous Mobile Robot Obstacle Avoidance
    Ribeiro, Tiago
    Goncalves, Fernando
    Garcia, Ines
    Lopes, Gil
    Fernando Ribeiro, A.
    [J]. 2019 19TH IEEE INTERNATIONAL CONFERENCE ON AUTONOMOUS ROBOT SYSTEMS AND COMPETITIONS (ICARSC 2019), 2019, : 243 - 249
  • [36] Optimal reinforcement learning and probabilistic-risk-based path planning and following of autonomous vehicles with obstacle avoidance
    Taghavifar, Hamid
    Taghavifar, Leyla
    Hu, Chuan
    Wei, Chongfeng
    Qin, Yechen
    [J]. PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART D-JOURNAL OF AUTOMOBILE ENGINEERING, 2024, 238 (06) : 1427 - 1439
  • [37] Reinforcement Learning for Autonomous Agents: Scene-Specific Dynamic Obstacle Avoidance and Target Pursuit in Unknown Environments
    Tang, Zixiang
    Fu, Fa
    Lu, Gaoshang
    Chen, Da
    [J]. IEEE ACCESS, 2024, 12 : 145496 - 145510
  • [38] OBSTACLE AVOIDANCE FOR AUTONOMOUS MOWING
    Creed, Ben C.
    Arsenault, Aaron
    Velinsky, Steven A.
    Lasky, Ty A.
    [J]. MECHANICS BASED DESIGN OF STRUCTURES AND MACHINES, 2012, 40 (03) : 334 - 348
  • [39] A fuzzy controller with supervised learning assisted reinforcement learning algorithm for obstacle avoidance
    Ye, C
    Yung, NHC
    Wang, DW
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2003, 33 (01): : 17 - 27
  • [40] Autonomous Navigation and Obstacle Avoidance for Small VTOL UAV in Unknown Environments
    Chen, Cheng
    Wang, Zian
    Gong, Zheng
    Cai, Pengcheng
    Zhang, Chengxi
    Li, Yi
    [J]. SYMMETRY-BASEL, 2022, 14 (12):