A Deep Reinforcement Learning Method for Collision Avoidance with Dense Speed-Constrained Multi-UAV

被引:0
作者
Han, Jiale [1 ]
Zhu, Yi [1 ]
Yang, Jian [1 ]
机构
[1] South China Univ Technol, Sch Automat Sci & Engn, Guangzhou 510640, Peoples R China
来源
IEEE ROBOTICS AND AUTOMATION LETTERS | 2025年 / 10卷 / 03期
基金
中国国家自然科学基金;
关键词
Collision avoidance; Autonomous aerial vehicles; Feature extraction; Safety; Recurrent neural networks; Deep reinforcement learning; Vectors; Turning; Training; Predictive models; reinforcement learning; autonomous aerial vehicles; soft actor-critic;
D O I
10.1109/LRA.2025.3527292
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
This letter introduces a novel deep reinforcement learning (DRL) method for collision avoidance problem of fixed-wing unmanned aerial vehicles (UAVs). First, with considering the characteristics of collision avoidance problem, a collision prediction method is proposed to identify the neighboring UAVs with a significant threat. A convolutional neural network model is devised to extract the dynamic environment features. Second, a trajectory tracking macro action is incorporated into the action space of the proposed DRL-based algorithm. Guided by the reward function that considers to reward for closing to the preset flight paths, UAVs could return to the preset flight path after completing the collision avoidance. The proposed method is trained in simulation scenarios, with model updates implemented using a soft actor-critic (SAC) algorithm. Validation experiments are conducted in several complex multi-UAV flight environments. The results demonstrate the superiority of our method over other advanced methods.
引用
收藏
页码:2152 / 2159
页数:8
相关论文
共 24 条
[1]   Interaction ruling animal collective behavior depends on topological rather than metric distance: Evidence from a field study [J].
Ballerini, M. ;
Calbibbo, N. ;
Candeleir, R. ;
Cavagna, A. ;
Cisbani, E. ;
Giardina, I. ;
Lecomte, V. ;
Orlandi, A. ;
Parisi, G. ;
Procaccini, A. ;
Viale, M. ;
Zdravkovic, V. .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2008, 105 (04) :1232-1237
[2]  
Beard R. W., 2012, Small unmanned aircraft:Theory and practice
[3]   Constant speed optimal reciprocal collision avoidance [J].
Durand, Nicolas .
TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2018, 96 :366-379
[4]  
Everett M, 2018, IEEE INT C INT ROBOT, P3052, DOI 10.1109/IROS.2018.8593871
[5]   A survey of safety separation management and collision avoidance approaches of civil UAS operating in integration national airspace system [J].
Guan, Xiangmin ;
Lyu, Renli ;
Shi, Hongxia ;
Chen, Jun .
CHINESE JOURNAL OF AERONAUTICS, 2020, 33 (11) :2851-2863
[6]  
Haarnoja T, 2018, PR MACH LEARN RES, V80
[7]   Reinforcement learning control of constrained dynamic systems with uniformly ultimate boundedness stability guarantee [J].
Han, Minghao ;
Tian, Yuan ;
Zhang, Lixian ;
Wang, Jun ;
Pan, Wei .
AUTOMATICA, 2021, 129
[8]   Reinforcement Learned Distributed Multi-Robot Navigation With Reciprocal Velocity Obstacle Shaped Rewards [J].
Han, Ruihua ;
Chen, Shengduo ;
Wang, Shuaijun ;
Zhang, Zeqing ;
Gao, Rui ;
Hao, Qi ;
Pan, Jia .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (03) :5896-5903
[9]  
Han RH, 2020, IEEE INT CONF ROBOT, P448, DOI [10.1109/icra40945.2020.9197209, 10.1109/ICRA40945.2020.9197209]
[10]   Reinforcement Learning-Based Collision Avoidance and Optimal Trajectory Planning in UAV Communication Networks [J].
Hsu, Yu-Hsin ;
Gau, Rung-Hung .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2022, 21 (01) :306-320