Bio-Inspired Collision Avoidance in Swarm Systems via Deep Reinforcement Learning

被引:31
作者
Na, Seongin [1 ]
Niu, Hanlin [1 ]
Lennox, Barry [1 ]
Arvin, Farshad [1 ]
机构
[1] Univ Manchester, Sch Engn, Dept Elect & Elect Engn, Swarm & Computat Intelligence Lab SwaCIL, Manchester M13 9PL, Lancs, England
基金
英国工程与自然科学研究理事会; 欧盟地平线“2020”;
关键词
Collision avoidance; Autonomous vehicles; Communication networks; Training; Task analysis; Robot sensing systems; Servers; autonomous vehicles; multi-agent systems; deep reinforcement learning; swarm robotics; MEMORY; ANT;
D O I
10.1109/TVT.2022.3145346
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Autonomous vehicles have been highlighted as a major growth area for future transportation systems and the deployment of large numbers of these vehicles is expected when safety and legal challenges are overcome. To meet the necessary safety standards, effective collision avoidance technologies are required to ensure that the number of accidents are kept to a minimum. As large numbers of autonomous vehicles, operating together on roads, can be regarded as a swarm system, we propose a bio-inspired collision avoidance strategy using virtual pheromones; an approach that has evolved effectively in nature over many millions of years. Previous research using virtual pheromones showed the potential of pheromone-based systems to maneuver a swarm of robots. However, designing an individual controller to maximise the performance of the entire swarm is a major challenge. In this paper, we propose a novel deep reinforcement learning (DRL) based approach that is able to train a controller that introduces collision avoidance behaviour. To accelerate training, we propose a novel sampling strategy called Highlight Experience Replay and integrate it with a Deep Deterministic Policy Gradient algorithm with noise added to the weights and biases of the artificial neural network to improve exploration. To evaluate the performance of the proposed DRL-based controller, we applied it to navigation and collision avoidance tasks in three different traffic scenarios. The experimental results showed that the proposed DRL-based controller outperformed the manually-tuned controller in terms of stability, effectiveness, robustness and ease of tuning process. Furthermore, the proposed Highlight Experience Replay method outperformed than the popular Prioritized Experience Replay sampling strategy by taking 27% of training time average over three stages.
引用
收藏
页码:2511 / 2526
页数:16
相关论文
共 50 条
[21]   Space Manipulator Collision Avoidance Using a Deep Reinforcement Learning Control [J].
Blaise, James ;
Bazzocchi, Michael C. F. .
AEROSPACE, 2023, 10 (09)
[22]   Vehicles Control: Collision Avoidance using Federated Deep Reinforcement Learning [J].
Ben Elallid, Badr ;
Abouaomar, Amine ;
Benamar, Nabil ;
Kobbane, Abdellatif .
IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, :4369-4374
[23]   COLREG-Compliant Collision Avoidance for Unmanned Surface Vehicle Using Deep Reinforcement Learning [J].
Meyer, Eivind ;
Heiberg, Amalie ;
Rasheed, Adil ;
San, Omer .
IEEE ACCESS, 2020, 8 :165344-165364
[24]   Formation Control with Collision Avoidance through Deep Reinforcement Learning [J].
Sui, Zezhi ;
Pu, Zhiqiang ;
Yi, Jianqiang ;
Xiong, Tianyi .
2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
[25]   Ship Collision Avoidance Using Constrained Deep Reinforcement Learning [J].
Zhang, Rui ;
Wang, Xiao ;
Liu, Kezhong ;
Wu, Xiaolie ;
Lu, Tianyou ;
Chao Zhaohui .
2018 5TH INTERNATIONAL CONFERENCE ON BEHAVIORAL, ECONOMIC, AND SOCIO-CULTURAL COMPUTING (BESC), 2018, :115-120
[26]   A Comprehensive Review of Shepherding as a Bio-Inspired Swarm-Robotics Guidance Approach [J].
Long, Nathan K. ;
Sammut, Karl ;
Sgarioto, Daniel ;
Garratt, Matthew ;
Abbass, Hussein A. .
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2020, 4 (04) :523-537
[27]   Deep reinforcement learning with predictive auxiliary task for autonomous train collision avoidance [J].
Plissonneau, Antoine ;
Jourdan, Luca ;
Trentesaux, Damien ;
Abdi, Lotfi ;
Sallak, Mohamed ;
Bekrar, Abdelghani ;
Quost, Benjamin ;
Schoen, Walter .
JOURNAL OF RAIL TRANSPORT PLANNING & MANAGEMENT, 2024, 31
[28]   Deep Reinforcement Learning Based Collision Avoidance Algorithm for Differential Drive Robot [J].
Lu, Xinglong ;
Cao, Yiwen ;
Zhao, Zhonghua ;
Yan, Yilin .
INTELLIGENT ROBOTICS AND APPLICATIONS (ICIRA 2018), PT I, 2018, 10984 :186-198
[29]   A Bio-inspired Collision Detector for Small Quadcopter [J].
Zhao, Jiannan ;
Hu, Cheng ;
Zhang, Chun ;
Wang, Zhihua ;
Yue, Shigang .
2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
[30]   Bio-inspired artificial pheromone system for swarm robotics applications [J].
Na, Seongin ;
Qiu, Yiping ;
Turgut, Ali E. ;
Ulrich, Jiri ;
Krajnik, Tomas ;
Yue, Shigang ;
Lennox, Barry ;
Arvin, Farshad .
ADAPTIVE BEHAVIOR, 2021, 29 (04) :395-415