Bio-Inspired Collision Avoidance in Swarm Systems via Deep Reinforcement Learning

被引:24
作者
Na, Seongin [1 ]
Niu, Hanlin [1 ]
Lennox, Barry [1 ]
Arvin, Farshad [1 ]
机构
[1] Univ Manchester, Sch Engn, Dept Elect & Elect Engn, Swarm & Computat Intelligence Lab SwaCIL, Manchester M13 9PL, Lancs, England
基金
英国工程与自然科学研究理事会; 欧盟地平线“2020”;
关键词
Collision avoidance; Autonomous vehicles; Communication networks; Training; Task analysis; Robot sensing systems; Servers; autonomous vehicles; multi-agent systems; deep reinforcement learning; swarm robotics; MEMORY; ANT;
D O I
10.1109/TVT.2022.3145346
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Autonomous vehicles have been highlighted as a major growth area for future transportation systems and the deployment of large numbers of these vehicles is expected when safety and legal challenges are overcome. To meet the necessary safety standards, effective collision avoidance technologies are required to ensure that the number of accidents are kept to a minimum. As large numbers of autonomous vehicles, operating together on roads, can be regarded as a swarm system, we propose a bio-inspired collision avoidance strategy using virtual pheromones; an approach that has evolved effectively in nature over many millions of years. Previous research using virtual pheromones showed the potential of pheromone-based systems to maneuver a swarm of robots. However, designing an individual controller to maximise the performance of the entire swarm is a major challenge. In this paper, we propose a novel deep reinforcement learning (DRL) based approach that is able to train a controller that introduces collision avoidance behaviour. To accelerate training, we propose a novel sampling strategy called Highlight Experience Replay and integrate it with a Deep Deterministic Policy Gradient algorithm with noise added to the weights and biases of the artificial neural network to improve exploration. To evaluate the performance of the proposed DRL-based controller, we applied it to navigation and collision avoidance tasks in three different traffic scenarios. The experimental results showed that the proposed DRL-based controller outperformed the manually-tuned controller in terms of stability, effectiveness, robustness and ease of tuning process. Furthermore, the proposed Highlight Experience Replay method outperformed than the popular Prioritized Experience Replay sampling strategy by taking 27% of training time average over three stages.
引用
收藏
页码:2511 / 2526
页数:16
相关论文
共 50 条
  • [21] Vehicles Control: Collision Avoidance using Federated Deep Reinforcement Learning
    Ben Elallid, Badr
    Abouaomar, Amine
    Benamar, Nabil
    Kobbane, Abdellatif
    IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 4369 - 4374
  • [22] COLREG-Compliant Collision Avoidance for Unmanned Surface Vehicle Using Deep Reinforcement Learning
    Meyer, Eivind
    Heiberg, Amalie
    Rasheed, Adil
    San, Omer
    IEEE ACCESS, 2020, 8 (08): : 165344 - 165364
  • [23] Formation Control with Collision Avoidance through Deep Reinforcement Learning
    Sui, Zezhi
    Pu, Zhiqiang
    Yi, Jianqiang
    Xiong, Tianyi
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [24] Ship Collision Avoidance Using Constrained Deep Reinforcement Learning
    Zhang, Rui
    Wang, Xiao
    Liu, Kezhong
    Wu, Xiaolie
    Lu, Tianyou
    Chao Zhaohui
    2018 5TH INTERNATIONAL CONFERENCE ON BEHAVIORAL, ECONOMIC, AND SOCIO-CULTURAL COMPUTING (BESC), 2018, : 115 - 120
  • [25] Deep reinforcement learning with predictive auxiliary task for autonomous train collision avoidance
    Plissonneau, Antoine
    Jourdan, Luca
    Trentesaux, Damien
    Abdi, Lotfi
    Sallak, Mohamed
    Bekrar, Abdelghani
    Quost, Benjamin
    Schoen, Walter
    JOURNAL OF RAIL TRANSPORT PLANNING & MANAGEMENT, 2024, 31
  • [26] A Comprehensive Review of Shepherding as a Bio-Inspired Swarm-Robotics Guidance Approach
    Long, Nathan K.
    Sammut, Karl
    Sgarioto, Daniel
    Garratt, Matthew
    Abbass, Hussein A.
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2020, 4 (04): : 523 - 537
  • [27] Deep Reinforcement Learning Based Collision Avoidance Algorithm for Differential Drive Robot
    Lu, Xinglong
    Cao, Yiwen
    Zhao, Zhonghua
    Yan, Yilin
    INTELLIGENT ROBOTICS AND APPLICATIONS (ICIRA 2018), PT I, 2018, 10984 : 186 - 198
  • [28] A Bio-inspired Collision Detector for Small Quadcopter
    Zhao, Jiannan
    Hu, Cheng
    Zhang, Chun
    Wang, Zhihua
    Yue, Shigang
    2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [29] Bio-inspired artificial pheromone system for swarm robotics applications
    Na, Seongin
    Qiu, Yiping
    Turgut, Ali E.
    Ulrich, Jiri
    Krajnik, Tomas
    Yue, Shigang
    Lennox, Barry
    Arvin, Farshad
    ADAPTIVE BEHAVIOR, 2021, 29 (04) : 395 - 415
  • [30] Collision-avoidance under COLREGS for unmanned surface vehicles via deep reinforcement learning
    Ma, Yong
    Zhao, Yujiao
    Wang, Yulong
    Gan, Langxiong
    Zheng, Yuanzhou
    MARITIME POLICY & MANAGEMENT, 2020, 47 (05) : 665 - 686