Relative Distributed Formation and Obstacle Avoidance with Multi-agent Reinforcement Learning

被引:12
作者
Yan, Yuzi [1 ]
Li, Xiaoxiang [1 ]
Qiu, Xinyou [1 ]
Qiu, Jiantao [2 ,3 ]
Wang, Jian [1 ]
Wang, Yu [1 ]
Shen, Yuan [1 ]
机构
[1] Tsinghua Univ, Dept Elect Engn, Beijing, Peoples R China
[2] Tsinghua Univ, Beijing, Peoples R China
[3] Shanghai AI Lab, Shanghai, Peoples R China
来源
2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022) | 2022年
基金
国家重点研发计划;
关键词
DYNAMIC-MODEL; SYSTEMS;
D O I
10.1109/ICRA46639.2022.9812263
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-agent formation as well as obstacle avoidance is one of the most actively studied topics in the field of multi-agent systems. Although some classic controllers like model predictive control (MPC) and fuzzy control achieve a certain measure of success, most of them require precise global information which is not accessible in harsh environments. On the other hand, some reinforcement learning (RL) based approaches adopt the leader-follower structure to organize different agents' behaviors, which sacrifices the collaboration between agents thus suffering from bottlenecks in maneuverability and robustness. In this paper, we propose a distributed formation and obstacle avoidance method based on multi-agent reinforcement learning (MARL). Agents in our system only utilize local and relative information to make decisions and control themselves distributively, and will reorganize themselves into a new topology quickly in case that any of them is disconnected. Our method achieves better performance regarding formation error, formation convergence rate and on-par success rate of obstacle avoidance compared with baselines (both classic control methods and another RL-based method). The feasibility of our method is verified by both simulation and hardware implementation with Ackermann-steering vehicles.
引用
收藏
页码:1661 / 1667
页数:7
相关论文
共 50 条
  • [1] Multi-Agent Deep Reinforcement Learning for Distributed Load Restoration
    Linh Vu
    Tuyen Vu
    Thanh Long Vu
    Srivastava, Anurag
    IEEE TRANSACTIONS ON SMART GRID, 2024, 15 (02) : 1749 - 1760
  • [2] A Nearly Optimal Multi-agent Formation Control with Reinforcement Learning
    Peng, Jiangwen
    Mu, Chaoxu
    Wang, Ke
    2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 5315 - 5320
  • [3] Deep Reinforcement Learning Agent for Negotiation in Multi-Agent Cooperative Distributed Predictive Control
    Aponte-Rengifo, Oscar
    Vega, Pastora
    Francisco, Mario
    APPLIED SCIENCES-BASEL, 2023, 13 (04):
  • [4] Swarm Multi-agent Trapping Multi-target Control with Obstacle Avoidance
    Li, Chenyang
    Jiang, Guanjie
    Yang, Yonghui
    Chen, XueBo
    ADVANCES IN SWARM INTELLIGENCE, ICSI 2023, PT II, 2023, 13969 : 49 - 61
  • [5] Multi-Agent Reinforcement Learning-Based Distributed Dynamic Spectrum Access
    Albinsaid, Hasan
    Singh, Keshav
    Biswas, Sudip
    Li, Chih-Peng
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2022, 8 (02) : 1174 - 1185
  • [6] A Distributed Multi-Agent Dynamic Area Coverage Algorithm Based on Reinforcement Learning
    Xiao, Jian
    Wang, Gang
    Zhang, Ying
    Cheng, Lei
    IEEE ACCESS, 2020, 8 : 33511 - 33521
  • [7] Adaptive Multi-Agent Control with Dynamic Obstacle Avoidance in a Limited Region
    Bai, Yang
    Wang, Yujie
    Xiong, Xiaogang
    Svinin, Mikhail
    Magid, Evgeni
    2022 AMERICAN CONTROL CONFERENCE, ACC, 2022, : 4695 - 4700
  • [8] Multi-agent deep reinforcement learning: a survey
    Gronauer, Sven
    Diepold, Klaus
    ARTIFICIAL INTELLIGENCE REVIEW, 2022, 55 (02) : 895 - 943
  • [9] Multi-agent reinforcement learning with weak ties☆
    Wang, Huan
    Zhou, Xu
    Kang, Yu
    Xue, Jian
    Yang, Chenguang
    Liu, Xiaofeng
    INFORMATION FUSION, 2025, 118
  • [10] Goal-Based Multi-Agent Tree Formation using Reinforcement Learning
    Prasad, B. K. Swathi
    Ramasangu, Hariharan
    2017 INTERNATIONAL CONFERENCE ON ADVANCES IN COMPUTING, COMMUNICATIONS AND INFORMATICS (ICACCI), 2017, : 2180 - 2184