Relative Distributed Formation and Obstacle Avoidance with Multi-agent Reinforcement Learning

被引:12
作者
Yan, Yuzi [1 ]
Li, Xiaoxiang [1 ]
Qiu, Xinyou [1 ]
Qiu, Jiantao [2 ,3 ]
Wang, Jian [1 ]
Wang, Yu [1 ]
Shen, Yuan [1 ]
机构
[1] Tsinghua Univ, Dept Elect Engn, Beijing, Peoples R China
[2] Tsinghua Univ, Beijing, Peoples R China
[3] Shanghai AI Lab, Shanghai, Peoples R China
来源
2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022) | 2022年
基金
国家重点研发计划;
关键词
DYNAMIC-MODEL; SYSTEMS;
D O I
10.1109/ICRA46639.2022.9812263
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-agent formation as well as obstacle avoidance is one of the most actively studied topics in the field of multi-agent systems. Although some classic controllers like model predictive control (MPC) and fuzzy control achieve a certain measure of success, most of them require precise global information which is not accessible in harsh environments. On the other hand, some reinforcement learning (RL) based approaches adopt the leader-follower structure to organize different agents' behaviors, which sacrifices the collaboration between agents thus suffering from bottlenecks in maneuverability and robustness. In this paper, we propose a distributed formation and obstacle avoidance method based on multi-agent reinforcement learning (MARL). Agents in our system only utilize local and relative information to make decisions and control themselves distributively, and will reorganize themselves into a new topology quickly in case that any of them is disconnected. Our method achieves better performance regarding formation error, formation convergence rate and on-par success rate of obstacle avoidance compared with baselines (both classic control methods and another RL-based method). The feasibility of our method is verified by both simulation and hardware implementation with Ackermann-steering vehicles.
引用
收藏
页码:1661 / 1667
页数:7
相关论文
共 50 条
[21]   Distributed Optimization for Distribution Grids With Stochastic DER Using Multi-Agent Deep Reinforcement Learning [J].
Al-Saffar, Mohammed ;
Musilek, Petr .
IEEE ACCESS, 2021, 9 :63059-63072
[22]   A Novel Multi-agent Formation Control Law With Collision Avoidance [J].
Mondal, Arindam ;
Behera, Laxmidhar ;
Sahoo, Soumya Ranjan ;
Shukla, Anupam .
IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2017, 4 (03) :558-568
[23]   Distributed multi-agent collision avoidance using robust differential game [J].
Xue, Wenyan ;
Zhan, Siyuan ;
Wu, Zhihong ;
Chen, Yutao ;
Huang, Jie .
ISA TRANSACTIONS, 2023, 134 :95-107
[24]   Obstacle-Avoidance Distributed Reinforcement Learning Optimal Control for Spacecraft Cluster Flight [J].
Meng, Yizhen ;
Liu, Chun ;
Zhao, Jing ;
Huang, Jing .
IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS, 2025, 61 (01) :443-456
[25]   Learning multi-agent communication with double attentional deep reinforcement learning [J].
Mao, Hangyu ;
Zhang, Zhengchao ;
Xiao, Zhen ;
Gong, Zhibo ;
Ni, Yan .
AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2020, 34 (01)
[26]   Smart Grid for Industry Using Multi-Agent Reinforcement Learning [J].
Roesch, Martin ;
Linder, Christian ;
Zimmermann, Roland ;
Rudolf, Andreas ;
Hohmann, Andrea ;
Reinhart, Gunther .
APPLIED SCIENCES-BASEL, 2020, 10 (19) :1-20
[27]   Classes of Dilemma Problems and Their Multi-Agent Reinforcement Learning Method [J].
Kuroe, Yasuaki ;
Iima, Hitoshi .
IEEE ACCESS, 2024, 12 :107353-107367
[28]   Multi-Agent Reinforcement Learning with Optimal Equivalent Action of Neighborhood [J].
Wang, Haixing ;
Yang, Yi ;
Lin, Zhiwei ;
Wang, Tian .
ACTUATORS, 2022, 11 (04)
[29]   Multi-agent reinforcement learning in a new transactive energy mechanism [J].
Mohsenzadeh-Yazdi, Hossein ;
Kebriaei, Hamed ;
Aminifar, Farrokh .
IET GENERATION TRANSMISSION & DISTRIBUTION, 2024, 18 (18) :2943-2955
[30]   Optimal control in microgrid using multi-agent reinforcement learning [J].
Li, Fu-Dong ;
Wu, Min ;
He, Yong ;
Chen, Xin .
ISA TRANSACTIONS, 2012, 51 (06) :743-751