Obstacle Avoidance in Multi-Agent Formation Process Based on Deep Reinforcement Learning

被引:5
作者
Ji X. [1 ]
Hai J. [1 ]
Luo W. [1 ]
Lin C. [1 ]
Xiong Y. [3 ]
Ou Z. [3 ]
Wen J. [1 ]
机构
[1] School of Electrical and Information Engineering, Guangxi University of Science and Technology, Liuzhou, Guangxi
[2] Technology Center of Dongfeng Liuzhou Automobile Co., Ltd., Liuzhou, Guangxi
基金
中国国家自然科学基金;
关键词
A; deep reinforcement learning (DRL); formation; multi-agent; O; 231.5; obstacle avoidance; wheelbarrow;
D O I
10.1007/s12204-021-2357-6
中图分类号
学科分类号
摘要
To solve the problems of difficult control law design, poor portability, and poor stability of traditional multi-agent formation obstacle avoidance algorithms, a multi-agent formation obstacle avoidance method based on deep reinforcement learning (DRL) is proposed. This method combines the perception ability of convolutional neural networks (CNNs) with the decision-making ability of reinforcement learning in a general form and realizes direct output control from the visual perception input of the environment to the action through an end-to-end learning method. The multi-agent system (MAS) model of the follow-leader formation method was designed with the wheelbarrow as the control object. An improved deep Q netwrok (DQN) algorithm (we improved its discount factor and learning efficiency and designed a reward value function that considers the distance relationship between the agent and the obstacle and the coordination factor between the multi-agents) was designed to achieve obstacle avoidance and collision avoidance in the process of multi-agent formation into the desired formation. The simulation results show that the proposed method achieves the expected goal of multi-agent formation obstacle avoidance and has stronger portability compared with the traditional algorithm. © 2021, Shanghai Jiao Tong University and Springer-Verlag GmbH Germany, part of Springer Nature.
引用
收藏
页码:680 / 685
页数:5
相关论文
共 13 条
[1]  
Xie G., Zhang Y., Survey of consensus problem in cooperative control of multi-agent systems, Application Research of Computers, 28, 6, pp. 2035-2039, (2011)
[2]  
Chen Z., Lin L., Yan G., An approach to scientific cooperative robotics: Through MAS (multi-agent system), Robot, 23, 4, pp. 368-373, (2001)
[3]  
Duan Y., Yang H., Cui B., Et al., Application of reinforcement learning to basic action learning of soccer robot, Robot, 30, 5, pp. 453-459, (2008)
[4]  
Littman M.L., Reinforcement learning improves behaviour from evaluative feedback, Nature, 521, 7553, pp. 445-451, (2015)
[5]  
Zhu Y., Zhao D., Probably approximately correct reinforcement learning solving continuous-state control problem, Control Theory and Applications, 33, 12, pp. 1603-1613, (2016)
[6]  
Zhou W., The application of deep learning algorithms in intelligent collaborative robots, China New Telecommunications, 19, 21, pp. 129-130, (2017)
[7]  
Polydoros A.S., Nalpantidis L., Survey of model-based reinforcement learning: Applications on robotics, Journal of Intelligent & Robotic Systems, 86, 2, pp. 153-173, (2017)
[8]  
Lima H., Kuroe Y., Swarm reinforcement learning methods improving certainty of learning for amulti-robot formation problem, pp. 3026-3033, (2015)
[9]  
Liu Q., Zhai J., Zhang Z., Et al., A survey on deep reinforcement learning, Chinese Journal of Computers, 41, 1, pp. 1-27, (2018)
[10]  
Riedmiller M., Neural fitted Q iteration: First experiences with a data efficient neural reinforcement learning method, pp. 317-328, (2005)