Distributed and Scalable Cooperative Formation of Unmanned Ground Vehicles Using Deep Reinforcement Learning

被引:2
|
作者
Huang, Shichun [1 ]
Wang, Tao [1 ,2 ,3 ]
Tang, Yong [4 ,5 ]
Hu, Yiwen [6 ]
Xin, Gu [7 ]
Zhou, Dianle [8 ]
机构
[1] Sun Yat sen Univ, Sch Intelligent Syst Engn, Guangzhou 510275, Peoples R China
[2] Southern Marine Sci & Engn Guangdong Lab Zhuhai, Zhuhai 519000, Peoples R China
[3] Guangdong Prov Key Lab Fire Sci & Intelligent Emer, Guangzhou 510006, Peoples R China
[4] Northwestern Polytech Univ, Sch Civil Aviat, Xian 710072, Peoples R China
[5] UAS Co Ltd, Aviat Ind Corp China Chengdu, Chengdu 610091, Peoples R China
[6] AVIC Chengdu Aircraft Design & Res Inst, Chengdu 610041, Peoples R China
[7] China Acad Launch Vehicle Technol, Dept Res & Dev Ctr, Beijing 100076, Peoples R China
[8] Natl Univ Def Technol, Coll Adv Interdisciplinary Studies, Changsha 410073, Peoples R China
基金
中国国家自然科学基金;
关键词
unmanned ground vehicles (UGVs); deep reinforcement learning; deep deterministic policy gradient (DDPG); multiagent systems; distributed formation control; MOBILE ROBOT;
D O I
10.3390/aerospace10020096
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
Cooperative formation control of unmanned ground vehicles (UGVs) has become one of the important research hotspots in the application of UGV and attracted more and more attention in the military and civil fields. Compared with traditional formation control algorithms, reinforcement-learning-based algorithms can provide a new solution with a lower complexity for real-time formation control by equipping UGVs with artificial intelligence. Therefore, in this paper, a distributed deep-reinforcement-learning-based cooperative formation control algorithm is proposed to solve the navigation, maintenance, and obstacle avoidance tasks of UGV formations. More importantly, the hierarchical triangular formation structure and the newly designed Markov decision process for UGV formations of leader and follower attributes make the control strategy learned by the algorithm reusable, so that the formation can arbitrarily increase the number of UGVs and realize a more flexible expansion. The effectiveness and scalability of the algorithm is verified by formation simulation experiments of different scales.
引用
收藏
页数:20
相关论文
共 50 条
  • [41] Distributed formation control with obstacle avoidance for multiple underactuated unmanned surface vehicles
    Tang, Xiangyu
    Yu, Jianglong
    Li, Xiaoduo
    Dong, Xiwang
    Ren, Zhang
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2024, 361 (12):
  • [42] Using Deep Reinforcement Learning to Automate Network Configurations for Internet of Vehicles
    Liu, Xing
    Qian, Cheng
    Yu, Wei
    Griffith, David
    Gopstein, Avi
    Golmie, Nada
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (12) : 15948 - 15958
  • [43] Obstacle avoidance planning of autonomous vehicles using deep reinforcement learning
    Qian, Yubin
    Feng, Song
    Hu, Wenhao
    Wang, Wanqiu
    ADVANCES IN MECHANICAL ENGINEERING, 2022, 14 (12)
  • [44] Speed and heading control of an unmanned surface vehicle using deep reinforcement learning
    Wu, Ting
    Ye, Hui
    Xiang, Zhengrong
    Yang, Xiaofei
    2023 IEEE 12TH DATA DRIVEN CONTROL AND LEARNING SYSTEMS CONFERENCE, DDCLS, 2023, : 573 - 578
  • [45] Double Critic Deep Reinforcement Learning for Mapless 3D Navigation of Unmanned Aerial Vehicles
    Bedin Grando, Ricardo
    de Jesus, Junior Costa
    Kich, Victor Augusto
    Kolling, Alisson Henrique
    Jorge Drews-Jr, Paulo Lilles
    JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2022, 104 (02)
  • [46] Target tracking control for Unmanned Surface Vehicles: An end-to-end deep reinforcement learning approach
    Wang, Zihao
    Hu, Qiyuan
    Wang, Chao
    Liu, Yi
    Xie, Wenbo
    OCEAN ENGINEERING, 2025, 317
  • [47] Deep reinforcement learning to control an unmanned swarm system
    Liang, Hongtao
    Wang, Yaonan
    Hua, Hean
    Zhong, Hang
    Zheng, Chenghong
    Zeng, Junhao
    Liang, Jiacheng
    Li, Zhengchen
    Gongcheng Kexue Xuebao/Chinese Journal of Engineering, 2024, 46 (09): : 1521 - 1534
  • [48] Distributed deep reinforcement learning method using profit sharing for learning acceleration
    Kodama, Naoki
    Harada, Taku
    Miyazaki, Kazuteru
    IEEJ TRANSACTIONS ON ELECTRICAL AND ELECTRONIC ENGINEERING, 2020, 15 (08) : 1188 - 1196
  • [49] Aerial and Ground Vehicles Collaboration for Automated Target Tracking Using Reinforcement Learning
    Zanone, R. Oliver
    Velni, Javad Mohammadpour
    IFAC PAPERSONLINE, 2024, 58 (28): : 456 - 461
  • [50] Intelligent Autonomous Navigation of Car-Like Unmanned Ground Vehicle via Deep Reinforcement Learning
    Sivashangaran, Shathushan
    Zheng, Minghui
    IFAC PAPERSONLINE, 2021, 54 (20): : 218 - 225