Hierarchical Relational Graph Learning for Autonomous Multirobot Cooperative Navigation in Dynamic Environments

被引:1
|
作者
Wang, Ting [1 ]
Du, Xiao [1 ]
Chen, Mingsong [1 ]
Li, Keqin [2 ]
机构
[1] East China Normal Univ, MoE Engn Res Ctr Software Hardware Codesign Techno, Shanghai 200241, Peoples R China
[2] SUNY, Dept Comp Sci, New York, NY 12561 USA
关键词
Cyber-physical systems (CPSs); deep reinforcement learning (DRL); multirobot cooperation; robot navigation; DESIGN AUTOMATION; REINFORCEMENT; FRAMEWORK; MODEL;
D O I
10.1109/TCAD.2023.3260710
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
As a specific kind of cyber-physical systems (CPSs), autonomous robot clusters play an important role in various intelligent manufacturing fields. However, due to the increasing design complexity of robot clusters, it is becoming more and more challenging to guarantee the safety and efficiency for multirobot cooperative navigation in dynamic and complex environments. Although deep reinforcement learning (DRL) shows great potential in learning multirobot cooperative navigation policies, existing DRL-based approaches suffer from scalability issues and rarely consider the transferability of trained policies to new tasks. To address these problems, this article presents a novel DRL-based multirobot cooperative navigation approach named HRMR-Navi that equips each robot with both a two-layered hierarchical graph network model and an attention-based communication model. In our approach, the hierarchical graph network model can efficiently figure out hierarchical relations among all agents that either cooperate for efficiency or avoid obstacles for safety to derive more advanced strategies, and the communication model can accurately form a global view of the environment for a specific robot, thus, the multirobot cooperation efficiency can be further strengthened. Meanwhile, we propose an improved proximal policy optimization (PPO) algorithm based on the Maximum Entropy Reinforcement Learning, named MEPPO, to enhance the robot exploration ability. Comprehensive experimental results demonstrate that, compared with state-of-the-art approaches, HRMR-Navi can achieve more efficient cooperative navigation with less time cost, lower collision rate, higher scalability, and better knowledge transferability.
引用
收藏
页码:3559 / 3570
页数:12
相关论文
共 50 条
  • [1] Autonomous Multirobot Navigation and Cooperative Mapping in Partially Unknown Environments
    Xie, Hongyu
    Zhang, Dong
    Hu, Xiaobo
    Zhou, MengChu
    Cao, Zhengcai
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72 : 1 - 12
  • [2] Visuomotor Reinforcement Learning for Multirobot Cooperative Navigation
    Liu, Zhe
    Liu, Qiming
    Tang, Ling
    Jin, Kefan
    Wang, Hongye
    Liu, Ming
    Wang, Hesheng
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2022, 19 (04) : 3234 - 3245
  • [3] Cooperative Deep Reinforcement Learning Policies for Autonomous Navigation in Complex Environments
    Tran, Van Manh
    Kim, Gon-Woo
    IEEE ACCESS, 2024, 12 : 101053 - 101065
  • [4] Deep Imitation Learning for Autonomous Navigation in Dynamic Pedestrian Environments
    Qin, Lei
    Huang, Zefan
    Zhang, Chen
    Guo, Hongliang
    Ang, Marcelo, Jr.
    Rus, Daniela
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 4108 - 4115
  • [5] Relational Graph Learning for Crowd Navigation
    Chen, Changan
    Hu, Sha
    Nikdel, Payam
    Mori, Greg
    Savva, Manolis
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 10007 - 10013
  • [6] Hierarchical Reinforcement Learning for Dynamic Autonomous Vehicle Navigation at Intelligent Intersections
    Sun, Qian
    Zhang, Le
    Yu, Huan
    Zhang, Weijia
    Mei, Yu
    Xiong, Hui
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 4852 - 4861
  • [7] Learning in Dynamic Environments with Ensemble Selection for Autonomous Outdoor Robot Navigation
    Procopio, Michael J.
    Mulligan, Jane
    Grudic, Greg
    2008 IEEE/RSJ INTERNATIONAL CONFERENCE ON ROBOTS AND INTELLIGENT SYSTEMS, VOLS 1-3, CONFERENCE PROCEEDINGS, 2008, : 620 - 627
  • [8] Autonomous Navigation of Quadrotors in Dynamic Complex Environments
    Li, Ruocheng
    Xin, Bin
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2025, 72 (03) : 2790 - 2800
  • [9] Intervention Force-based Imitation Learning for Autonomous Navigation in Dynamic Environments
    Yokoyama, Tomoya
    Seiya, Shunya
    Takeuchi, Eijiro
    Takeda, Kazuya
    2020 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2020, : 1679 - 1688
  • [10] Graph Relational Reinforcement Learning for Mobile Robot Navigation in Large-Scale Crowded Environments
    Liu, Zhe
    Zhai, Yu
    Li, Jiaming
    Wang, Guangming
    Miao, Yanzi
    Wang, Hesheng
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (08) : 8776 - 8787