MACNS: A generic graph neural network integrated deep reinforcement learning based multi-agent collaborative navigation system for dynamic trajectory planning

被引:10
作者
Xiao, Ziren [1 ]
Li, Peisong [1 ]
Liu, Chang [1 ]
Gao, Honghao [2 ]
Wang, Xinheng [1 ]
机构
[1] Xian Jiaotong Liverpool Univ, Sch Adv Technol, Suzhou, Peoples R China
[2] Shanghai Univ, Sch Comp Engn & Sci, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
Graph proximal policy optimisation; Graph neural network; Multi-agent navigation;
D O I
10.1016/j.inffus.2024.102250
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi -agent collaborative navigation is prevalent in modern transportation systems, including delivery logistics, warehouse automation, and personalised tourism, where multiple moving agents (e.g., robots and people) must meet at a common destination from different starting points. However, traditional methods face challenges in efficiently and accurately optimising routes for multiple moving agents (e.g., robots). These include dynamically adjusting the common destination in response to changing traffic conditions, encoding the realworld map, and the training speed. Therefore, we propose a generic Multi -Agent Collaborative Navigation System (MACNS) to address those challenges. First, we formalise the solution of the problem and challenges into a Markov Decision Process (MDP), which is further developed as a training environment where a Deep Reinforcement Learning (DRL) agent can learn patterns efficiently. Second, the proposed framework integrates a Graph Neural Network (GNN) into the policy network of Proximal Policy Optimisation for the homogeneous decision -making of each individual agent, showing excellent generalisation, extensibility and convergence speed. Finally, we demonstrate how MACNS can be applied and implemented in a real -world use case: tour groups gathering and picking up. Extensive simulations and real -world tests validate the effectiveness of the MACNS-based use case, showcasing its superiority over other state-of-the-art PPO-related methods.
引用
收藏
页数:10
相关论文
共 39 条
[1]   Multi-Agent Deep Reinforcement Learning to Manage Connected Autonomous Vehicles at Tomorrow's Intersections [J].
Antonio, Guillen-Perez ;
Maria-Dolores, Cano .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2022, 71 (07) :7033-7043
[2]   Survey of Deep Reinforcement Learning for Motion Planning of Autonomous Vehicles [J].
Aradi, Szilard .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (02) :740-759
[3]   A Framework for Real-World Multi-Robot Systems Running Decentralized GNN-Based Policies [J].
Blumenkamp, Jan ;
Morad, Steven ;
Gielis, Jennifer ;
Li, Qingbiao ;
Prorok, Amanda .
2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA 2022, 2022, :8772-8778
[4]   Cooperative Path Planning of Multi-Agent Based on Graph Neural Network [J].
Dai, Hanqi ;
Lu, Weining ;
Yang, Jun ;
Liang, Bin .
2022 34TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2022, :5620-5624
[5]   Hierarchical path planner for unknown space exploration using reinforcement learning-based intelligent frontier selection [J].
Fan, Jie ;
Zhang, Xudong ;
Zou, Yuan .
EXPERT SYSTEMS WITH APPLICATIONS, 2023, 230
[6]   ALGORITHM-97 - SHORTEST PATH [J].
FLOYD, RW .
COMMUNICATIONS OF THE ACM, 1962, 5 (06) :345-345
[7]   A learning approach for multi-agent travelling problem with dynamic service requirement in mobile IoT [J].
Fu, Chenchen ;
Gao, Zhengxuan ;
Wu, Weiwei ;
Chau, Vincent ;
Wang, Jie ;
Xu, Xueyong ;
Luo, Junzhou .
COMPUTERS & ELECTRICAL ENGINEERING, 2022, 104
[8]  
Gong SZ, 2023, Arxiv, DOI arXiv:2307.12639
[9]   Asynchronous Multithreading Reinforcement-Learning-Based Path Planning and Tracking for Unmanned Underwater Vehicle [J].
He, Zichen ;
Dong, Lu ;
Sun, Changyin ;
Wang, Jiawei .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2022, 52 (05) :2757-2769
[10]  
Huang S., 2022, INT FLAIRS C P U FLO, P35, DOI DOI 10.32473/FLAIRS.V35I.130584