GCMA: An Adaptive Multiagent Reinforcement Learning Framework With Group Communication for Complex and Similar Tasks Coordination

被引:1
|
作者
Peng, Kexing [1 ]
Ma, Tinghuai [2 ,3 ]
Yu, Xin [2 ]
Rong, Huan [4 ]
Qian, Yurong [5 ]
Al-Nabhan, Najla [6 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Comp Sci, Nanjing 210044, Peoples R China
[2] Nanjing Univ Informat Sci & Technol, Sch Software, Nanjing 210044, Peoples R China
[3] Jiangsu Ocean Univ, Lianyungang 222005, Peoples R China
[4] Nanjing Univ Informat Sci & Technol, Sch Artificial Intelligence, Nanjing 210044, Peoples R China
[5] Xinjiang Univ, Urumqi 830008, Peoples R China
[6] King Saud Univ, Dept Comp Sci, Riyadh 11362, Saudi Arabia
基金
中国国家自然科学基金;
关键词
Task analysis; Training; Games; Reinforcement learning; Resource management; Redundancy; Graph neural networks; Complex task policy; group communication; multiagent reinforcement learning (MARL); multitasks;
D O I
10.1109/TG.2023.3346394
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Coordinating multiple agents with diverse tasks and changing goals without interference is a challenge. Multiagent reinforcement learning (MARL) aims to develop effective communication and joint policies using group learning. Some of the previous approaches required each agent to maintain a set of networks independently, resulting in no consideration of interactions. Joint communication work causes agents receiving information unrelated to their own tasks. Currently, agents with different task divisions are often grouped by action tendency, but this can lead to poor dynamic grouping. This article presents a two-phase solution for multiple agents, addressing these issues. The first phase develops heterogeneous agent communication joint policies using a group communication MARL framework (GCMA). The framework employs a periodic grouping strategy, reducing exploration and communication redundancy by dynamically assigning agent group hidden features through hypernetwork and graph communication. The scheme efficiently utilizes resources for adapting to multiple similar tasks. In the second phase, each agent's policy network is distilled into a generalized simple network, adapting to similar tasks with varying quantities and sizes. GCMA is tested in complex environments, such as StarCraft II and unmanned aerial vehicle (UAV) take-off, showing its well-performing for large-scale, coordinated tasks. It shows GCMA's effectiveness for solid generalization in multitask tests with simulated pedestrians.
引用
收藏
页码:670 / 682
页数:13
相关论文
共 50 条
  • [1] A Decentralized Communication Framework Based on Dual-Level Recurrence for Multiagent Reinforcement Learning
    Li, Xuesi
    Li, Jingchen
    Shi, Haobin
    Hwang, Kao-Shing
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2024, 16 (02) : 640 - 649
  • [2] Adaptive Individual Q-Learning-A Multiagent Reinforcement Learning Method for Coordination Optimization
    Zhang, Zhen
    Wang, Dongqing
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 12
  • [3] Adaptive Learning: A New Decentralized Reinforcement Learning Approach for Cooperative Multiagent Systems
    Li, Meng-Lin
    Chen, Shaofei
    Chen, Jing
    IEEE ACCESS, 2020, 8 : 99404 - 99421
  • [4] Learning Automata-Based Multiagent Reinforcement Learning for Optimization of Cooperative Tasks
    Zhang, Zhen
    Wang, Dongqing
    Gao, Junwei
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (10) : 4639 - 4652
  • [5] Inferring Latent Temporal Sparse Coordination Graph for Multiagent Reinforcement Learning
    Duan, Wei
    Lu, Jie
    Xuan, Junyu
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [6] Multiagent Meta-Reinforcement Learning for Adaptive Multipath Routing Optimization
    Chen, Long
    Hu, Bin
    Guan, Zhi-Hong
    Zhao, Lian
    Shen, Xuemin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (10) : 5374 - 5386
  • [7] Coordination of Electric Vehicle Charging Through Multiagent Reinforcement Learning
    Da Silva, Felipe Leno
    Nishida, Cyntia E. H.
    Roijers, Diederik M.
    Costa, Anna H. Reali
    IEEE TRANSACTIONS ON SMART GRID, 2020, 11 (03) : 2347 - 2356
  • [8] Optimal Group Consensus of Multiagent Systems in Graphical Games Using Reinforcement Learning
    Wang, Yuhan
    Wang, Zhuping
    Zhang, Hao
    Yan, Huaicheng
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2025, 55 (03): : 2343 - 2353
  • [9] Coordination of Communication in Robot Teams by Reinforcement Learning
    Maravall, Dario
    de Lope, Javier
    Dominguez, Raul
    FOUNDATIONS ON NATURAL AND ARTIFICIAL COMPUTATION: 4TH INTERNATIONAL WORK-CONFERENCE ON THE INTERPLAY BETWEEN NATURAL AND ARTIFICIAL COMPUTATION, IWINAC 2011, PART I, 2011, 6686 : 156 - 164
  • [10] Coordination of communication in robot teams by reinforcement learning
    Maravall, Dario
    de Lope, Javier
    Dominguez, Raul
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2013, 61 (07) : 661 - 666