GCMA: An Adaptive Multiagent Reinforcement Learning Framework With Group Communication for Complex and Similar Tasks Coordination

被引:1
|
作者
Peng, Kexing [1 ]
Ma, Tinghuai [2 ,3 ]
Yu, Xin [2 ]
Rong, Huan [4 ]
Qian, Yurong [5 ]
Al-Nabhan, Najla [6 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Comp Sci, Nanjing 210044, Peoples R China
[2] Nanjing Univ Informat Sci & Technol, Sch Software, Nanjing 210044, Peoples R China
[3] Jiangsu Ocean Univ, Lianyungang 222005, Peoples R China
[4] Nanjing Univ Informat Sci & Technol, Sch Artificial Intelligence, Nanjing 210044, Peoples R China
[5] Xinjiang Univ, Urumqi 830008, Peoples R China
[6] King Saud Univ, Dept Comp Sci, Riyadh 11362, Saudi Arabia
基金
中国国家自然科学基金;
关键词
Task analysis; Training; Games; Reinforcement learning; Resource management; Redundancy; Graph neural networks; Complex task policy; group communication; multiagent reinforcement learning (MARL); multitasks;
D O I
10.1109/TG.2023.3346394
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Coordinating multiple agents with diverse tasks and changing goals without interference is a challenge. Multiagent reinforcement learning (MARL) aims to develop effective communication and joint policies using group learning. Some of the previous approaches required each agent to maintain a set of networks independently, resulting in no consideration of interactions. Joint communication work causes agents receiving information unrelated to their own tasks. Currently, agents with different task divisions are often grouped by action tendency, but this can lead to poor dynamic grouping. This article presents a two-phase solution for multiple agents, addressing these issues. The first phase develops heterogeneous agent communication joint policies using a group communication MARL framework (GCMA). The framework employs a periodic grouping strategy, reducing exploration and communication redundancy by dynamically assigning agent group hidden features through hypernetwork and graph communication. The scheme efficiently utilizes resources for adapting to multiple similar tasks. In the second phase, each agent's policy network is distilled into a generalized simple network, adapting to similar tasks with varying quantities and sizes. GCMA is tested in complex environments, such as StarCraft II and unmanned aerial vehicle (UAV) take-off, showing its well-performing for large-scale, coordinated tasks. It shows GCMA's effectiveness for solid generalization in multitask tests with simulated pedestrians.
引用
收藏
页码:670 / 682
页数:13
相关论文
共 50 条
  • [21] Adaptive Multiagent Model Based on Reinforcement Learning for Distributed Generation Systems
    Divenyi, Daniel
    Dan, Andras
    2012 23RD INTERNATIONAL WORKSHOP ON DATABASE AND EXPERT SYSTEMS APPLICATIONS (DEXA), 2012, : 303 - 307
  • [22] Hybrid Trajectory and Force Learning of Complex Assembly Tasks: A Combined Learning Framework
    Wang, Yan
    Beltran-Hernandez, Cristian C.
    Wan, Weiwei
    Harada, Kensuke
    IEEE ACCESS, 2021, 9 : 60175 - 60186
  • [23] SON Coordination in Heterogeneous Networks: A Reinforcement Learning Framework
    Iacoboaiea, Ovidiu-Constantin
    Sayrac, Berna
    Ben Jemaa, Sana
    Bianchi, Pascal
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2016, 15 (09) : 5835 - 5847
  • [24] Reward shaping in multiagent reinforcement learning for self-organizing systems in assembly tasks
    Huang, Bingling
    Jin, Yan
    ADVANCED ENGINEERING INFORMATICS, 2022, 54
  • [25] ARL: An adaptive reinforcement learning framework for complex question answering over knowledge base
    Zhang, Qixuan
    Weng, Xinyi
    Zhou, Guangyou
    Zhang, Yi
    Huang, Jimmy Xiangji
    INFORMATION PROCESSING & MANAGEMENT, 2022, 59 (03)
  • [26] Fast Adaptive Task Offloading and Resource Allocation in Large-Scale MEC Systems via Multiagent Graph Reinforcement Learning
    Gao, Zhen
    Yang, Lei
    Dai, Yu
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (01) : 758 - 776
  • [27] Multirobot coordination with deep reinforcement learning in complex environments
    Wang, Di
    Deng, Hongbin
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 180
  • [28] Cascaded Attention: Adaptive and Gated Graph Attention Network for Multiagent Reinforcement Learning
    Qi, Shuhan
    Huang, Xinhao
    Peng, Peixi
    Huang, Xuzhong
    Zhang, Jiajia
    Wang, Xuan
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (03) : 3769 - 3779
  • [29] Adaptive Social Metaverse Streaming Based on Federated Multiagent Deep Reinforcement Learning
    Long, Zijian
    Wang, Haopeng
    Dong, Haiwei
    El Saddik, Abdulmotaleb
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2025,
  • [30] Efficient Communication via Self-Supervised Information Aggregation for Online and Offline Multiagent Reinforcement Learning
    Guan, Cong
    Chen, Feng
    Yuan, Lei
    Zhang, Zongzhang
    Yu, Yang
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,