GCMA: An Adaptive Multiagent Reinforcement Learning Framework With Group Communication for Complex and Similar Tasks Coordination

被引:1
|
作者
Peng, Kexing [1 ]
Ma, Tinghuai [2 ,3 ]
Yu, Xin [2 ]
Rong, Huan [4 ]
Qian, Yurong [5 ]
Al-Nabhan, Najla [6 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Comp Sci, Nanjing 210044, Peoples R China
[2] Nanjing Univ Informat Sci & Technol, Sch Software, Nanjing 210044, Peoples R China
[3] Jiangsu Ocean Univ, Lianyungang 222005, Peoples R China
[4] Nanjing Univ Informat Sci & Technol, Sch Artificial Intelligence, Nanjing 210044, Peoples R China
[5] Xinjiang Univ, Urumqi 830008, Peoples R China
[6] King Saud Univ, Dept Comp Sci, Riyadh 11362, Saudi Arabia
基金
中国国家自然科学基金;
关键词
Task analysis; Training; Games; Reinforcement learning; Resource management; Redundancy; Graph neural networks; Complex task policy; group communication; multiagent reinforcement learning (MARL); multitasks;
D O I
10.1109/TG.2023.3346394
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Coordinating multiple agents with diverse tasks and changing goals without interference is a challenge. Multiagent reinforcement learning (MARL) aims to develop effective communication and joint policies using group learning. Some of the previous approaches required each agent to maintain a set of networks independently, resulting in no consideration of interactions. Joint communication work causes agents receiving information unrelated to their own tasks. Currently, agents with different task divisions are often grouped by action tendency, but this can lead to poor dynamic grouping. This article presents a two-phase solution for multiple agents, addressing these issues. The first phase develops heterogeneous agent communication joint policies using a group communication MARL framework (GCMA). The framework employs a periodic grouping strategy, reducing exploration and communication redundancy by dynamically assigning agent group hidden features through hypernetwork and graph communication. The scheme efficiently utilizes resources for adapting to multiple similar tasks. In the second phase, each agent's policy network is distilled into a generalized simple network, adapting to similar tasks with varying quantities and sizes. GCMA is tested in complex environments, such as StarCraft II and unmanned aerial vehicle (UAV) take-off, showing its well-performing for large-scale, coordinated tasks. It shows GCMA's effectiveness for solid generalization in multitask tests with simulated pedestrians.
引用
收藏
页码:670 / 682
页数:13
相关论文
共 50 条
  • [31] Prior Knowledge-Augmented Broad Reinforcement Learning Framework for Fault Diagnosis of Heterogeneous Multiagent Systems
    Guo, Li
    Ren, Yiran
    Li, Runze
    Jiang, Bin
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2024, 16 (01) : 115 - 123
  • [32] HTTP Adaptive Streaming Framework with Online Reinforcement Learning
    Kang, Jeongho
    Chung, Kwangsue
    APPLIED SCIENCES-BASEL, 2022, 12 (15):
  • [33] SON Coordination for parameter conflict resolution: A reinforcement learning framework
    Iacoboaiea, Ovidiu
    Sayrac, Berna
    Ben Jemaa, Sana
    Bianchi, Pascal
    2014 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE WORKSHOPS (WCNCW), 2014, : 196 - +
  • [34] Adaptive Discount Factor for Deep Reinforcement Learning in Continuing Tasks with Uncertainty
    Kim, MyeongSeop
    Kim, Jung-Su
    Choi, Myoung-Su
    Park, Jae-Han
    SENSORS, 2022, 22 (19)
  • [36] Multiagent Bayesian Deep Reinforcement Learning for Microgrid Energy Management under Communication Failures
    Zhou, Hao
    Aral, Atakan
    Brandic, Ivona
    Erol-Kantarci, Melike
    IEEE Internet of Things Journal, 2022, 9 (14) : 11685 - 11698
  • [37] Multiple mini-robots navigation using a collaborative multiagent reinforcement learning framework
    Chaysri, Piyabhum
    Blekas, Konstantinos
    Vlachos, Kostas
    ADVANCED ROBOTICS, 2020, 34 (13) : 902 - 916
  • [38] Fast Adaptive Task Offloading and Resource Allocation via Multiagent Reinforcement Learning in Heterogeneous Vehicular Fog Computing
    Gao, Zhen
    Yang, Lei
    Dai, Yu
    IEEE INTERNET OF THINGS JOURNAL, 2023, 10 (08) : 6818 - 6835
  • [39] Adaptive Event-Triggered Bipartite Formation for Multiagent Systems via Reinforcement Learning
    Zhao, Huarong
    Shan, Jinjun
    Peng, Li
    Yu, Hongnian
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (12) : 17817 - 17828
  • [40] MASK-RL: Multiagent Video Object Segmentation Framework Through Reinforcement Learning
    Vecchio, Giuseppe
    Palazzo, Simone
    Giordano, Daniela
    Rundo, Francesco
    Spampinato, Concetto
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (12) : 5103 - 5115