A leader-following paradigm based deep reinforcement learning method for multi-agent cooperation games

被引:10
|
作者
Zhang, Feiye [1 ]
Yang, Qingyu [1 ,2 ]
An, Dou [1 ,2 ]
机构
[1] Xi An Jiao Tong Univ, Fac Elect & Informat Engn, 28, West Xianning Rd, Xian 710049, Shaanxi, Peoples R China
[2] Xi An Jiao Tong Univ, MOE Key Lab Intelligent Networks & Network Secur, 28, West Xianning Rd, Xian 710049, Shaanxi, Peoples R China
基金
美国国家科学基金会; 中国国家自然科学基金;
关键词
Multi-agent systems; Deep reinforcement learning; Centralized training with decentralized; execution; Cooperative games; LEVEL;
D O I
10.1016/j.neunet.2022.09.012
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-agent deep reinforcement learning algorithms with centralized training with decentralized execution (CTDE) paradigm has attracted growing attention in both industry and research community. However, the existing CTDE methods follow the action selection paradigm that all agents choose actions at the same time, which ignores the heterogeneous roles of different agents. Motivated by the human wisdom in cooperative behaviors, we present a novel leader-following paradigm based deep multi-agent cooperation method (LFMCO) for multi-agent cooperative games. Specifically, we define a leader as someone who broadcasts a message representing the selected action to all subordinates. After that, the followers choose their individual action based on the received message from the leader. To measure the influence of leader's action on followers, we introduced a concept of information gain, i.e., the change of followers' value function entropy, which is positively correlated with the influence of leader's action. We evaluate the LFMCO on several cooperation scenarios of StarCraft2. Simulation results confirm the significant performance improvements of LFMCO compared with four state-of-the-art benchmarks on the challenging cooperative environment.(c) 2022 Elsevier Ltd. All rights reserved.
引用
收藏
页码:1 / 12
页数:12
相关论文
共 50 条
  • [1] Leader-following multi-agent consensus control
    Zhang W.-G.
    Qu S.-L.
    Yuhang Xuebao/Journal of Astronautics, 2010, 31 (09): : 2172 - 2176
  • [2] LEADER-FOLLOWING CONSENSUS OF MULTI-AGENT SYSTEM WITH DIFFUSION
    Guo, Ling
    Nian, Xiaohong
    Pan, Huan
    ASIAN JOURNAL OF CONTROL, 2014, 16 (01) : 188 - 197
  • [3] On Consensus Speed of the Leader-following Multi-agent System
    Zhang Xiaoyong
    Zhang Guoshan
    PROCEEDINGS OF THE 31ST CHINESE CONTROL CONFERENCE, 2012, : 6384 - 6388
  • [4] Leader-following consensus of multi-agent systems with memory
    Feng, Xin-Lei
    Huang, Ting-Zhu
    JOURNAL OF COMPUTATIONAL ANALYSIS AND APPLICATIONS, 2014, 16 (02) : 315 - 328
  • [5] A fully value distributional deep reinforcement learning framework for multi-agent cooperation
    Fu, Mingsheng
    Huang, Liwei
    Li, Fan
    Qu, Hong
    Xu, Chengzhong
    NEURAL NETWORKS, 2025, 184
  • [6] Formation Control of Multi-agent Based on Deep Reinforcement Learning
    Pan, Chao
    Nian, Xiaohong
    Dai, Xunhua
    Wang, Haibo
    Xiong, Hongyun
    PROCEEDINGS OF 2022 INTERNATIONAL CONFERENCE ON AUTONOMOUS UNMANNED SYSTEMS, ICAUS 2022, 2023, 1010 : 1149 - 1159
  • [7] Multi-agent deep reinforcement learning: a survey
    Gronauer, Sven
    Diepold, Klaus
    ARTIFICIAL INTELLIGENCE REVIEW, 2022, 55 (02) : 895 - 943
  • [8] Multi-Agent Cognition Difference Reinforcement Learning for Multi-Agent Cooperation
    Wang, Huimu
    Qiu, Tenghai
    Liu, Zhen
    Pu, Zhiqiang
    Yi, Jianqiang
    Yuan, Wanmai
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [9] Robust Observer-Based Leader-Following Consensus for Multi-Agent Systems
    Vazquez Trejo, Jesus Avelino
    Adam-Medina, Manuel
    Vazquez, Juan A.
    Garcia, C. D.
    Garcia, J.
    IEEE LATIN AMERICA TRANSACTIONS, 2021, 19 (11) : 1949 - 1958
  • [10] Adaptive Leader-Following Consensus of Linear Multi-Agent Systems
    Lee S.
    Journal of Institute of Control, Robotics and Systems, 2022, 28 (09): : 787 - 792