A leader-following paradigm based deep reinforcement learning method for multi-agent cooperation games

被引:10
|
作者
Zhang, Feiye [1 ]
Yang, Qingyu [1 ,2 ]
An, Dou [1 ,2 ]
机构
[1] Xi An Jiao Tong Univ, Fac Elect & Informat Engn, 28, West Xianning Rd, Xian 710049, Shaanxi, Peoples R China
[2] Xi An Jiao Tong Univ, MOE Key Lab Intelligent Networks & Network Secur, 28, West Xianning Rd, Xian 710049, Shaanxi, Peoples R China
基金
美国国家科学基金会; 中国国家自然科学基金;
关键词
Multi-agent systems; Deep reinforcement learning; Centralized training with decentralized; execution; Cooperative games; LEVEL;
D O I
10.1016/j.neunet.2022.09.012
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-agent deep reinforcement learning algorithms with centralized training with decentralized execution (CTDE) paradigm has attracted growing attention in both industry and research community. However, the existing CTDE methods follow the action selection paradigm that all agents choose actions at the same time, which ignores the heterogeneous roles of different agents. Motivated by the human wisdom in cooperative behaviors, we present a novel leader-following paradigm based deep multi-agent cooperation method (LFMCO) for multi-agent cooperative games. Specifically, we define a leader as someone who broadcasts a message representing the selected action to all subordinates. After that, the followers choose their individual action based on the received message from the leader. To measure the influence of leader's action on followers, we introduced a concept of information gain, i.e., the change of followers' value function entropy, which is positively correlated with the influence of leader's action. We evaluate the LFMCO on several cooperation scenarios of StarCraft2. Simulation results confirm the significant performance improvements of LFMCO compared with four state-of-the-art benchmarks on the challenging cooperative environment.(c) 2022 Elsevier Ltd. All rights reserved.
引用
收藏
页码:1 / 12
页数:12
相关论文
共 50 条
  • [41] Leader-following consensus of multi-agent systems under antagonistic networks
    Wang, Qianyao
    Liu, Kexin
    Wang, Xiong
    Wu, Lulu
    Lu, Jinhu
    NEUROCOMPUTING, 2020, 413 : 339 - 347
  • [42] Deep reinforcement learning for multi-agent interaction
    Ahmed, Ibrahim H.
    Brewitt, Cillian
    Carlucho, Ignacio
    Christianos, Filippos
    Dunion, Mhairi
    Fosong, Elliot
    Garcin, Samuel
    Guo, Shangmin
    Gyevnar, Balint
    McInroe, Trevor
    Papoudakis, Georgios
    Rahman, Arrasy
    Schafer, Lukas
    Tamborski, Massimiliano
    Vecchio, Giuseppe
    Wang, Cheng
    Albrecht, Stefano, V
    AI COMMUNICATIONS, 2022, 35 (04) : 357 - 368
  • [43] Multi-agent deep reinforcement learning: a survey
    Sven Gronauer
    Klaus Diepold
    Artificial Intelligence Review, 2022, 55 : 895 - 943
  • [44] Intelligent multicast routing method based on multi-agent deep reinforcement learning in SDWN
    Hu, Hongwen
    Ye, Miao
    Zhao, Chenwei
    Jiang, Qiuxiang
    Xue, Xingsi
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2023, 20 (09) : 17158 - 17196
  • [45] Generalized multi-synchronization: A leader-following consensus problem of multi-agent systems
    Cruz-Ancona, Christopher D.
    Martinez-Guerra, Rafael
    Perez-Pinacho, Claudia A.
    NEUROCOMPUTING, 2017, 233 : 52 - 60
  • [46] Observer-based fault-tolerant leader-following control for multi-agent systems
    Trejo, Juan Antonio Vazquez
    Rotondo, Damiano
    Medina, Manuel Adam
    Theilliol, Didier
    2021 EUROPEAN CONTROL CONFERENCE (ECC), 2021, : 2285 - 2290
  • [47] Multi-agent deep reinforcement learning for online request scheduling in edge cooperation networks
    Zhang, Yaqiang
    Li, Ruyang
    Zhao, Yaqian
    Li, Rengang
    Wang, Yanwei
    Zhou, Zhangbing
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2023, 141 : 258 - 268
  • [48] Multi-Agent Deep Reinforcement Learning for Distributed Load Restoration
    Linh Vu
    Tuyen Vu
    Thanh Long Vu
    Srivastava, Anurag
    IEEE TRANSACTIONS ON SMART GRID, 2024, 15 (02) : 1749 - 1760
  • [49] Leader-following Consensus Problem of Multi-agent Networks Based on Homogeneous Polynomial Lyapunov Functions
    Tan Chong
    Duan Guang-Ren
    Liu Guoping
    PROCEEDINGS OF THE 29TH CHINESE CONTROL CONFERENCE, 2010, : 5747 - 5752
  • [50] Leader-following consensus of multi-agent systems under fixed and switching topologies
    Ni, Wei
    Cheng, Daizhan
    SYSTEMS & CONTROL LETTERS, 2010, 59 (3-4) : 209 - 217