Multi-agent reinforcement learning with weak ties☆

被引:0
作者
Wang, Huan [1 ,2 ]
Zhou, Xu [1 ]
Kang, Yu [3 ]
Xue, Jian [4 ]
Yang, Chenguang [5 ]
Liu, Xiaofeng [1 ]
机构
[1] Hohai Univ, Coll Artificial Intelligence & Automat, Changzhou 213022, Peoples R China
[2] Chuzhou Polytechig, Sch Informat Engn, Chuzhou 239000, Peoples R China
[3] Univ Sci & Technol China, Sch Automat Engn, Hefei 230026, Peoples R China
[4] Univ Chinese Acad Sci, Sch Engn Sci, Beijing 100049, Peoples R China
[5] UWE, Bristol Robot Lab, Bristol BS161QY, England
关键词
Weak ties; Graph modeling; Information exchange; Optimal policy; Multi-agent reinforcement learning; COMMUNICATION; STRENGTH; SYSTEMS;
D O I
10.1016/j.inffus.2025.102942
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing multi-agent reinforcement learning (MARL) algorithms focus primarily on maximizing global game gains or encouraging cooperation between agents, often overlooking the weak ties between them. In multi- agent environments, the quality of the information exchanged is crucial for optimal policy learning. To this end, we propose a novel MARL framework that integrates weak-tie theory with graph modeling to forma weak- tie modeling module. We use the distribution of tie strengths and the dominant agent which is computed based on tie graph to control the information exchange between agents. Our method is evaluated against various baseline models indifferent multi-agent environments. Experimental results show that our method significantly improves the adversarial win rates and rewards of agents, and reduces the combat losses of agents in confrontation. Our method provides insights into how to reduce information redundancy in the training of large-scale agents.
引用
收藏
页数:12
相关论文
共 37 条
  • [1] Anastassacos N, 2020, AAAI CONF ARTIF INTE, V34, P7047
  • [2] Andrychowicz M., 2017, HINDSIGHT EXPERIENCE
  • [3] Structural holes and good ideas
    Burt, RS
    [J]. AMERICAN JOURNAL OF SOCIOLOGY, 2004, 110 (02) : 349 - 399
  • [4] Chen XY, 2020, Arxiv, DOI arXiv:1910.12179
  • [5] Das A, 2019, PR MACH LEARN RES, V97
  • [6] Ding Ziluo, 2020, Advances in Neural Information Processing Systems, V33
  • [7] Foerster J, 2017, PR MACH LEARN RES, V70
  • [8] Foerster JN, 2018, AAAI CONF ARTIF INTE, P2974
  • [9] Fujimoto S, 2018, PR MACH LEARN RES, V80
  • [10] Hu J., 2021, CoRR abs/2106.14334