Multi-Agent Evolutionary Reinforcement Learning Based on Cooperative Games

被引:0
作者
Yu, Jin [1 ,2 ]
Zhang, Ya [1 ,2 ]
Sun, Changyin [1 ,2 ]
机构
[1] Southeast Univ, Sch Automat, Nanjing 210096, Peoples R China
[2] Minist Educ, Key Lab Measurement & Control Complex Syst Engn, Nanjing 210096, Peoples R China
来源
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE | 2024年
关键词
Cooperative game; evolutionary algorithm; evolutionary reinforcement learning; multi-agent; reinforcement learning (RL);
D O I
10.1109/TETCI.2024.3452119
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Despite the significant advancements in single-agent evolutionary reinforcement learning, research exploring evolutionary reinforcement learning within multi-agent systems is still in its nascent stage. The integration of evolutionary algorithms (EA) and reinforcement learning (RL) has partially mitigated RL's reliance on the environment and provided it with an ample supply of data. Nonetheless, existing studies primarily focus on the indirect collaboration between RL and EA, which lacks sufficient exploration on the effective balance of individual and team rewards. To address this problem, this study introduces game theory to establish a dynamic cooperation framework between EA and RL, and proposes a multi-agent evolutionary reinforcement learning based on cooperative games. This framework facilitates more efficient direct collaboration between RL and EA, enhancing individual rewards while ensuring the attainment of team objectives. Initially, a cooperative policy is formed through a joint network to simplify the parameters of each agent to speed up the overall training process. Subsequently, RL and EA engage in cooperative games to determine whether RL jointly optimizes the same policy based on Pareto optimal results. Lastly, through double objectives optimization, a balance between the two types of rewards is achieved, with EA focusing on team rewards and RL focusing on individual rewards. Experimental results demonstrate that the proposed algorithm outperforms its single-algorithm counterparts in terms of competitiveness.
引用
收藏
页数:9
相关论文
共 31 条
  • [1] Bhalla Sushrut, 2020, Advances in Artificial Intelligence. 33rd Canadian Conference on Artificial Intelligence, Canadian AI 2020. Proceedings. Lecture Notes in Artificial Intelligence. Subseries of Lecture Notes in Computer Science (LNAI 12109), P67, DOI 10.1007/978-3-030-47358-7_7
  • [2] Branzei R, 2008, Models in cooperative game theory, V556
  • [3] Multi-Agent Reinforcement Learning: A Review of Challenges and Applications
    Canese, Lorenzo
    Cardarilli, Gian Carlo
    Di Nunzio, Luca
    Fazzolari, Rocco
    Giardino, Daniele
    Re, Marco
    Spano, Sergio
    [J]. APPLIED SCIENCES-BASEL, 2021, 11 (11):
  • [4] Curiel I., 2013, COOPERATIVE GAME THE, V16
  • [5] Du YL, 2019, ADV NEUR IN, V32
  • [6] Fujimoto S, 2018, PR MACH LEARN RES, V80
  • [7] An Evolutionary Transfer Reinforcement Learning Framework for Multiagent Systems
    Hou, Yaqing
    Ong, Yew-Soon
    Feng, Liang
    Zurada, Jacek M.
    [J]. IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2017, 21 (04) : 601 - 615
  • [8] A general deep reinforcement learning hyperheuristic framework for solving combinatorial optimization problems
    Kallestad, Jakob
    Hasibi, Ramin
    Hemmati, Ahmad
    Soerensen, Kenneth
    [J]. EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2023, 309 (01) : 446 - 468
  • [9] Lambora Annu, 2019, 2019 International Conference on Machine Learning, Big Data, Cloud and Parallel Computing (COMITCon), P380, DOI 10.1109/COMITCon.2019.8862255
  • [10] Liu H., 2024, J SUPERCOMPUTING, V80, P2346