Action Prediction for Cooperative Exploration in Multi-agent Reinforcement Learning

被引:0
|
作者
Zhang, Yanqiang [1 ]
Feng, Dawei [1 ]
Ding, Bo [1 ]
机构
[1] Natl Univ Def Technol, Natl Lab Parallel & Distributed Proc, Changsha 410073, Peoples R China
来源
NEURAL INFORMATION PROCESSING, ICONIP 2023, PT II | 2024年 / 14448卷
关键词
Multi-agent Systems; Reinforcement Learning; Intrinsic Reward;
D O I
10.1007/978-981-99-8082-6_28
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-agent reinforcement learning methods have shown significant progress, however, they continue to exhibit exploration problems in complex and challenging environments. To address the above issue, current research has introduced several exploration-enhanced methods for multi-agent reinforcement learning, they are still faced with the issues of inefficient exploration and low performance in challenging tasks that necessitate complex cooperation among agents. This paper proposes the prediction-action Qmix (PQmix) method, an action prediction-based multi-agent intrinsic reward construction approach. The PQmix method employs the joint local observation of agents and the next joint local observation after executing actions to predict the real joint action of agents. The method calculates the action prediction error as the intrinsic reward to measure the novel of the joint state and encourages agents to actively explore the action and state spaces in the environment. We compare PQmix with strong baselines on the MARL benchmark to validate it. The result of experiments demonstrates that PQmix outperforms the state-of-the-art algorithms on the StarCraft Multi-Agent Challenge (SMAC). In the end, the stability of the method is verified by experiments.
引用
收藏
页码:358 / 372
页数:15
相关论文
共 50 条
  • [41] Intrinsic Reward with Peer Incentives for Cooperative Multi-Agent Reinforcement Learning
    Zhang, Tianle
    Liu, Zhen
    Wu, Shiguang
    Pu, Zhiqiang
    Yi, Jianqiang
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [42] Commander-Soldiers Reinforcement Learning for Cooperative Multi-Agent Systems
    Chen, Yiqun
    Yang, Wei
    Zhang, Tianle
    Wu, Shiguang
    Chang, Hongxing
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [43] A Cooperative Multi-Agent Reinforcement Learning Method Based on Coordination Degree
    Cui, Haoyan
    Zhang, Zhen
    IEEE ACCESS, 2021, 9 : 123805 - 123814
  • [44] Testing Reinforcement Learning Explainability Methods in a Multi-Agent Cooperative Environment
    Domenech i Vila, Marc
    Gnatyshak, Dmitry
    Tormos, Adrian
    Alvarez-Napagao, Sergio
    ARTIFICIAL INTELLIGENCE RESEARCH AND DEVELOPMENT, 2022, 356 : 355 - 364
  • [45] Autonomous and cooperative control of UAV cluster with multi-agent reinforcement learning
    Xu, D.
    Chen, G.
    AERONAUTICAL JOURNAL, 2022, 126 (1300): : 932 - 951
  • [46] Cooperative reinforcement learning in topology-based multi-agent systems
    Xiao, Dan
    Tan, Ah-Hwee
    AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2013, 26 (01) : 86 - 119
  • [47] Microscopic Traffic Simulation by Cooperative Multi-agent Deep Reinforcement Learning
    Bacchiani, Giulio
    Molinari, Daniele
    Patander, Marco
    AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 1547 - 1555
  • [48] QDAP: Downsizing adaptive policy for cooperative multi-agent reinforcement learning
    Zhao, Zhitong
    Zhang, Ya
    Wang, Siying
    Zhang, Fan
    Zhang, Malu
    Chen, Wenyu
    KNOWLEDGE-BASED SYSTEMS, 2024, 294
  • [49] Generalized learning automata for multi-agent reinforcement learning
    De Hauwere, Yann-Michael
    Vrancx, Peter
    Nowe, Ann
    AI COMMUNICATIONS, 2010, 23 (04) : 311 - 324
  • [50] Decentralized Computation Offloading with Cooperative UAVs: Multi-Agent Deep Reinforcement Learning Perspective
    Hwang, Sangwon
    Lee, Hoon
    Park, Juseong
    Lee, Inkyu
    IEEE WIRELESS COMMUNICATIONS, 2022, 29 (04) : 24 - 31