Collective Intrinsic Motivation of a Multi-agent System Based on Reinforcement Learning Algorithms

被引:0
|
作者
Bolshakov, Vladislav [1 ]
Sakulin, Sergey [1 ]
Alfimtsev, Alexander [1 ]
机构
[1] BMSTU, Moscow, Russia
来源
INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 4, INTELLISYS 2023 | 2024年 / 825卷
关键词
Multi-agent reinforcement learning; Intrinsic motivation; Reward shaping; LEVEL;
D O I
10.1007/978-3-031-47718-8_42
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
One of the great challenges in reinforcement learning is learning an optimal behavior in environments with sparse rewards. Solving tasks in such setting require effective exploration methods that are often based on intrinsic rewards. Plenty of real-world problems involve sparse rewards and many of them are further complicated by multi-agent setting, where the majority of intrinsic motivation methods are ineffective. In this paper we address the problem of multi-agent environments with sparse rewards and propose to combine intrinsic rewards and multi-agent reinforcement learning (MARL) technics to create the Collective Intrinsic Motivation of Agents (CIMA) method. CIMA uses both the external reward and the intrinsic collective reward from the cooperative multi-agent system. The proposed method can be used along with any MARL method as base reinforcement learning algorithm. We compare CIMA with several state-of-the-art MARL methods within multi-agent environment with sparse rewards designed in StarCraft II.
引用
收藏
页码:655 / 670
页数:16
相关论文
共 50 条
  • [21] OptimizingMARL: Developing Cooperative Game Environments Based on Multi-agent Reinforcement Learning
    Ferreira, Thais
    Clua, Esteban
    Kohwalter, Troy Costa
    Santos, Rodrigo
    ENTERTAINMENT COMPUTING, ICEC 2022, 2022, 13477 : 89 - 102
  • [22] A Multi-Agent Reinforcement Learning Approach for Blockchain-based Electricity Trading System
    Cao, Yifan
    Ren, Xiaoxu
    Qiu, Chao
    Wang, Xiaofei
    Yao, Haipeng
    Yu, F. Richard
    2021 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2021,
  • [23] Multi-agent reinforcement learning based coordinated control of PEMFC gas supply system
    Wang, Guowei
    Wang, Xinli
    Wang, Lei
    Jia, Lei
    Shao, Mingjun
    Yu, Youliang
    2022 IEEE 17TH CONFERENCE ON INDUSTRIAL ELECTRONICS AND APPLICATIONS (ICIEA), 2022, : 132 - 137
  • [24] Aggregation Transfer Learning for Multi-Agent Reinforcement learning
    Xu, Dongsheng
    Qiao, Peng
    Dou, Yong
    2021 2ND INTERNATIONAL CONFERENCE ON BIG DATA & ARTIFICIAL INTELLIGENCE & SOFTWARE ENGINEERING (ICBASE 2021), 2021, : 547 - 551
  • [25] Multi-AGV Scheduling based on Hierarchical Intrinsically Rewarded Multi-Agent Reinforcement Learning
    Zhang, Jiangshan
    Guo, Bin
    Sun, Zhuo
    Li, Mengyuan
    Liu, Jiaqi
    Yu, Zhiwen
    Fan, Xiaopeng
    2022 IEEE 19TH INTERNATIONAL CONFERENCE ON MOBILE AD HOC AND SMART SYSTEMS (MASS 2022), 2022, : 155 - 161
  • [26] Concept Learning for Interpretable Multi-Agent Reinforcement Learning
    Zabounidis, Renos
    Campbell, Joseph
    Stepputtis, Simon
    Hughes, Dana
    Sycara, Katia
    CONFERENCE ON ROBOT LEARNING, VOL 205, 2022, 205 : 1828 - 1837
  • [27] Learning structured communication for multi-agent reinforcement learning
    Sheng, Junjie
    Wang, Xiangfeng
    Jin, Bo
    Yan, Junchi
    Li, Wenhao
    Chang, Tsung-Hui
    Wang, Jun
    Zha, Hongyuan
    AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2022, 36 (02)
  • [28] Learning structured communication for multi-agent reinforcement learning
    Junjie Sheng
    Xiangfeng Wang
    Bo Jin
    Junchi Yan
    Wenhao Li
    Tsung-Hui Chang
    Jun Wang
    Hongyuan Zha
    Autonomous Agents and Multi-Agent Systems, 2022, 36
  • [29] Multi-agent reinforcement learning for character control
    Li, Cheng
    Fussell, Levi
    Komura, Taku
    VISUAL COMPUTER, 2021, 37 (12): : 3115 - 3123
  • [30] HALFTONING WITH MULTI-AGENT DEEP REINFORCEMENT LEARNING
    Jiang, Haitian
    Xiong, Dongliang
    Jiang, Xiaowen
    Yin, Aiguo
    Ding, Li
    Huang, Kai
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 641 - 645