Utilizing Observed Information for No-Communication Multi-Agent Reinforcement Learning toward Cooperation in Dynamic Environment

被引:1
|
作者
Uwano F. [1 ]
Takadama K. [1 ]
机构
[1] Department of Informatics, The University of Electro-Communications
关键词
dynamic environment; memory management; multi-agent system; reinforcement learning;
D O I
10.9746/jcmsi.12.199
中图分类号
学科分类号
摘要
This paper proposes a multi-agent reinforcement learning method without communication toward dynamic environments, called profit minimizing reinforcement learning with oblivion of memory (PMRL-OM). PMRL-OM is extended from PMRL and defines a memory range that only utilizes the valuable information from the environment. Since agents do not require information observed before an environmental change, the agents utilize the information acquired after a certain iteration, which is performed by the memory range. In addition, PMRL-OM improves the update function for a goal value as a priority of purpose and updates the goal value based on newer information. To evaluate the effectiveness of PMRL-OM, this study compares PMRL-OM with PMRL in five dynamic maze environments, including state changes for two types of cooperation, position changes for two types of cooperation, and a combined case from these four cases. The experimental results revealed that: (a) PMRL-OM was an effective method for cooperation in all five cases of dynamic environments examined in this study; (b) PMRL-OM was more effective than PMRL was in these dynamic environments; and (c) in a memory range of 100 to 500, PMRL-OM performs well. © Taylor & Francis Group, LLC 2019.
引用
收藏
页码:199 / 208
页数:9
相关论文
共 50 条
  • [31] Safe Multi-Agent Reinforcement Learning via Dynamic Shielding
    Qiu, Yunbo
    Jin, Yue
    Yu, Lebin
    Wang, Jian
    Zhang, Xudong
    2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024, 2024, : 1254 - 1257
  • [32] A Multi-Agent Reinforcement Learning Approach to the Dynamic Job Shop Scheduling Problem
    Inal, Ali Firat
    Sel, Cagri
    Aktepe, Adnan
    Turker, Ahmet Kursad
    Ersoz, Suleyman
    SUSTAINABILITY, 2023, 15 (10)
  • [33] Reinforcement learning of multi-agent communicative acts
    Hoet S.
    Sabouret N.
    Revue d'Intelligence Artificielle, 2010, 24 (02) : 159 - 188
  • [34] Developing multi-agent adversarial environment using reinforcement learning and imitation learning
    Ziyao Han
    Yupeng Liang
    Kazuhiro Ohkura
    Artificial Life and Robotics, 2023, 28 : 703 - 709
  • [35] Developing multi-agent adversarial environment using reinforcement learning and imitation learning
    Han, Ziyao
    Liang, Yupeng
    Ohkura, Kazuhiro
    ARTIFICIAL LIFE AND ROBOTICS, 2023, 28 (04) : 703 - 709
  • [36] Innovative Approach Towards Cooperation Models for Multi-agent Reinforcement Learning (CMMARL)
    Vidhate, Deepak A.
    Kulkarni, Parag
    SMART TRENDS IN INFORMATION TECHNOLOGY AND COMPUTER COMMUNICATIONS, SMARTCOM 2016, 2016, 628 : 468 - 478
  • [37] HyperComm: Hypergraph-based communication in multi-agent reinforcement learning
    Zhu, Tianyu
    Shi, Xinli
    Xu, Xiangping
    Gui, Jie
    Cao, Jinde
    NEURAL NETWORKS, 2024, 178
  • [38] Multi-Agent Uncertainty Sharing for Cooperative Multi-Agent Reinforcement Learning
    Chen, Hao
    Yang, Guangkai
    Zhang, Junge
    Yin, Qiyue
    Huang, Kaiqi
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [39] Emergent cooperation from mutual acknowledgment exchange in multi-agent reinforcement learning
    Phan, Thomy
    Sommer, Felix
    Ritz, Fabian
    Altmann, Philipp
    Nuesslein, Jonas
    Koelle, Michael
    Belzner, Lenz
    Linnhoff-Popien, Claudia
    AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2024, 38 (02)
  • [40] Testing Reinforcement Learning Explainability Methods in a Multi-Agent Cooperative Environment
    Domenech i Vila, Marc
    Gnatyshak, Dmitry
    Tormos, Adrian
    Alvarez-Napagao, Sergio
    ARTIFICIAL INTELLIGENCE RESEARCH AND DEVELOPMENT, 2022, 356 : 355 - 364