Utilizing Observed Information for No-Communication Multi-Agent Reinforcement Learning toward Cooperation in Dynamic Environment

被引:1
|
作者
Uwano F. [1 ]
Takadama K. [1 ]
机构
[1] Department of Informatics, The University of Electro-Communications
关键词
dynamic environment; memory management; multi-agent system; reinforcement learning;
D O I
10.9746/jcmsi.12.199
中图分类号
学科分类号
摘要
This paper proposes a multi-agent reinforcement learning method without communication toward dynamic environments, called profit minimizing reinforcement learning with oblivion of memory (PMRL-OM). PMRL-OM is extended from PMRL and defines a memory range that only utilizes the valuable information from the environment. Since agents do not require information observed before an environmental change, the agents utilize the information acquired after a certain iteration, which is performed by the memory range. In addition, PMRL-OM improves the update function for a goal value as a priority of purpose and updates the goal value based on newer information. To evaluate the effectiveness of PMRL-OM, this study compares PMRL-OM with PMRL in five dynamic maze environments, including state changes for two types of cooperation, position changes for two types of cooperation, and a combined case from these four cases. The experimental results revealed that: (a) PMRL-OM was an effective method for cooperation in all five cases of dynamic environments examined in this study; (b) PMRL-OM was more effective than PMRL was in these dynamic environments; and (c) in a memory range of 100 to 500, PMRL-OM performs well. © Taylor & Francis Group, LLC 2019.
引用
收藏
页码:199 / 208
页数:9
相关论文
共 50 条
  • [21] The evolution of cooperation in continuous dilemmas via multi-agent reinforcement learning
    Zhu, Congcong
    Ye, Dayong
    Zhu, Tianqing
    Zhou, Wanlei
    KNOWLEDGE-BASED SYSTEMS, 2025, 315
  • [22] Multi-agent Pathfinding with Communication Reinforcement Learning and Deadlock Detection
    Ye, Zhaohui
    Li, Yanjie
    Guo, Ronghao
    Gao, Jianqi
    Fu, Wen
    INTELLIGENT ROBOTICS AND APPLICATIONS (ICIRA 2022), PT I, 2022, 13455 : 493 - 504
  • [23] Self-Clustering Hierarchical Multi-Agent Reinforcement Learning With Extensible Cooperation Graph
    Fu, Qingxu
    Qiu, Tenghai
    Yi, Jianqiang
    Pu, Zhiqiang
    Ai, Xiaolin
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024,
  • [24] Coordinated Reinforcement Learning Agents in a Multi-Agent Virtual Environment
    Sause, William
    2013 12TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2013), VOL 1, 2013, : 227 - 230
  • [25] Leveraging Expert Demonstrations in Robot Cooperation with Multi-Agent Reinforcement Learning
    Zhang, Zhaolong
    Li, Yihui
    Rojas, Juan
    Guan, Yisheng
    INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2021, PT II, 2021, 13014 : 211 - 222
  • [26] Research on cooperation and learning in multi-agent system
    Zheng, SL
    Luo, XF
    Luo, ZH
    Yang, JG
    2002 6TH INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING PROCEEDINGS, VOLS I AND II, 2002, : 1159 - 1162
  • [27] Reinforcement Learning in Multi-Agent Environment and Ant Colony for Packet Scheduling in Routers
    Bournenane, Malika
    Mellouk, Abdelhamid
    Benhamamouche, Djilali
    MOBIWAC'07: PROCEEDINGS OF THE FIFTH ACM INTERNATIONAL WORKSHOP ON MOBILITY MANAGEMENT AND WIRELESS ACCESS, 2007, : 137 - 143
  • [28] A Review of Multi-Agent Reinforcement Learning Algorithms
    Liang, Jiaxin
    Miao, Haotian
    Li, Kai
    Tan, Jianheng
    Wang, Xi
    Luo, Rui
    Jiang, Yueqiu
    ELECTRONICS, 2025, 14 (04):
  • [29] SCM network with multi-agent reinforcement learning
    Zhao, Gang
    Sun, Ruoying
    FIFTH WUHAN INTERNATIONAL CONFERENCE ON E-BUSINESS, VOLS 1-3, 2006, : 1294 - 1300
  • [30] A multi-agent reinforcement learning approach to dynamic service composition
    Wang, Hongbing
    Wang, Xiaojun
    Hu, Xingguo
    Zhang, Xingzhi
    Gu, Mingzhu
    INFORMATION SCIENCES, 2016, 363 : 96 - 119