Cooperative Multi-Agent Reinforcement Learning with Hierarchical Relation Graph under Partial Observability

被引:8
|
作者
Li, Yang [1 ]
Wang, Xinzhi [1 ]
Wang, Jianshu [1 ]
Wang, Wei [1 ]
Luo, Xiangfeng [1 ]
Xie, Shaorong [1 ]
机构
[1] Shanghai Univ, Sch Comp Engn & Sci, Shanghai, Peoples R China
来源
2020 IEEE 32ND INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI) | 2020年
基金
中国国家自然科学基金;
关键词
Reinforcement Learning; Multi-Agent; Hierarchical Relation Graph;
D O I
10.1109/ICTAI50040.2020.00011
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cooperation among agents with partial observation is an important task in multi-agent reinforcement learning (MARL), aiming to maximize a common reward. Most existing cooperative MARL approaches focus on building different model frameworks, such as centralized, decentralized, and centralized training with decentralized execution. These methods employ partial observation of agents as input directly, but rarely consider the local relationship between agents. The local relationship can help agents integrate observation information among different agents in a local range, and then adopt a more effective cooperation policy. In this paper, we propose a MARL method based on spatial relationship called hierarchical relation graph soft actorcritic (HRG-SAC). The method first uses a hierarchical relation graph generation module to represent the spatial relationship between agents in local space. Second, it integrates feature information of the relation graph through the graph convolution network (GCN). Finally, the soft actor-critic (SAC) is used to optimize agents' actions in training for compliance control. We conduct experiments on the Food Collector task and compare HRG-SAC with three baseline methods. The results demonstrate that the hierarchical relation graph can significantly improve MARL performance in the cooperative task.T
引用
收藏
页码:1 / 8
页数:8
相关论文
共 50 条
  • [1] Asynchronous multi-agent deep reinforcement learning under partial observability
    Xiao, Yuchen
    Tan, Weihao
    Hoffman, Joshua
    Xia, Tian
    Amato, Christopher
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2025,
  • [2] Distributed Cooperative Multi-Agent Reinforcement Learning with Directed Coordination Graph
    Jing, Gangshan
    Bai, He
    George, Jemin
    Chakrabortty, Aranya
    Sharma, Piyush K.
    2022 AMERICAN CONTROL CONFERENCE, ACC, 2022, : 3273 - 3278
  • [3] Multi-agent Cooperative Search based on Reinforcement Learning
    Sun, Yinjiang
    Zhang, Rui
    Liang, Wenbao
    Xu, Cheng
    PROCEEDINGS OF 2020 3RD INTERNATIONAL CONFERENCE ON UNMANNED SYSTEMS (ICUS), 2020, : 891 - 896
  • [4] Centralized reinforcement learning for multi-agent cooperative environments
    Chengxuan Lu
    Qihao Bao
    Shaojie Xia
    Chongxiao Qu
    Evolutionary Intelligence, 2024, 17 : 267 - 273
  • [5] Centralized reinforcement learning for multi-agent cooperative environments
    Lu, Chengxuan
    Bao, Qihao
    Xia, Shaojie
    Qu, Chongxiao
    EVOLUTIONARY INTELLIGENCE, 2024, 17 (01) : 267 - 273
  • [6] Transform networks for cooperative multi-agent deep reinforcement learning
    Hongbin Wang
    Xiaodong Xie
    Lianke Zhou
    Applied Intelligence, 2023, 53 : 9261 - 9269
  • [7] Transform networks for cooperative multi-agent deep reinforcement learning
    Wang, Hongbin
    Xie, Xiaodong
    Zhou, Lianke
    APPLIED INTELLIGENCE, 2023, 53 (08) : 9261 - 9269
  • [8] Cooperative Behavior by Multi-agent Reinforcement Learning with Abstractive Communication
    Tanda, Jin
    Moustafa, Ahmed
    Ito, Takayuki
    2019 IEEE INTERNATIONAL CONFERENCE ON AGENTS (ICA), 2019, : 8 - 13
  • [9] Knowledge Reuse of Multi-Agent Reinforcement Learning in Cooperative Tasks
    Shi, Daming
    Tong, Junbo
    Liu, Yi
    Fan, Wenhui
    ENTROPY, 2022, 24 (04)
  • [10] Distributed hierarchical reinforcement learning in multi-agent adversarial environments
    Naderializadeh, Navid
    Soleyman, Sean
    Hung, Fan
    Khosla, Deepak
    Chen, Yang
    Fadaie, Joshua G.
    ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING FOR MULTI-DOMAIN OPERATIONS APPLICATIONS IV, 2022, 12113