A Decentralized Communication Framework Based on Dual-Level Recurrence for Multiagent Reinforcement Learning

被引:3
|
作者
Li, Xuesi [1 ]
Li, Jingchen [1 ]
Shi, Haobin [1 ]
Hwang, Kao-Shing [2 ]
机构
[1] Northwestern Polytech Univ, Sch Comp Sci & Engn, Xian 710129, Shaanxi, Peoples R China
[2] Natl Sun Yat Sen Univ, Dept Elect Engn, Kaohsiung 804, Taiwan
基金
中国国家自然科学基金;
关键词
Reinforcement learning; Logic gates; Training; Adaptation models; Multi-agent systems; Task analysis; Decision making; Gated recurrent network; multiagent reinforcement learning; multiagent system;
D O I
10.1109/TCDS.2023.3281878
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Designing communication channels for multiagent is a feasible method to conduct decentralized learning, especially in partially observable environments or large-scale multiagent systems. In this work, a communication model with dual-level recurrence is developed to provide a more efficient communication mechanism for the multiagent reinforcement learning field. The communications are conducted by a gated-attention-based recurrent network, in which the historical states are taken into account and regarded as the second-level recurrence. We separate communication messages from memories in the recurrent model so that the proposed communication flow can adapt changeable communication objects in the case of limited communication, and the communication results are fair to every agent. We provide a sufficient discussion about our method in both partially observable and fully observable environments. The results of several experiments suggest our method outperforms the existing decentralized communication frameworks and the corresponding centralized training method.
引用
收藏
页码:640 / 649
页数:10
相关论文
共 50 条
  • [41] A multiagent reinforcement learning approach based on different states
    李珺
    潘启树
    Journal of Harbin Institute of Technology(New series), 2010, (03) : 419 - 423
  • [42] A MULTIAGENT REINFORCEMENT LEARNING FRAMEWORK FOR OFF-POLICY EVALUATION IN TWO-SIDED MARKETS
    Shi, Chengchun
    Wan, Runzhe
    Song, Ge
    Luo, Shikai
    Zhu, Hongtu
    Song, Rui
    ANNALS OF APPLIED STATISTICS, 2023, 17 (04) : 2701 - 2722
  • [43] Multiagent Reinforcement Learning Based Distributed Channel Access for Industrial Edge-Cloud Web 3.0
    Yang, Chen
    Wang, Yushi
    Lan, Shulin
    Zhu, Liehuang
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2024, 11 (05): : 3943 - 3954
  • [44] A Deep Reinforcement Learning-Based Framework for PolSAR Imagery Classification
    Nie, Wen
    Huang, Kui
    Yang, Jie
    Li, Pingxiang
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [45] Reinforcement Learning-Based Decentralized Control for Networked Interconnected Systems With Communication and Control Constraints
    Liu, Jinliang
    Zhang, Nan
    Zha, Lijuan
    Xie, Xiangpeng
    Tian, Engang
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2024, 21 (03) : 4674 - 4685
  • [46] Adaptive Social Metaverse Streaming Based on Federated Multiagent Deep Reinforcement Learning
    Long, Zijian
    Wang, Haopeng
    Dong, Haiwei
    El Saddik, Abdulmotaleb
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2025,
  • [47] Experience Sharing Based Memetic Transfer Learning for Multiagent Reinforcement Learning
    Wang, Tonghao
    Peng, Xingguang
    Jin, Yaochu
    Xu, Demin
    MEMETIC COMPUTING, 2022, 14 (01) : 3 - 17
  • [48] Opinion Dynamics in Gossiper-Media Networks Based on Multiagent Reinforcement Learning
    Zhang, Chengwei
    Fang, Dina
    Sen, Sandip
    Li, Xiaohong
    Feng, Zhiyong
    Xue, Wanli
    An, Dou
    Zhao, Xintian
    Chen, Rong
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2023, 10 (02): : 1143 - 1156
  • [49] Experience Sharing Based Memetic Transfer Learning for Multiagent Reinforcement Learning
    Tonghao Wang
    Xingguang Peng
    Yaochu Jin
    Demin Xu
    Memetic Computing, 2022, 14 : 3 - 17
  • [50] Decentralized, Safe, Multiagent Motion Planning for Drones Under Uncertainty via Filtered Reinforcement Learning
    Vinod, Abraham P.
    Safaoui, Sleiman
    Summers, Tyler H.
    Yoshikawa, Nobuyuki
    Di Cairano, Stefano
    IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, 2024, 32 (06) : 2492 - 2499