SADMA: Scalable Asynchronous Distributed Multi-agent Reinforcement Learning Training Framework

被引:0
|
作者
Wang, Sizhe [1 ]
Qian, Long [1 ]
Yi, Cairun [1 ]
Wu, Fan [1 ]
Kou, Qian [1 ]
Li, Mingyang [1 ]
Chen, Xingyug [1 ]
Lan, Xuguang [1 ]
机构
[1] Xi An Jiao Tong Univ, Inst Artificial Intelligence & Robot, Natl Engn Res Ctr Visual Informat & Applicat, Natl Key Lab Human Machine Hybrid Augmented Intel, Xian 710049, Shaanxi, Peoples R China
来源
基金
国家重点研发计划;
关键词
Multi-agent Reinforcement Learning; Distributed Training; Large Scale Multi-agent Training;
D O I
10.1007/978-3-031-71152-7_4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-agent Reinforcement Learning (MARL) has shown significant success in solving large-scale complex decision-making problems in multi-agent systems (MAS) while facing the challenge of increasing computational cost and training time. MARL algorithms often require sufficient environment exploration to achieve good performance, especially for complex environments, where the interaction frequency and synchronous training scheme can severely limit the overall speed. Most existing RL training frameworks, which utilize distributed training for acceleration, focus on simple single-agent settings and are not scalable to extend to large-scale MARL scenarios. To address this problem, we introduce a Scalable Asynchronous Distributed Multi-Agent RL training framework called SADMA, which modularizes the training process and executes the modules in an asynchronous and distributed manner for efficient training. Our framework is powerfully scalable and provides an efficient solution for distributed training of multi-agent reinforcement learning in large-scale complex environments. Code is available at https://github.com/sadmaenv/sadma.
引用
收藏
页码:64 / 81
页数:18
相关论文
共 50 条
  • [31] Optimal Tracking Agent: A New Framework for Multi-Agent Reinforcement Learning
    Cao, Weihua
    Chen, Gang
    Chen, Xin
    Wu, Min
    TRUSTCOM 2011: 2011 INTERNATIONAL JOINT CONFERENCE OF IEEE TRUSTCOM-11/IEEE ICESS-11/FCST-11, 2011, : 1328 - 1334
  • [32] Multi-Agent Deep Reinforcement Learning for Distributed Load Restoration
    Linh Vu
    Tuyen Vu
    Thanh Long Vu
    Srivastava, Anurag
    IEEE TRANSACTIONS ON SMART GRID, 2024, 15 (02) : 1749 - 1760
  • [33] Distributed Inverse Constrained Reinforcement Learning for Multi-agent Systems
    Liu, Shicheng
    Zhu, Minghui
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [34] Multi-agent deep reinforcement learning strategy for distributed energy
    Xi, Lei
    Sun, Mengmeng
    Zhou, Huan
    Xu, Yanchun
    Wu, Junnan
    Li, Yanying
    MEASUREMENT, 2021, 185
  • [35] Multi-Agent Deep Reinforcement Learning for Distributed Satellite Routing
    Lozano-Cuadra, Federico
    Soret, Beatriz
    2024 IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING FOR COMMUNICATION AND NETWORKING, ICMLCN 2024, 2024, : 554 - 555
  • [36] A Multi-agent Reinforcement Learning Perspective on Distributed Traffic Engineering
    Geng, Nan
    Lan, Tian
    Aggarwal, Vaneet
    Yang, Yuan
    Xu, Mingwei
    2020 IEEE 28TH INTERNATIONAL CONFERENCE ON NETWORK PROTOCOLS (IEEE ICNP 2020), 2020,
  • [37] Distributed hierarchical reinforcement learning in multi-agent adversarial environments
    Naderializadeh, Navid
    Soleyman, Sean
    Hung, Fan
    Khosla, Deepak
    Chen, Yang
    Fadaie, Joshua G.
    ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING FOR MULTI-DOMAIN OPERATIONS APPLICATIONS IV, 2022, 12113
  • [38] PowerNet: Multi-Agent Deep Reinforcement Learning for Scalable Powergrid Control
    Chen, Dong
    Chen, Kaian
    Li, Zhaojian
    Chu, Tianshu
    Yao, Rui
    Qiu, Feng
    Lin, Kaixiang
    IEEE TRANSACTIONS ON POWER SYSTEMS, 2022, 37 (02) : 1007 - 1017
  • [39] Scalable Autonomous Separation Assurance With Heterogeneous Multi-Agent Reinforcement Learning
    Brittain, Marc
    Wei, Peng
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2022, 19 (04) : 2837 - 2848
  • [40] Scalable Multi-Agent Reinforcement Learning for Networked Systems with Average Reward
    Qu, Guannan
    Lin, Yiheng
    Wierman, Adam
    Li, Na
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33