Characterizing and Optimizing the End-to-End Performance of Multi-Agent Reinforcement Learning Systems

被引:0
作者
Gogineni, Kailash [1 ]
Mei, Yongsheng [1 ]
Gogineni, Karthikeya
Wei, Peng [1 ]
Lan, Tian [1 ]
Venkataramani, Guru [1 ]
机构
[1] George Washington Univ, Washington, DC 20052 USA
来源
2024 IEEE INTERNATIONAL SYMPOSIUM ON WORKLOAD CHARACTERIZATION, IISWC 2024 | 2024年
基金
美国国家科学基金会;
关键词
Multi-Agent Systems; Performance Analysis; Reinforcement Learning; Performance Optimization;
D O I
10.1109/IISWC63097.2024.00028
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Multi-Agent Reinforcement Learning Systems (MARL) can unlock the potential to model and control multiple autonomous decision-making agents simultaneously. During online training, MARL algorithms involve performance-intensive computations, such as exploration and exploitation phases originating from a large observation-action space and a huge number of training steps. Understanding and mitigating the MARL performance limiters is key to their practical adoption. In this paper, we first present a detailed workload characterization of MARL workloads under different multi-agent settings. Our experimental analysis identifies a critical performance bottleneck that affects scaling within the mini-batch sampling on transition data. To mitigate this issue, we explore a series of optimization strategies. First, we investigate cache locality-aware sampling that prioritizes intra-agent neighbor transitions over other randomly picked transition data samples within the baseline MARL algorithms. Next, we explore importance sampling techniques that preserve the learning performance/distribution and capture the neighbors of important transitions. Finally, we design an additional algorithmic optimization that reorganizes the transition data layout to improve the cache locality between different agents during the mini-batch sampling process. We evaluate our optimizations using popular MARL workloads on multi-agent particle games. Our work highlights several opportunities for enhancing the performance of multi-agent systems, with end-to-end training time improvements ranging from 8.2% (3 agents) to 20.5% (24 agents) compared to the baseline MADDPG, affirming the usefulness of deeply understanding MARL performance bottlenecks and mitigating them effectively.
引用
收藏
页码:224 / 235
页数:12
相关论文
共 50 条
  • [1] End-to-end Deep Reinforcement Learning for Multi-agent Collaborative Exploration
    Chen, Zichen
    Subagdja, Budhitama
    Tan, Ah-Hwee
    2019 IEEE INTERNATIONAL CONFERENCE ON AGENTS (ICA), 2019, : 99 - 102
  • [2] Cooperative Learning of Multi-Agent Systems Via Reinforcement Learning
    Wang, Xin
    Zhao, Chen
    Huang, Tingwen
    Chakrabarti, Prasun
    Kurths, Juergen
    IEEE TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING OVER NETWORKS, 2023, 9 : 13 - 23
  • [3] Shaping multi-agent systems with gradient reinforcement learning
    Buffet, Olivier
    Dutech, Alain
    Charpillet, Francois
    AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2007, 15 (02) : 197 - 220
  • [4] Shaping multi-agent systems with gradient reinforcement learning
    Olivier Buffet
    Alain Dutech
    François Charpillet
    Autonomous Agents and Multi-Agent Systems, 2007, 15 : 197 - 220
  • [5] Output synchronization of multi-agent systems via reinforcement learning
    Liu, Yingying
    Wang, Zhanshan
    NEUROCOMPUTING, 2022, 508 : 110 - 119
  • [6] Verifiably Safe Exploration for End-to-End Reinforcement Learning
    Hunt, Nathan
    Fulton, Nathan
    Magliacane, Sara
    Trong Nghia Hoang
    Das, Subhro
    Solar-Lezama, Armando
    HSCC2021: PROCEEDINGS OF THE 24TH INTERNATIONAL CONFERENCE ON HYBRID SYSTEMS: COMPUTATION AND CONTROL (PART OF CPS-IOT WEEK), 2021,
  • [7] Building Collaboration in Multi-agent Systems Using Reinforcement Learning
    Aydin, Mehmet Emin
    Fellows, Ryan
    COMPUTATIONAL COLLECTIVE INTELLIGENCE, ICCCI 2018, PT II, 2018, 11056 : 201 - 212
  • [8] PowerGridworld: A Framework for Multi-Agent Reinforcement Learning in Power Systems
    Biagioni, David
    Zhang, Xiangyu
    Wald, Dylan
    Vaidhynathan, Deepthi
    Chintala, Rohit
    King, Jennifer
    Zamzam, Ahmed S.
    PROCEEDINGS OF THE 2022 THE THIRTEENTH ACM INTERNATIONAL CONFERENCE ON FUTURE ENERGY SYSTEMS, E-ENERGY 2022, 2022, : 565 - 570
  • [9] Decentralized Incremental Fuzzy Reinforcement Learning for Multi-Agent Systems
    Hamzeloo, Sam
    Jahromi, Mansoor Zolghadri
    INTERNATIONAL JOURNAL OF UNCERTAINTY FUZZINESS AND KNOWLEDGE-BASED SYSTEMS, 2020, 28 (01) : 79 - 98
  • [10] ORACLE: End-to-End Model Based Reinforcement Learning
    Andersen, Per-Arne
    Goodwin, Morten
    Granmo, Ole-Christoffer
    ARTIFICIAL INTELLIGENCE XXXVIII, 2021, 13101 : 44 - 57