Multi-Task Multi-Agent Reinforcement Learning via Skill Graphs

被引:0
作者
Zhu, Guobin [1 ,2 ]
Zhou, Rui [1 ]
Ji, Wenkang [2 ]
Zhang, Hongyin [3 ]
Wang, Donglin [3 ]
Zhao, Shiyu [2 ]
机构
[1] Beihang Univ, Sch Automat Sci & Elect Engn, Beijing 100191, Peoples R China
[2] Westlake Univ, Dept Artificial Intelligence, WINDY Lab, Hangzhou 310024, Peoples R China
[3] Westlake Univ, Dept Artificial Intelligence, MiLAB Lab, Hangzhou 310024, Peoples R China
来源
IEEE ROBOTICS AND AUTOMATION LETTERS | 2025年 / 10卷 / 09期
基金
中国国家自然科学基金;
关键词
Vectors; Knowledge transfer; Reinforcement learning; Multitasking; Training; Standards; Knowledge graphs; Libraries; Trajectory; Supervised learning; Multi-robot systems; multi-agent reinforcement learning; multi-task reinforcement learning;
D O I
10.1109/LRA.2025.3588784
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Multi-task multi-agent reinforcement learning (M T-MARL) has recently gained attention for its potential to enhance MARL's adaptability across multiple tasks. However, it is challenging for existing multi-task learning methods to handle complex problems, as they are unable to handle unrelated tasks and possess limited knowledge transfer capabilities. In this paper, we propose a hierarchical approach that efficiently addresses these challenges. The high-level module utilizes a skill graph, while the low-level module employs a standard MARL algorithm. Our approach offers two contributions. First, we consider the MT-MARL problem in the context of unrelated tasks, expanding the scope of MTRL. Second, the skill graph is used as the upper layer of the standard hierarchical approach, with training independent of the lower layer, effectively handling unrelated tasks and enhancing knowledge transfer capabilities. Extensive experiments are conducted to validate these advantages and demonstrate that the proposed method outperforms the latest hierarchical MAPPO algorithms. Videos and code are available at https://github.com/WindyLab/MT-MARL-SG
引用
收藏
页码:8650 / 8657
页数:8
相关论文
共 35 条
[1]  
Rusu AA, 2016, Arxiv, DOI arXiv:1511.06295
[2]  
Andreas J, 2017, PR MACH LEARN RES, V70
[3]  
Bordes A., 2013, Adv. Neural Inf. Process. Syst., P26
[4]   Learning Functionally Decomposed Hierarchies for Continuous Control Tasks With Path Planning [J].
Christen, Sammy ;
Jendele, Lukas ;
Aksan, Emre ;
Hilliges, Otmar .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02) :3623-3630
[5]   Decentralized Multi-Agent Pursuit Using Deep Reinforcement Learning [J].
de Souza, Cristino, Jr. ;
Newbury, Rhys ;
Cosgun, Akansel ;
Castillo, Pedro ;
Vidolov, Boris ;
Kulic, Dana .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (03) :4552-4559
[6]   Swarm Robotics: Past, Present, and Future [J].
Dorigo, Marco ;
Theraulaz, Guy ;
Trianni, Vito .
PROCEEDINGS OF THE IEEE, 2021, 109 (07) :1152-1165
[7]   Distributed multi-robot collision avoidance via deep reinforcement learning for navigation in complex scenarios [J].
Fan, Tingxiang ;
Long, Pinxin ;
Liu, Wenxi ;
Pan, Jia .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2020, 39 (07) :856-892
[8]   Planning-Augmented Hierarchical Reinforcement Learning [J].
Gieselmann, Robert ;
Pokorny, Florian T. .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (03) :5097-5104
[9]  
Hessel M, 2019, AAAI CONF ARTIF INTE, P3796
[10]   A Survey on Knowledge Graphs: Representation, Acquisition, and Applications [J].
Ji, Shaoxiong ;
Pan, Shirui ;
Cambria, Erik ;
Marttinen, Pekka ;
Yu, Philip S. .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (02) :494-514