Unsupervised Task Clustering for Multi-task Reinforcement Learning

被引:3
|
作者
Ackermann, Johannes [1 ]
Richter, Oliver [2 ]
Wattenhofer, Roger [2 ]
机构
[1] Tech Univ Munich, Munich, Germany
[2] Swiss Fed Inst Technol, Zurich, Switzerland
关键词
MIXTURES;
D O I
10.1007/978-3-030-86486-6_14
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Meta-learning, transfer learning and multi-task learning have recently laid a path towards more generally applicable reinforcement learning agents that are not limited to a single task. However, most existing approaches implicitly assume a uniform similarity between tasks. We argue that this assumption is limiting in settings where the relationship between tasks is unknown a-priori. In this work, we propose a general approach to automatically cluster together similar tasks during training. Our method, inspired by the expectation-maximization algorithm, succeeds at finding clusters of related tasks and uses these to improve sample complexity. We achieve this by designing an agent with multiple policies. In the expectation step, we evaluate the performance of the policies on all tasks and assign each task to the best performing policy. In the maximization step, each policy trains by sampling tasks from its assigned set. This method is intuitive, simple to implement and orthogonal to other multi-task learning algorithms. We show the generality of our approach by evaluating on simple discrete and continuous control tasks, as well as complex bipedal walker tasks and Atari games. Results show improvements in sample complexity as well as a more general applicability when compared to other approaches.
引用
收藏
页码:222 / 237
页数:16
相关论文
共 50 条
  • [21] Multi-Task Clustering with Model Relation Learning
    Zhang, Xiaotong
    Zhang, Xianchao
    Liu, Han
    Luo, Jiebo
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 3132 - 3140
  • [22] Skills Regularized Task Decomposition for Multi-task Offline Reinforcement Learning
    Yoo, Minjong
    Cho, Sangwoo
    Woo, Honguk
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [23] Multi-task Deep Reinforcement Learning for Scalable Parallel Task Scheduling
    Zhang, Lingxin
    Qi, Qi
    Wang, Jingyu
    Sun, Haifeng
    Liao, Jianxin
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 2992 - 3001
  • [24] Projected Task-Specific Layers for Multi-Task Reinforcement Learning
    Roberts, Josselin Somerville
    Di, Julia
    2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA 2024, 2024, : 2887 - 2893
  • [25] Unsupervised Reinforcement Learning for Multi-Task Autonomous Driving: Expanding Skills and Cultivating Curiosity
    Ma, Zhenyu
    Liu, Xinyi
    Huang, Yanjun
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (10) : 14209 - 14219
  • [26] Unsupervised Multi-task Learning with Hierarchical Data Structure
    Cao, Wenming
    Qian, Sheng
    Wu, Si
    Wong, Hau-San
    PATTERN RECOGNITION, 2019, 86 : 248 - 264
  • [27] Unsupervised learning of multi-task deep variational model
    Tan, Lu
    Li, Ling
    Liu, Wan-Quan
    An, Sen-Jian
    Munyard, Kylie
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2022, 87
  • [28] Unsupervised Multi-Task Feature Learning on Point Clouds
    Hassani, Kaveh
    Haley, Mike
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 8159 - 8170
  • [29] Multi-task gradient descent for multi-task learning
    Lu Bai
    Yew-Soon Ong
    Tiantian He
    Abhishek Gupta
    Memetic Computing, 2020, 12 : 355 - 369
  • [30] Multi-task gradient descent for multi-task learning
    Bai, Lu
    Ong, Yew-Soon
    He, Tiantian
    Gupta, Abhishek
    MEMETIC COMPUTING, 2020, 12 (04) : 355 - 369