Unsupervised Task Clustering for Multi-task Reinforcement Learning

被引:3
|
作者
Ackermann, Johannes [1 ]
Richter, Oliver [2 ]
Wattenhofer, Roger [2 ]
机构
[1] Tech Univ Munich, Munich, Germany
[2] Swiss Fed Inst Technol, Zurich, Switzerland
关键词
MIXTURES;
D O I
10.1007/978-3-030-86486-6_14
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Meta-learning, transfer learning and multi-task learning have recently laid a path towards more generally applicable reinforcement learning agents that are not limited to a single task. However, most existing approaches implicitly assume a uniform similarity between tasks. We argue that this assumption is limiting in settings where the relationship between tasks is unknown a-priori. In this work, we propose a general approach to automatically cluster together similar tasks during training. Our method, inspired by the expectation-maximization algorithm, succeeds at finding clusters of related tasks and uses these to improve sample complexity. We achieve this by designing an agent with multiple policies. In the expectation step, we evaluate the performance of the policies on all tasks and assign each task to the best performing policy. In the maximization step, each policy trains by sampling tasks from its assigned set. This method is intuitive, simple to implement and orthogonal to other multi-task learning algorithms. We show the generality of our approach by evaluating on simple discrete and continuous control tasks, as well as complex bipedal walker tasks and Atari games. Results show improvements in sample complexity as well as a more general applicability when compared to other approaches.
引用
收藏
页码:222 / 237
页数:16
相关论文
共 50 条
  • [31] Zero-Shot Task Generalization with Multi-Task Deep Reinforcement Learning
    Oh, Junhyuk
    Singh, Satinder
    Lee, Honglak
    Kohli, Pushmeet
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [32] Optimization of Deep Reinforcement Learning with Hybrid Multi-Task Learning
    Varghese, Nelson Vithayathil
    Mahmoud, Qusay H.
    2021 15TH ANNUAL IEEE INTERNATIONAL SYSTEMS CONFERENCE (SYSCON 2021), 2021,
  • [33] Learning potential functions and their representations for multi-task reinforcement learning
    Matthijs Snel
    Shimon Whiteson
    Autonomous Agents and Multi-Agent Systems, 2014, 28 : 637 - 681
  • [34] Learning potential functions and their representations for multi-task reinforcement learning
    Snel, Matthijs
    Whiteson, Shimon
    AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2014, 28 (04) : 637 - 681
  • [35] Multi-task subspace clustering
    Zhong, Guo
    Pun, Chi -Man
    INFORMATION SCIENCES, 2024, 661
  • [36] Episodic memory transfer for multi-task reinforcement learning
    Sorokin, Artyom Y.
    Burtsev, Mikhail S.
    BIOLOGICALLY INSPIRED COGNITIVE ARCHITECTURES, 2018, 26 : 91 - 95
  • [37] Multi-task Hierarchical Adversarial Inverse Reinforcement Learning
    Chen, Jiayu
    Tamboli, Dipesh
    Lan, Tian
    Aggarwal, Vaneet
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202
  • [38] Attributed graph clustering with multi-task embedding learning
    Zhang, Xiaotong
    Liu, Han
    Zhang, Xianchao
    Liu, Xinyue
    NEURAL NETWORKS, 2022, 152 : 224 - 233
  • [39] Distributed Clustering for Cooperative Multi-Task Learning Networks
    Li, Jiani
    Wang, Weihan
    Abbas, Waseem
    Koutsoukos, Xenofon
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2023, 10 (06): : 3933 - 3942
  • [40] Multi-task learning regression via convex clustering
    Okazaki, Akira
    Kawano, Shuichi
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2024, 195