LEAGUE: Guided Skill Learning and Abstraction for Long-Horizon Manipulation

被引:3
作者
Cheng, Shuo [1 ]
Xu, Danfei [1 ]
机构
[1] Georgia Inst Technol, Atlanta, GA 30308 USA
关键词
Reinforcement learning; Libraries; Behavioral sciences; Silver; task and motion planning; continual learning;
D O I
10.1109/LRA.2023.3308061
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
To assist with everyday human activities, robots must solve complex long-horizon tasks and generalize to new settings. Recent deep reinforcement learning (RL) methods show promise in fully autonomous learning, but they struggle to reach long-term goals in large environments. On the other hand, Task and Motion Planning (TAMP) approaches excel at solving and generalizing across long-horizon tasks, thanks to their powerful state and action abstractions. But they assume predefined skill sets, which limits their real-world applications. In this work, we combine the benefits of these two paradigms and propose an integrated task planning and skill learning framework named LEAGUE (Learning and Abstraction with Guidance). LEAGUE leverages the symbolic interface of a task planner to guide RL-based skill learning and creates abstract state space to enable skill reuse. More importantly, LEAGUE learns manipulation skills in-situ of the task planning system, continuously growing its capability and the set of tasks that it can solve. We evaluate LEAGUE on four challenging simulated task domains and show that LEAGUE outperforms baselines by large margins. We also show that the learned skills can be reused to accelerate learning in new tasks domains and transfer to a physical robot platform.
引用
收藏
页码:6451 / 6458
页数:8
相关论文
共 29 条
  • [1] Skill Learning for Long-Horizon Sequential Tasks
    Alves, Joao
    Lau, Nuno
    Silva, Filipe
    PROGRESS IN ARTIFICIAL INTELLIGENCE, EPIA 2022, 2022, 13566 : 713 - 724
  • [2] Modular Reinforcement Learning In Long-Horizon Manipulation Tasks
    Vavrecka, Michal
    Kriz, Jonas
    Sokovnin, Nikita
    Sejnova, Gabriela
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING-ICANN 2024, PT X, 2024, 15025 : 299 - 312
  • [3] Bottom-Up Skill Discovery From Unsegmented Demonstrations for Long-Horizon Robot Manipulation
    Zhu, Yifeng
    Stone, Peter
    Zhu, Yuke
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02) : 4126 - 4133
  • [4] Generative Skill Chaining: Long-Horizon Skill Planning with Diffusion Models
    Mishra, Utkarsh A.
    Xue, Shangjie
    Chen, Yongxin
    Xu, Danfei
    CONFERENCE ON ROBOT LEARNING, VOL 229, 2023, 229
  • [5] Sequential Dexterity: Chaining Dexterous Policies for Long-Horizon Manipulation
    Chen, Yuanpei
    Wang, Chen
    Li Fei-Fei
    Liu, Karen
    CONFERENCE ON ROBOT LEARNING, VOL 229, 2023, 229
  • [6] Bimanual Long-Horizon Manipulation Via Temporal-Context Transformer RL
    Oh, Ji-Heon
    Espinoza, Ismael
    Jung, Danbi
    Kim, Tae-Seong
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (12): : 10898 - 10905
  • [7] Learning to Correct Mistakes: Backjumping in Long-Horizon Task and Motion Planning
    Sung, Yoonchang
    Wang, Zizhao
    Stone, Peter
    CONFERENCE ON ROBOT LEARNING, VOL 205, 2022, 205 : 2115 - 2124
  • [8] ERRA: An Embodied Representation and Reasoning Architecture for Long-Horizon Language-Conditioned Manipulation Tasks
    Zhao, Chao
    Yuan, Shuai
    Jiang, Chunli
    Cai, Junhao
    Yu, Hongyu
    Wang, Michael Yu
    Chen, Qifeng
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (06) : 3230 - 3237
  • [9] SyDeBO: Symbolic-Decision-Embedded Bilevel Optimization for Long-Horizon Manipulation in Dynamic Environments
    Zhao, Zhigen
    Zhou, Ziyi
    Park, Michael
    Zhao, Ye
    IEEE ACCESS, 2021, 9 : 128817 - 128826
  • [10] State-Dependent Maximum Entropy Reinforcement Learning for Robot Long-Horizon Task Learning
    Deshuai Zheng
    Jin Yan
    Tao Xue
    Yong Liu
    Journal of Intelligent & Robotic Systems, 2024, 110