LEAGUE: Guided Skill Learning and Abstraction for Long-Horizon Manipulation

被引:6
作者
Cheng, Shuo [1 ]
Xu, Danfei [1 ]
机构
[1] Georgia Inst Technol, Atlanta, GA 30308 USA
关键词
Reinforcement learning; Libraries; Behavioral sciences; Silver; task and motion planning; continual learning;
D O I
10.1109/LRA.2023.3308061
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
To assist with everyday human activities, robots must solve complex long-horizon tasks and generalize to new settings. Recent deep reinforcement learning (RL) methods show promise in fully autonomous learning, but they struggle to reach long-term goals in large environments. On the other hand, Task and Motion Planning (TAMP) approaches excel at solving and generalizing across long-horizon tasks, thanks to their powerful state and action abstractions. But they assume predefined skill sets, which limits their real-world applications. In this work, we combine the benefits of these two paradigms and propose an integrated task planning and skill learning framework named LEAGUE (Learning and Abstraction with Guidance). LEAGUE leverages the symbolic interface of a task planner to guide RL-based skill learning and creates abstract state space to enable skill reuse. More importantly, LEAGUE learns manipulation skills in-situ of the task planning system, continuously growing its capability and the set of tasks that it can solve. We evaluate LEAGUE on four challenging simulated task domains and show that LEAGUE outperforms baselines by large margins. We also show that the learned skills can be reused to accelerate learning in new tasks domains and transfer to a physical robot platform.
引用
收藏
页码:6451 / 6458
页数:8
相关论文
共 32 条
[21]   To imitate or not to imitate: Boosting reinforcement learning-based construction robotic control for long-horizon tasks using virtual demonstrations [J].
Huang, Lei ;
Zhu, Zihan ;
Zou, Zhengbo .
AUTOMATION IN CONSTRUCTION, 2023, 146
[22]   Multi-State-Space Reasoning Reinforcement Learning for Long-Horizon RFID-Based Robotic Searching and Planning Tasks [J].
Yu Z. ;
Zhang J. ;
Mao S. ;
Periaswamy S.C.G. ;
Patton J. .
Journal of Communications and Information Networks, 2022, 7 (03) :239-251
[23]   Unsupervised Reinforcement Learning for Transferable Manipulation Skill Discovery [J].
Cho, Daesol ;
Kim, Jigang ;
Kim, H. Jin .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (03) :7455-7462
[24]   Learn-Gen-Plan: Bridging the Gap Between Vision Language Models and Real-World Long-Horizon Dexterous Manipulations [J].
Hao, Peng ;
Cui, Shaowei ;
Wei, Junhang ;
Lu, Tao ;
Cai, Yinghao ;
Wang, Shuo .
IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2025, 22 :15638-15649
[25]   Variable Impedance Skill Learning for Contact-Rich Manipulation [J].
Yang, Quantao ;
Durr, Alexander ;
Topp, Elin Anna ;
Stork, Johannes A. ;
Stoyanov, Todor .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (03) :8391-8398
[26]   A Reinforcement Learning-Based Framework for Robot Manipulation Skill Acquisition [J].
Liu, Dong ;
Wang, Zitu ;
Lu, Binpeng ;
Cong, Ming ;
Yu, Honghua ;
Zou, Qiang .
IEEE ACCESS, 2020, 8 :108429-108437
[27]   Human skill knowledge guided global trajectory policy reinforcement learning method [J].
Zang, Yajing ;
Wang, Pengfei ;
Zha, Fusheng ;
Guo, Wei ;
Li, Chuanfeng ;
Sun, Lining .
FRONTIERS IN NEUROROBOTICS, 2024, 18
[28]   Skill generalization of tubular object manipulation with tactile sensing and Sim2Real learning [J].
Zhao, Yongqiang ;
Jing, Xingshuo ;
Qian, Kun ;
Gomes, Daniel Fernandes ;
Luo, Shan .
ROBOTICS AND AUTONOMOUS SYSTEMS, 2023, 160
[29]   Reinforcement Learning-Guided Long-Timescale Simulation of Hydrogen Transport in Metals [J].
Tang, Hao ;
Li, Boning ;
Song, Yixuan ;
Liu, Mengren ;
Xu, Haowei ;
Wang, Guoqing ;
Chung, Heejung ;
Li, Ju .
ADVANCED SCIENCE, 2024, 11 (05)
[30]   Overcoming the Long Horizon Barrier for Sample-Efficient Reinforcement Learning with Latent Low-Rank Structure [J].
Sam, Tyler ;
Chen, Yudong ;
Yu, Christina Lee .
PROCEEDINGS OF THE ACM ON MEASUREMENT AND ANALYSIS OF COMPUTING SYSTEMS, 2023, 7 (02)