Human-In-The-Loop Task and Motion Planning for Imitation Learning

被引:0
|
作者
Mandlekar, Ajay [1 ]
Garrett, Caelan [1 ]
Xu, Danfei [1 ,2 ]
Fox, Dieter [1 ]
机构
[1] NVIDIA, Santa Clara, CA 95051 USA
[2] Georgia Inst Technol, Atlanta, GA USA
来源
CONFERENCE ON ROBOT LEARNING, VOL 229 | 2023年 / 229卷
关键词
Imitation Learning; Task and Motion Planning; Teleoperation; SKILLS; MODELS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Imitation learning from human demonstrations can teach robots complex manipulation skills, but is time-consuming and labor intensive. In contrast, Task and Motion Planning (TAMP) systems are automated and excel at solving long-horizon tasks, but they are difficult to apply to contact-rich tasks. In this paper, we present Human-in-the-Loop Task and Motion Planning (HITL-TAMP), a novel system that leverages the benefits of both approaches. The system employs a TAMP-gated control mechanism, which selectively gives and takes control to and from a human teleoperator. This enables the human teleoperator to manage a fleet of robots, maximizing data collection efficiency. The collected human data is then combined with an imitation learning framework to train a TAMPgated policy, leading to superior performance compared to training on full task demonstrations. We compared HITL-TAMP to a conventional teleoperation system - users gathered more than 3x the number of demos given the same time budget. Furthermore, proficient agents (75%+ success) could be trained from just 10 minutes of non-expert teleoperation data. Finally, we collected 2.1K demos with HITL-TAMP across 12 contact-rich, long-horizon tasks and show that the system often produces near-perfect agents. Videos and additional results at https://hitltamp.github.io.
引用
收藏
页数:31
相关论文
共 50 条
  • [1] Human-in-the-loop Multi-task Tracking Improved by Interactive Learning
    Wen, Xupeng
    Wang, Chang
    Zhu, Yuting
    Niu, Yifeng
    Wu, Lizhen
    Yin, Dong
    2018 CHINESE AUTOMATION CONGRESS (CAC), 2018, : 2289 - 2294
  • [2] PACE: Learning Effective Task Decomposition for Human-in-the-loop Healthcare Delivery
    Zheng, Kaiping
    Chen, Gang
    Herschel, Melanie
    Ngiam, Kee Yuan
    Ooi, Beng Chin
    Gao, Jinyang
    SIGMOD '21: PROCEEDINGS OF THE 2021 INTERNATIONAL CONFERENCE ON MANAGEMENT OF DATA, 2021, : 2156 - 2168
  • [3] Human-in-the-loop Reinforcement Learning
    Liang, Huanghuang
    Yang, Lu
    Cheng, Hong
    Tu, Wenzhe
    Xu, Mengjie
    2017 CHINESE AUTOMATION CONGRESS (CAC), 2017, : 4511 - 4518
  • [4] A Hybrid Human-in-the-Loop Deep Reinforcement Learning Method for UAV Motion Planning for Long Trajectories with Unpredictable Obstacles
    Zhang, Sitong
    Li, Yibing
    Ye, Fang
    Geng, Xiaoyu
    Zhou, Zitao
    Shi, Tuo
    DRONES, 2023, 7 (05)
  • [5] Human-In-The-Loop Control and Task Learning for Pneumatically Actuated Muscle Based Robots
    Teramae, Tatsuya
    Ishihara, Koji
    Bahic, Jan
    Morimoto, Jun
    Oztop, Erhan
    FRONTIERS IN NEUROROBOTICS, 2018, 12
  • [6] Trust-Based Multi-Robot Symbolic Motion Planning with a Human-in-the-Loop
    Wang, Yue
    Humphrey, Laura R.
    Liao, Zhanrui
    Zheng, Huanfei
    ACM TRANSACTIONS ON INTERACTIVE INTELLIGENT SYSTEMS, 2018, 8 (04)
  • [7] RUNTIME VERIFICATION OF TRUST-BASED SYMBOLIC ROBOT MOTION PLANNING WITH HUMAN-IN-THE-LOOP
    Mahani, Maziar Fooladi
    Wang, Yue
    PROCEEDINGS OF THE ASME 9TH ANNUAL DYNAMIC SYSTEMS AND CONTROL CONFERENCE, 2016, VOL 2, 2017,
  • [8] A survey of human-in-the-loop for machine learning
    Wu, Xingjiao
    Xiao, Luwei
    Sun, Yixuan
    Zhang, Junhang
    Ma, Tianlong
    He, Liang
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2022, 135 : 364 - 381
  • [9] Human-in-the-loop Applied Machine Learning
    Brodley, Carla E.
    2017 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2017, : 1 - 1
  • [10] Robot Task Control Utilizing Human-in-the-loop Perception
    Yu, Wonpil
    Lee, Jae-Yeong
    Chae, Heesung
    Han, Kyuseo
    Lee, Yucheol
    Jang, Minsu
    2008 17TH IEEE INTERNATIONAL SYMPOSIUM ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION, VOLS 1 AND 2, 2008, : 395 - 400