Grasp Pose Learning from Human Demonstration with Task Constraints

被引:0
作者
Yinghui Liu
Kun Qian
Xin Xu
Bo Zhou
Fang Fang
机构
[1] Southeast University,School of Automation
[2] Southeast University,The Key Laboratory of Measurement and Control of CSE, Ministry of Education
来源
Journal of Intelligent & Robotic Systems | 2022年 / 105卷
关键词
Learning from demonstration; Robot grasping; Grasp pose detection; Superquadric; Task constraints;
D O I
暂无
中图分类号
学科分类号
摘要
To learn grasp constraints from human demonstrations, we propose a method that combines data-driven grasp constraint learning and one-shot human demonstration of tasks. By presenting task constraints in a GMM-based gripper-independent form, the task constraints are learned from simulated data with self-labeled grasp quality scores. By observing a human demonstration of the task and a real-world object, the learned task constraint model can be utilized to infer both the unknown grasping task and the probability density distributions of the task constraints on the object point cloud. In addition, we extend the superquadric-based grasp estimation method for reproducing the grasping task with 2-finger grippers. The task constraints restrict the searching scope of the grasp pose, so the geometrically best grasp pose within the task-constrained regions can be obtained. The effectiveness of our methodology is verified in experiments with a UR5 robot with a 2-finger gripper.
引用
收藏
相关论文
共 50 条
  • [31] iTP-LfD: Improved task parametrised learning from demonstration for adaptive path generation of cobot
    El Zaatari, Shirine
    Wang, Yuqi
    Li, Weidong
    Peng, Yiqun
    ROBOTICS AND COMPUTER-INTEGRATED MANUFACTURING, 2021, 69
  • [32] A learning from demonstration framework for adaptive task and motion planning in varying package-to-order scenarios
    Ma, Ruidong
    Chen, Jingyu
    Oyekan, John
    ROBOTICS AND COMPUTER-INTEGRATED MANUFACTURING, 2023, 82
  • [33] Fuzzy dynamical system for robot learning motion skills from human demonstration
    Teng, Tao
    Gatti, Matteo
    Poni, Stefano
    Caldwell, Darwin
    Chen, Fei
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2023, 164
  • [34] ROBOT LEARNING FROM HUMAN DEMONSTRATION OF ACTIVITIES OF DAILY LIVING (ADL) TASKS
    Trivedi, Urvish
    Alqasemi, Redwan
    Dubey, Rajiv
    PROCEEDINGS OF ASME 2021 INTERNATIONAL MECHANICAL ENGINEERING CONGRESS AND EXPOSITION (IMECE2021), VOL 6, 2021,
  • [35] Learning Gentle Grasping From Human-Free Force Control Demonstration
    Li, Mingxuan
    Zhang, Lunwei
    Li, Tiemin
    Jiang, Yao
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2025, 10 (03): : 2391 - 2398
  • [36] The Effects of a Robot's Performance on Human Teachers for Learning from Demonstration Tasks
    Hedlund, Erin
    Johnson, Michael
    Gombolay, Matthew
    2021 16TH ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, HRI, 2021, : 207 - 215
  • [37] Robot Performing Peg-in-Hole Operations by Learning from Human Demonstration
    Zhu, Zuyuan
    Hu, Huosheng
    Gu, Dongbing
    2018 10TH COMPUTER SCIENCE AND ELECTRONIC ENGINEERING CONFERENCE (CEEC), 2018, : 30 - 35
  • [38] Learning efficient push and grasp policy in a totebox from simulation
    Ni, Peiyuan
    Zhang, Wenguang
    Zhang, Haoruo
    Cao, Qixin
    ADVANCED ROBOTICS, 2020, 34 (13) : 873 - 887
  • [39] Learning generalizable behaviors from demonstration
    Rivera, Corban
    Popek, Katie M.
    Ashcraft, Chace
    Staley, Edward W.
    Katyal, Kapil D.
    Paulhamus, Bart L.
    FRONTIERS IN NEUROROBOTICS, 2022, 16
  • [40] Learning motion primitives from demonstration
    Chi, Mingshan
    Yao, Yufeng
    Liu, Yaxin
    Teng, Yiqian
    Zhong, Ming
    ADVANCES IN MECHANICAL ENGINEERING, 2017, 9 (12)