Learning to Regrasp Using Visual-Tactile Representation-Based Reinforcement Learning

被引:0
作者
Zhang, Zhuangzhuang [1 ]
Sun, Han [1 ]
Zhou, Zhenning [1 ]
Wang, Yizhao [1 ]
Huang, Huang [2 ]
Zhang, Zhinan [1 ]
Cao, Qixin [1 ]
机构
[1] Shanghai Jiao Tong Univ, State Key Lab Mech Syst & Vibrat, Shanghai 200240, Peoples R China
[2] Beijing Inst Control Engn, Beijing 100191, Peoples R China
关键词
Visualization; Force; Grasping; Training; Representation learning; Tactile sensors; Feature extraction; Stability analysis; Optimization; Hardware; Reinforcement learning; representation learning; robotic regrasp; transfer learning; visual-tactile fusion; VISION; SENSOR;
D O I
10.1109/TIM.2024.3470030
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The open-loop grasp planner, which relies on vision, is prone to failure caused by calibration errors, visual occlusions, and other factors. Additionally, it cannot adapt the grasp pose and gripping force in real time, thereby increasing the risk of potential damage to unidentified objects. This work presents a multimodal regrasp control framework based on deep reinforcement learning (RL). Given a coarse initial grasp pose, the proposed regrasping policy efficiently optimizes grasp pose and gripping force by deeply fusing visual and high-resolution tactile data in a closed-loop fashion. To enhance the sample efficiency and generalization capability of the RL algorithm, this work leverages self-supervision to pretrain a visual-tactile representation model, which serves as a feature extraction network during RL policy training. The RL policy is trained purely in simulation and successfully deployed to a real-world environment via domain adaptation and domain randomization techniques. Extensive experimental results in simulation and real-world environments indicate that the robot guided by the regrasping policy is able to achieve gentle grasping of unknown objects with high success rates. Finally, the comparison results with the state-of-the-art algorithm also demonstrate the superiority of our algorithm.
引用
收藏
页数:11
相关论文
共 39 条
  • [1] More Than a Feeling: Learning to Grasp and Regrasp Using Vision and Touch
    Calandra, Roberto
    Owens, Andrew
    Jayaraman, Dinesh
    Lin, Justin
    Yuan, Wenzhen
    Malik, Jitendra
    Adelson, Edward H.
    Levine, Sergey
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04): : 3300 - 3307
  • [2] Yale-CMU-Berkeley dataset for robotic manipulation research
    Calli, Berk
    Singh, Arjun
    Bruce, James
    Walsman, Aaron
    Konolige, Kurt
    Srinivasa, Siddhartha
    Abbeel, Pieter
    Dollar, Aaron M.
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2017, 36 (03) : 261 - 268
  • [3] Chebotar Y, 2016, 2016 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2016), P1960, DOI 10.1109/IROS.2016.7759309
  • [4] de Bruin Tim, 2018, IEEE Robotics and Automation Letters, V3, P1394, DOI 10.1109/LRA.2018.2800101
  • [5] Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
  • [6] A review on reinforcement learning for contact-rich robotic manipulation tasks
    Elguea-Aguinaco, Inigo
    Serrano-Munoz, Antonio
    Chrysostomou, Dimitrios
    Inziarte-Hidalgo, Ibai
    Bogh, Simon
    Arana-Arexolaleiba, Nestor
    [J]. ROBOTICS AND COMPUTER-INTEGRATED MANUFACTURING, 2023, 81
  • [7] Self-Supervised Representation Learning: Introduction, advances, and challenges
    Ericsson, Linus
    Gouk, Henry
    Loy, Chen Change
    Hospedales, Timothy M.
    [J]. IEEE SIGNAL PROCESSING MAGAZINE, 2022, 39 (03) : 42 - 62
  • [8] Feng Q, 2020, IEEE INT CONF ROBOT, P610, DOI [10.1109/icra40945.2020.9196815, 10.1109/ICRA40945.2020.9196815]
  • [9] Generalization in Reinforcement Learning by Soft Data Augmentation
    Hansen, Nicklas
    Wang, Xiaolong
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 13611 - 13617
  • [10] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778