Learning to Regrasp Using Visual-Tactile Representation-Based Reinforcement Learning

被引:0
|
作者
Zhang, Zhuangzhuang [1 ]
Sun, Han [1 ]
Zhou, Zhenning [1 ]
Wang, Yizhao [1 ]
Huang, Huang [2 ]
Zhang, Zhinan [1 ]
Cao, Qixin [1 ]
机构
[1] Shanghai Jiao Tong Univ, State Key Lab Mech Syst & Vibrat, Shanghai 200240, Peoples R China
[2] Beijing Inst Control Engn, Beijing 100191, Peoples R China
关键词
Visualization; Force; Grasping; Training; Representation learning; Tactile sensors; Feature extraction; Stability analysis; Optimization; Hardware; Reinforcement learning; representation learning; robotic regrasp; transfer learning; visual-tactile fusion; VISION; SENSOR;
D O I
10.1109/TIM.2024.3470030
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The open-loop grasp planner, which relies on vision, is prone to failure caused by calibration errors, visual occlusions, and other factors. Additionally, it cannot adapt the grasp pose and gripping force in real time, thereby increasing the risk of potential damage to unidentified objects. This work presents a multimodal regrasp control framework based on deep reinforcement learning (RL). Given a coarse initial grasp pose, the proposed regrasping policy efficiently optimizes grasp pose and gripping force by deeply fusing visual and high-resolution tactile data in a closed-loop fashion. To enhance the sample efficiency and generalization capability of the RL algorithm, this work leverages self-supervision to pretrain a visual-tactile representation model, which serves as a feature extraction network during RL policy training. The RL policy is trained purely in simulation and successfully deployed to a real-world environment via domain adaptation and domain randomization techniques. Extensive experimental results in simulation and real-world environments indicate that the robot guided by the regrasping policy is able to achieve gentle grasping of unknown objects with high success rates. Finally, the comparison results with the state-of-the-art algorithm also demonstrate the superiority of our algorithm.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Visual-Tactile Multimodality for Following Deformable Linear Objects Using Reinforcement Learning
    Pecyna, Leszek
    Dong, Siyuan
    Luo, Shan
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 3987 - 3994
  • [2] Learning cross-modal visual-tactile representation using ensembled generative adversarial networks
    Li, Xinwu
    Liu, Huaping
    Zhou, Junfeng
    Sun, FuChun
    COGNITIVE COMPUTATION AND SYSTEMS, 2019, 1 (02) : 40 - 44
  • [3] Lifelong robotic visual-tactile perception learning
    Dong, Jiahua
    Cong, Yang
    Sun, Gan
    Zhang, Tao
    PATTERN RECOGNITION, 2022, 121
  • [4] Representation-Based Robustness in Goal-Conditioned Reinforcement Learning
    Yin, Xiangyu
    Wu, Sihao
    Liu, Jiaxu
    Fang, Meng
    Zhao, Xingyu
    Huang, Xiaowei
    Ruan, Wenjie
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 19, 2024, : 21761 - 21769
  • [5] Visual-Tactile Fused Graph Learning for Object Clustering
    Zhang, Tao
    Cong, Yang
    Sun, Gan
    Dong, Jiahua
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (11) : 12275 - 12289
  • [6] Partial Visual-Tactile Fused Learning for Robotic Object Recognition
    Zhang, Tao
    Cong, Yang
    Dong, Jiahua
    Hou, Dongdong
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2022, 52 (07): : 4349 - 4361
  • [7] Event-Driven Visual-Tactile Sensing and Learning for Robots
    Taunyazov, Tasbolat
    Sng, Weicong
    See, Hian Hian
    Lim, Brian
    Kuan, Jethro
    Ansari, Abdul Fatir
    Tee, Benjamin C. K.
    Soh, Harold
    ROBOTICS: SCIENCE AND SYSTEMS XVI, 2020,
  • [8] VISUAL-TACTILE AND TACTILE-TACTILE PAIRED-ASSOCIATE LEARNING BY NORMAL AND POOR READERS
    STEGER, JA
    VELLUTIN.FR
    MESHOULA.U
    PERCEPTUAL AND MOTOR SKILLS, 1972, 35 (01) : 263 - &
  • [9] Multimodal Visual-Tactile Representation Learning through Self-Supervised Contrastive Pre-Training
    Dave, Vedant
    Lygerakis, Fotios
    Rueckert, Elmar
    2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA 2024, 2024, : 8013 - 8020
  • [10] Self-Attention Based Visual-Tactile Fusion Learning for Predicting Grasp Outcomes
    Cui, Shaowei
    Wang, Rui
    Wei, Junhang
    Hu, Jingyi
    Wang, Shuo
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04) : 5827 - 5834