Learning to Regrasp Using Visual-Tactile Representation-Based Reinforcement Learning

被引:0
|
作者
Zhang, Zhuangzhuang [1 ]
Sun, Han [1 ]
Zhou, Zhenning [1 ]
Wang, Yizhao [1 ]
Huang, Huang [2 ]
Zhang, Zhinan [1 ]
Cao, Qixin [1 ]
机构
[1] Shanghai Jiao Tong Univ, State Key Lab Mech Syst & Vibrat, Shanghai 200240, Peoples R China
[2] Beijing Inst Control Engn, Beijing 100191, Peoples R China
关键词
Visualization; Force; Grasping; Training; Representation learning; Tactile sensors; Feature extraction; Stability analysis; Optimization; Hardware; Reinforcement learning; representation learning; robotic regrasp; transfer learning; visual-tactile fusion; VISION; SENSOR;
D O I
10.1109/TIM.2024.3470030
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The open-loop grasp planner, which relies on vision, is prone to failure caused by calibration errors, visual occlusions, and other factors. Additionally, it cannot adapt the grasp pose and gripping force in real time, thereby increasing the risk of potential damage to unidentified objects. This work presents a multimodal regrasp control framework based on deep reinforcement learning (RL). Given a coarse initial grasp pose, the proposed regrasping policy efficiently optimizes grasp pose and gripping force by deeply fusing visual and high-resolution tactile data in a closed-loop fashion. To enhance the sample efficiency and generalization capability of the RL algorithm, this work leverages self-supervision to pretrain a visual-tactile representation model, which serves as a feature extraction network during RL policy training. The RL policy is trained purely in simulation and successfully deployed to a real-world environment via domain adaptation and domain randomization techniques. Extensive experimental results in simulation and real-world environments indicate that the robot guided by the regrasping policy is able to achieve gentle grasping of unknown objects with high success rates. Finally, the comparison results with the state-of-the-art algorithm also demonstrate the superiority of our algorithm.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Enhancing visual reinforcement learning with State-Action Representation
    Yan, Mengbei
    Lyu, Jiafei
    Li, Xiu
    KNOWLEDGE-BASED SYSTEMS, 2024, 304
  • [32] Tensor Representation-Based Multiview Graph Contrastive Learning for IoE Intelligence
    Gao, Yuan
    Zhao, Qian
    Yang, Laurence T.
    Yang, Jing
    Yang, Jieming
    IEEE INTERNET OF THINGS JOURNAL, 2025, 12 (04): : 3482 - 3492
  • [33] Flexible Material Quality Assessment Based on Visual-Tactile Fusion
    Xu, Shuchang
    Xu, Haohao
    Mao, Fangtao
    Su, Wen
    Ji, Menghui
    Gan, Haiyong
    Yang, Wenzhen
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 : 1 - 10
  • [34] REVERSAL IN TACTILE AND VISUAL LEARNING
    BIEDERMAN, GB
    SCIENCE, 1966, 154 (3749) : 677 - +
  • [35] REVERSAL IN TACTILE AND VISUAL LEARNING
    MOFFETT, A
    ETTLINGER, G
    SCIENCE, 1966, 154 (3750) : 799 - +
  • [36] Variational Empowerment as Representation Learning for Goal-Based Reinforcement Learning
    Choi, Jongwook
    Sharma, Archit
    Lee, Honglak
    Levine, Sergey
    Gu, Shixiang Shane
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [37] Dictionary learning method for joint sparse representation-based image fusion
    Zhang, Qiheng
    Fu, Yuli
    Li, Haifeng
    Zou, Jian
    OPTICAL ENGINEERING, 2013, 52 (05)
  • [38] Web Service Classification Based on Reinforcement Learning and Structured Representation Learning
    Sheng, Hankang
    Li, Zhangbing
    Liu, Jianxun
    Zhang, Xiao
    2021 INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND BLOCKCHAIN TECHNOLOGY (AIBT 2021), 2021, : 21 - 27
  • [39] Using pupae as appetitive reinforcement to study visual and tactile associative learning in the Ponerine ant Diacamma indicum
    Chandak, Parth
    Chakraborti, Udipta
    Annagiri, Sumana
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [40] Using pupae as appetitive reinforcement to study visual and tactile associative learning in the Ponerine ant Diacamma indicum
    Parth Chandak
    Udipta Chakraborti
    Sumana Annagiri
    Scientific Reports, 13