In-Hand Object Classification and Pose Estimation With Sim-to-Real Tactile Transfer for Robotic Manipulation

被引:4
作者
Yang, Sanghoon [1 ]
Kim, Won Dong [1 ]
Park, Hyunkyu [1 ]
Min, Seojung [1 ]
Han, Hyonyoung [2 ]
Kim, Jung [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Dept Mech Engn, Daejeon 34141, South Korea
[2] ETRI, Elect & Telecommun Res Inst, Daejeon 34129, South Korea
关键词
Perception for grasping and manipulation; deep learning in grasping and manipulation; transfer learning; IDENTIFYING OBJECTS; HAPTIC EXPLORATION;
D O I
10.1109/LRA.2023.3334971
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Dexterous robotic grasping gains great benefits from tactile sensation, but delicate object exploration by tactile information is challenged by difficulty in rich and efficient data production. In this letter, we propose a tactile-based in-hand object perception approach, empowered by a sim-to-real approach toward a data-efficient learning process. A high-fidelity tactile input, measured by a pair of vision-based tactile sensors, was represented as a point cloud facilitating dual functionality of object classification and the associated pose estimation. For the classification, we constructed PoinTacNet, a variation of PointNet to fit into tactile data processing. A reliable simulation methodology on tactile input was employed for the pretraining of the model, transferred to the fine-tuning process facing real tactile data. Taking inspiration from human behaviors, a re-grasping strategy was imparted by means of conditional accumulation of class likelihood distribution. The result of the framework facilitates a high object classification accuracy of 83.89$\%$ on the ten objects from McMaster-Carr's CAD models, which is significantly improved by the re-grasping. In addition, a set of benchmarks displays the computational efficiency in the sim-to-real transfer. In line with the successful classification, the posture of in-hand objects is estimated using point cloud registration algorithms, achieving an average angular and translational RMSE of 5.03$<^>\circ$ and 2.41 mm, respectively. The proposed approach has the potential to enable robots to attain human-like haptic exploration skills for perceiving unstructured environments.
引用
收藏
页码:659 / 666
页数:8
相关论文
共 31 条
  • [1] Allen PK, 1997, IEEE INT CONF ROBOT, P676, DOI 10.1109/ROBOT.1997.620114
  • [2] Tac2Pose: Tactile object pose estimation from the first touch
    Bauza, Maria
    Bronars, Antonia
    Rodriguez, Alberto
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2023, 42 (13) : 1185 - 1209
  • [3] Bauza M, 2019, IEEE INT CONF ROBOT, P3811, DOI [10.1109/ICRA.2019.8794298, 10.1109/icra.2019.8794298]
  • [4] Shape Completion using 3D-Encoder-Predictor CNNs and Shape Synthesis
    Dai, Angela
    Qi, Charles Ruizhongtai
    Niessner, Matthias
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 6545 - 6554
  • [5] VisuoTactile 6D Pose Estimation of an In-Hand Object Using Vision and Tactile Sensor Data
    Dikhale, Snehal
    Patel, Karankumar
    Dhingra, Daksh
    Naramura, Itoshi
    Hayashi, Akinobu
    Iba, Soshi
    Jamali, Nawid
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02) : 2148 - 2155
  • [6] Learning efficient haptic shape exploration with a rigid tactile sensor array
    Fleer, Sascha
    Moringen, Alexandra
    Klatzky, Roberta L.
    Ritter, Helge
    [J]. PLOS ONE, 2020, 15 (01):
  • [7] FilterReg: Robust and Efficient Probabilistic Point-Set Registration using Gaussian Filter and Twist Parameterization
    Gao, Wei
    Tedrake, Russ
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 11087 - 11096
  • [8] Jamali N, 2016, IEEE-RAS INT C HUMAN, P179, DOI 10.1109/HUMANOIDS.2016.7803275
  • [9] Coding and use of tactile signals from the fingertips in object manipulation tasks
    Johansson, Roland S.
    Flanagan, J. Randall
    [J]. NATURE REVIEWS NEUROSCIENCE, 2009, 10 (05) : 345 - 359
  • [10] Johnson MK, 2009, PROC CVPR IEEE, P1070, DOI 10.1109/CVPRW.2009.5206534