Tactile-Visual Fusion Based Robotic Grasp Detection Method with a Reproducible Sensor

被引:0
作者
Song, Yaoxian [1 ,2 ,3 ]
Luo, Yun [3 ]
Yu, Changbin [4 ,5 ,6 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai, Peoples R China
[2] Fudan Univ, Inst Intelligent Robots, Shanghai, Peoples R China
[3] Westlake Univ, Sch Engn, Hangzhou, Peoples R China
[4] Shandong First Med Univ, Coll Artificial Intelligence & Big Data, Tai An, Shandong, Peoples R China
[5] Shandong Acad Med Sci, Tai An, Shandong, Peoples R China
[6] Univ Johannesburg, Fac Engn & Built Environm, Johannesburg, South Africa
关键词
Tactile sensor; Tactile-visual dataset; Multi-modal fusion; Deep learning; Grasp detection; VISION; MODEL;
D O I
10.2991/ijcis.d.210531.001;
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Robotic grasp detection is a fundamental problem in robotic manipulation. The conventional grasp methods, using vision information only, can cause potential damage in force-sensitive tasks. In this paper, we propose a tactile-visual based method using a reproducible sensor to realize a fine-grained and haptic grasping. Although there exist several tactile-based methods, they require expensive custom sensors in coordination with their specific datasets. In order to overcome the limitations, we introduce a low-cost and reproducible tactile fingertip and build a general tactile-visual fusion grasp dataset including 5,110 grasping trials. We further propose a hierarchical encoder-decoder neural network to predict grasp points and force in an end-to-end manner. Then comparisons of our method with the state-of-the-art methods in the benchmark are shown both in vision-based and tactile-visual fusion schemes, and our method outperforms in most scenarios. Furthermore, we also compare our fusion method with the only vision-based method in the physical experiment, and the results indicate that our end-to-end method empowers the robot with a more fine-grained grasp ability, reducing force redundancy by 41%. Our project is available at https://sites.google.com/view/tvgd (c) 2021 The Authors. Published by Atlantis Press B.V.
引用
收藏
页码:1753 / 1762
页数:10
相关论文
共 24 条
[1]   Using a Deep Learning Model on Images to Obtain a 2D Laser People Detector for a Mobile Robot [J].
Aguirre, Eugenio ;
Garcia-Silvente, Miguel .
INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE SYSTEMS, 2019, 12 (02) :476-484
[2]   Surface roughness and sliding friction [J].
Bikerman, JJ .
REVIEWS OF MODERN PHYSICS, 1944, 16 (01) :0053-0068
[3]  
Bin Fang, 2019, 2019 IEEE International Conference on Robotics and Biomimetics (ROBIO), P203, DOI 10.1109/ROBIO49542.2019.8961671
[4]  
Calandra R., 2017, ANN C ROBOT LEARNING, V78, P314
[5]   More Than a Feeling: Learning to Grasp and Regrasp Using Vision and Touch [J].
Calandra, Roberto ;
Owens, Andrew ;
Jayaraman, Dinesh ;
Lin, Justin ;
Yuan, Wenzhen ;
Malik, Jitendra ;
Adelson, Edward H. ;
Levine, Sergey .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04) :3300-3307
[6]  
Depierre A, 2018, IEEE INT C INT ROBOT, P3511, DOI 10.1109/IROS.2018.8593950
[7]  
Detry R, 2013, IEEE INT CONF ROBOT, P601, DOI 10.1109/ICRA.2013.6630635
[8]  
Fang B, 2018, IEEE INT CONF ROBOT, P4740
[9]  
Hanbo Zhang, 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), P4768, DOI 10.1109/IROS40897.2019.8967869
[10]  
Hara K., 2017, ARXIV PREPRINT ARXIV