Deep Robotic Grasping Prediction with Hierarchical RGB-D Fusion

被引:0
作者
Yaoxian Song
Jun Wen
Dongfang Liu
Changbin Yu
机构
[1] Fudan University,School of Engineering
[2] Westlake University,College of Computer Science and Technology
[3] Zhejiang University,Department of Computer Engineering
[4] Rochester Institute of Technology,College of Artificial Intelligence and Big Data
[5] Shandong First Medical University & Shandong Academy of Medical Sciences,undefined
来源
International Journal of Control, Automation and Systems | 2022年 / 20卷
关键词
Deep learning; depth estimation; multi-modal fusion; uncertainty quantification; vision-based grasping;
D O I
暂无
中图分类号
学科分类号
摘要
Vision-based robotic grasping is a fundamental task in robotic control. Dexterous and precise grasp control of the robotic arm is challenging and a critical technique for the manufacturing and emerging robot service industry. Current state-of-art methods adopt RGB-D images or point clouds in an attempt to obtain an accurate, robust, and real-time policy. However, most of these methods only use single modal data or ignore the uncertainty of sampling data especially the depth information. Even they leverage multi-modal data, they seldom fuse the features in different scales. All of these results in unreliable grasp prediction inevitably. In this paper, we propose a novel multi-modal neural network to predict grasps in real-time. The key idea is to fuse RGB and depth information hierarchically and quantify the uncertainty of raw depth data to re-weight the depth features. For higher grasping performance, a background extraction module and depth re-estimation module are used to reduce the influence caused by the incompletion and low-quality of the raw data. We evaluate the performance on the Cornell Grasp Dataset and provide a series of extensive experiments to demonstrate the advantages of our method on a real robot. The results indicate the superiority of our proposed method by outperforming the state-of-the-art methods significantly in all metrics.
引用
收藏
页码:243 / 254
页数:11
相关论文
共 29 条
  • [1] Bicchi A(2000)Robotic grasping and contact: A review Proc. of ICRA. Millennium Conference. IEEE International Conference on Robotics and Automation. Symposia Proceedings (Cat. No. 00CH37065) 1 348-353
  • [2] Kumar V(1985)The mechanics of manipulation Proc. of IEEE International Conference on Robotics and Automation 2 544-548
  • [3] Mason M(2015)Deep learning for detecting robotic grasps The International Journal of Robotics Research 34 705-724
  • [4] Lenz I(2013)Datadriven grasp synthesis—a survey IEEE Transactions on Robotics 30 289-309
  • [5] Lee H(2012)Physical human interactive guidance: Identifying grasping principles from human-planned grasps IEEE Transactions on Robotics 28 899-910
  • [6] Saxena A(2018)Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection The International Journal of Robotics Research 37 421-436
  • [7] Bohg J(2015)Benchmarking in manipulation research: Using the yale-cmu-berkeley object and model set IEEE Robotics & Automation Magazine 22 36-52
  • [8] Morales A(2017)RGB-D object recognition and grasp detection using hierarchical cascaded forests IEEE Transactions on Robotics 33 547-564
  • [9] Asfour T(undefined)undefined undefined undefined undefined-undefined
  • [10] Kragic D(undefined)undefined undefined undefined undefined-undefined