Vision-based robotic grasping from object localization, object pose estimation to grasp estimation for parallel grippers: a review

被引:288
作者
Du, Guoguang [1 ]
Wang, Kai [1 ]
Lian, Shiguo [1 ]
Zhao, Kaiyong [1 ]
机构
[1] CloudMinds Technol Inc, Beijing, Peoples R China
关键词
Robotic grasping; Object localization; Object pose estimation; Grasp estimation; SALIENT; NETWORK; SEGMENTATION; REGISTRATION; RECOGNITION; HISTOGRAMS; VERSATILE; BENCHMARK; SURFACE;
D O I
10.1007/s10462-020-09888-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents a comprehensive survey on vision-based robotic grasping. We conclude three key tasks during vision-based robotic grasping, which are object localization, object pose estimation and grasp estimation. In detail, the object localization task contains object localization without classification, object detection and object instance segmentation. This task provides the regions of the target object in the input data. The object pose estimation task mainly refers to estimating the 6D object pose and includes correspondence-based methods, template-based methods and voting-based methods, which affords the generation of grasp poses for known objects. The grasp estimation task includes 2D planar grasp methods and 6DoF grasp methods, where the former is constrained to grasp from one direction. These three tasks could accomplish the robotic grasping with different combinations. Lots of object pose estimation methods need not object localization, and they conduct object localization and object pose estimation jointly. Lots of grasp estimation methods need not object localization and object pose estimation, and they conduct grasp estimation in an end-to-end manner. Both traditional methods and latest deep learning-based methods based on the RGB-D image inputs are reviewed elaborately in this survey. Related datasets and comparisons between state-of-the-art methods are summarized as well. In addition, challenges about vision-based robotic grasping and future directions in addressing these challenges are also pointed out.
引用
收藏
页码:1677 / 1734
页数:58
相关论文
共 324 条
[1]  
Akkaya I., 2019, SOLVING RUBIKS CUBE
[2]  
Aldoma A, 2011, 2011 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCV WORKSHOPS)
[3]  
[Anonymous], 2005, ISPRS Workshop on Laser Scanning, DOI 10.1.1.118.1736
[4]  
[Anonymous], 2018, ARXIV180210367
[5]   PointNetLK: Robust & Efficient Point Cloud Registration using PointNet [J].
Aoki, Yasuhiro ;
Goforth, Hunter ;
Srivatsan, Rangaprasad Arun ;
Lucey, Simon .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :7156-7165
[6]   Learning Grasp Affordance Reasoning Through Semantic Relations [J].
Ardon, Paola ;
Pairet, Eric ;
Petrick, Ronald P. A. ;
Ramamoorthy, Subramanian ;
Lohan, Katrin S. .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2019, 4 (04) :4571-4578
[7]  
Asir U, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P4875
[8]   Speeded-Up Robust Features (SURF) [J].
Bay, Herbert ;
Ess, Andreas ;
Tuytelaars, Tinne ;
Van Gool, Luc .
COMPUTER VISION AND IMAGE UNDERSTANDING, 2008, 110 (03) :346-359
[9]   Projections of ipRGCs and conventional RGCs to retinorecipient brain nuclei [J].
Beier, Corinne ;
Zhang, Ze ;
Yurgel, Maria ;
Hattar, Samer .
JOURNAL OF COMPARATIVE NEUROLOGY, 2021, 529 (08) :1863-1875
[10]  
Bellekens Ben, 2014, AMBIENT 2014 4 INT C, P8