A Vision-Based Robot Grasping System

被引:35
作者
Cheng, Hu [1 ]
Wang, Yingying [1 ]
Meng, Max Q-H [2 ,3 ,4 ]
机构
[1] Chinese Univ Hong Kong, Dept Elect Engn, Robot Percept & Artificial Intelligence Lab, Hong Kong, Peoples R China
[2] Southern Univ Sci & Technol, Shenzhen Key Lab Robot Percept & Intelligence, Shenzhen 518055, Peoples R China
[3] Southern Univ Sci & Technol, Dept Elect & Elect Engn, Shenzhen 518055, Peoples R China
[4] Chinese Univ Hong Kong, Shenzhen Res Inst, Shenzhen 518057, Peoples R China
关键词
Feature extraction; Grasping; Robots; Robot sensing systems; Sensors; Detectors; Grippers; Robot sensing system; visual sensing; neural networks; vision-based perception; real-time system; grasp pose estimation;
D O I
10.1109/JSEN.2022.3163730
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Grasping is critical for intelligent robots to accomplish sophisticated tasks. Even with multimodal sensor fusion, accurately and reliably estimating grasp poses for complex-shaped objects remains a challenge. In this paper, we design a vision-based grasping platform for a more general case, that is, grasping a variety of objects by a simple parallel gripper with the grasp detection model consuming RGB sensing or depth sensing. Focusing on the grasp pose estimation part, we propose a deep grasp detector that uses a densely connected Feature Pyramid Network (FPN) feature extractor and multiple two-stage detection units to achieve dense grasp pose predictions. Specifically, for the feature extractor, the fusion of different layer feature maps can increase both the model's capacity to detect the various size grasp areas and the accuracy of the regressed grasp positions. For each of the two-stage detection unit, the first stage generates horizontal candidate grasp areas, while the second stage refines them to predict the rotated grasp poses. We train and validate our grasp pose estimation algorithm on the Cornell Grasp Dataset and the Jacquard Dataset. The model achieves the detection accuracy of 93.3% and 89.6%, respectively. We further design real-world grasp experiments to verify the effectiveness of our vision-based robotic grasping system. The real scenario trials validate that the system is capable of grasping unseen objects, in particular, achieving robust and accurate grasp pose detection and gripper opening width measurement based on depth sensing only.
引用
收藏
页码:9610 / 9620
页数:11
相关论文
共 34 条
[1]   End-to-end Trainable Deep Neural Network for Robotic Grasp Detection and Semantic Segmentation from RGB [J].
Ainetter, Stefan ;
Fraundorfer, Friedrich .
2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, :13452-13458
[2]  
[Anonymous], 2019, 33 AAAI C ART INT
[3]  
Asir U, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P4875
[4]  
Cao H.., 2021, ARXIV210110226
[5]  
Chalvatzaki G., 2020, ARXIV200605123
[6]   FINDING ANTIPODAL POINT GRASPS ON IRREGULARLY SHAPED OBJECTS [J].
CHEN, IM ;
BURDICK, JW .
IEEE TRANSACTIONS ON ROBOTICS AND AUTOMATION, 1993, 9 (04) :507-512
[7]  
Cheng H., PROC IEEERSJ INT C I, P4686
[8]  
Cheng H, 2020, IEEE INT CONF ROBOT, P3604, DOI [10.1109/ICRA40945.2020.9197333, 10.1109/icra40945.2020.9197333]
[9]  
Cheng H, 2018, 2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (ROBIO), P544, DOI 10.1109/ROBIO.2018.8665219
[10]   Real-World Multiobject, Multigrasp Detection [J].
Chu, Fu-Jen ;
Xu, Ruinian ;
Vela, Patricio A. .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04) :3355-3362