SISG-Net: Simultaneous instance segmentation and grasp detection for robot grasp in clutter

被引:9
作者
Yan, Yunhui [1 ]
Tong, Ling
Song, Kechen [1 ]
Tian, Hongkun
Man, Yi
Yang, Wenkang
机构
[1] Northeastern Univ, Sch Mech Engn & Automat, Shenyang, Liaoning, Peoples R China
基金
中国国家自然科学基金;
关键词
Robot vision; Grasp detection; Instance segmentation; Artificial intelligence;
D O I
10.1016/j.aei.2023.102189
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Robots have always found it challenging to grasp in cluttered scenes because of the complex background in-formation and changing operating environment. Therefore, in order to enable robots to perform multi-object grasping tasks in a wider range of application scenarios, such as object sorting on industrial production lines and object manipulation by home service robots, we innovatively integrated segmentation and grasp detection into the same framework, and designed a simultaneous instance segmentation and grasp detection network (SISG-Net). By using the network, robots can better interact with complex environments and perform grasp tasks. In order to solve the problem of insufficient fusion of the existing RGBD fusion strategy in the robot field, we propose a lightweight RGB-D fusion module called SMCF to make modal fusion more efficient. In order to solve the problem of inaccurate perception of small objects in different scenes, we propose the FFASP module. Finally, we use the AFF module to adaptively fuse multi-scale features. Segmentation can remove noise information from the background, enabling the robot to grasp in different backgrounds with robustness. Using the segmentation result, we refine grasp detection and find the best grasp pose for robot grasping in complex scenes. Our grasp detection model performs similarly to state-of-the-art grasp detection algorithms on the Cornell Dataset. Our model achieves state-of-the-art performance on the OCID Dataset. We show that the method is stable and robust in real-world grasping experiments. The code and video of our experiment used in this paper can be found at: https://github.com/meiguiz/SISG-Net.
引用
收藏
页数:17
相关论文
共 39 条
[1]  
Ainetter S., 2021, arXiv
[2]   End-to-end Trainable Deep Neural Network for Robotic Grasp Detection and Semantic Segmentation from RGB [J].
Ainetter, Stefan ;
Fraundorfer, Friedrich .
2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, :13452-13458
[3]   Charuco Board-Based Omnidirectional Camera Calibration Method [J].
An, Gwon Hwan ;
Lee, Siyeong ;
Seo, Min-Woo ;
Yun, Kugjin ;
Cheong, Won-Sik ;
Kang, Suk-Ju .
ELECTRONICS, 2018, 7 (12)
[4]  
Araki Ryosuke, 2020, 2020 IEEE International Conference on Robotics and Automation (ICRA), P10487, DOI 10.1109/ICRA40945.2020.9197251
[5]  
Back S, 2020, IEEE IMAGE PROC, P828, DOI [10.1109/icip40778.2020.9190804, 10.1109/ICIP40778.2020.9190804]
[6]   Yale-CMU-Berkeley dataset for robotic manipulation research [J].
Calli, Berk ;
Singh, Arjun ;
Bruce, James ;
Walsman, Aaron ;
Konolige, Kurt ;
Srinivasa, Siddhartha ;
Abbeel, Pieter ;
Dollar, Aaron M. .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2017, 36 (03) :261-268
[7]  
Chalvatzaki G, 2021, Arxiv, DOI arXiv:2006.05123
[8]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[9]   Real-World Multiobject, Multigrasp Detection [J].
Chu, Fu-Jen ;
Xu, Ruinian ;
Vela, Patricio A. .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04) :3355-3362
[10]  
Couprie C, 2013, Arxiv, DOI arXiv:1301.3572