Attention-based efficient robot grasp detection network

被引:3
作者
Qin, Xiaofei [1 ]
Hu, Wenkai [1 ]
Xiao, Chen [2 ]
He, Changxiang [2 ]
Pei, Songwen [1 ,3 ,4 ]
Zhang, Xuedian [1 ,3 ,4 ,5 ]
机构
[1] Univ Shanghai Sci & Technol, Sch Opt Elect & Comp Engn, Shanghai 200093, Peoples R China
[2] Univ Shanghai Sci & Technol, Coll Sci, Shanghai 200093, Peoples R China
[3] Shanghai Key Lab Modern Opt Syst, Shanghai 200093, Peoples R China
[4] Minist Educ, Key Lab Biomed Opt Technol & Devices, Shanghai 200093, Peoples R China
[5] Tongji Univ, Shanghai Inst Intelligent Sci & Technol, Shanghai 201210, Peoples R China
关键词
Robot grasp detection; Attention mechanism; Encoder-decoder; Neural network;
D O I
10.1631/FITEE.2200502
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To balance the inference speed and detection accuracy of a grasp detection algorithm, which are both important for robot grasping tasks, we propose an encoder-decoder structured pixel-level grasp detection neural network named the attention-based efficient robot grasp detection network (AE-GDN). Three spatial attention modules are introduced in the encoder stages to enhance the detailed information, and three channel attention modules are introduced in the decoder stages to extract more semantic information. Several lightweight and efficient DenseBlocks are used to connect the encoder and decoder paths to improve the feature modeling capability of AE-GDN. A high intersection over union (IoU) value between the predicted grasp rectangle and the ground truth does not necessarily mean a high-quality grasp configuration, but might cause a collision. This is because traditional IoU loss calculation methods treat the center part of the predicted rectangle as having the same importance as the area around the grippers. We design a new IoU loss calculation method based on an hourglass box matching mechanism, which will create good correspondence between high IoUs and high-quality grasp configurations. AEGDN achieves the accuracy of 98.9% and 96.6% on the Cornell and Jacquard datasets, respectively. The inference speed reaches 43.5 frames per second with only about 1.2 x 10(6) parameters. The proposed AE-GDN has also been deployed on a practical robotic arm grasping system and performs grasping well.
引用
收藏
页码:1430 / 1444
页数:15
相关论文
共 40 条
  • [1] End-to-end Trainable Deep Neural Network for Robotic Grasp Detection and Semantic Segmentation from RGB
    Ainetter, Stefan
    Fraundorfer, Friedrich
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 13452 - 13458
  • [2] [Anonymous], P IEEE RSJ INT C INT, P9626, DOI [10.1109/IROS45743.2020.9340777, DOI 10.1109/IROS45743.2020.9340777]
  • [3] Asif U, 2019, AAAI CONF ARTIF INTE, P8085
  • [4] RGB-D Object Recognition and Grasp Detection Using Hierarchical Cascaded Forests
    Asif, Umar
    Bennamoun, Mohammed
    Sohel, Ferdous A.
    [J]. IEEE TRANSACTIONS ON ROBOTICS, 2017, 33 (03) : 547 - 564
  • [5] Asir U, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P4875
  • [6] Chellappa R, 2017, Designing deep convolutional neural networks for continuous object orientation estimation
  • [7] Detecting Graspable Rectangles of Objects in Robotic Grasping
    Chen, Lu
    Huang, Panfeng
    Li, Yuanhao
    Meng, Zhongjie
    [J]. INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2020, 18 (05) : 1343 - 1352
  • [8] Convolutional multi-grasp detection using grasp path for RGBD images
    Chen, Lu
    Huang, Panfeng
    Meng, Zhongjie
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2019, 113 : 94 - 103
  • [9] Chu F.J., 2018, arXiv
  • [10] Real-World Multiobject, Multigrasp Detection
    Chu, Fu-Jen
    Xu, Ruinian
    Vela, Patricio A.
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04): : 3355 - 3362