Visual Manipulation Relationship Detection based on Gated Graph Neural Network for Robotic Grasping

被引:9
作者
Ding, Mengyuan [1 ]
Liu, Yaxin [1 ]
Yang, Chenjie [1 ]
Lan, Xuguang [1 ]
机构
[1] Xi An Jiao Tong Univ, Inst Artificial Intelligence & Robot, Coll Artificial Intelligence, Natl Engn Lab Visual Informat Applicat, Xian 710049, Peoples R China
来源
2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS) | 2022年
关键词
D O I
10.1109/IROS47612.2022.9981077
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Exploring the relationship among objects and giving the correct operation sequence is vital for robotic manipulation. However, most previous algorithms only model the relationship between pairs of objects independently, ignoring the interaction effect between them, which may generate redundant or missing relations in complex scenes, such as multi-object stacking and partial occlusion. To solve this problem, a Gated Graph Neural Network (GGNN) is designed for visual manipulation relationship detection, which can help robots detect targets in complex scenes and obtain the appropriate grasping order. Firstly, the robot extracts feature from the input image and estimate object categories. Then GGNN is used to effectively capture the dependencies between objects in the whole scene, update the relevant features, and output the grasping sequence. In addition, by embedding positional encoding into pair object features, accurate context information is obtained to reduce the adverse effects of complex scenes. Finally, the constructed algorithm is applied to the physical robot for grasping. Experiment results on the Visual Manipulation Relationship Dataset (VMRD) and the large-scale relational grasp dataset named REGRAD show that our method significantly improves the accuracy of relationship detection in complex scenes, and can be well generalized in the real world.
引用
收藏
页码:1404 / 1410
页数:7
相关论文
共 20 条
  • [1] Graph-Based Global Reasoning Networks
    Chen, Yunpeng
    Rohrbach, Marcus
    Yan, Zhicheng
    Yan, Shuicheng
    Feng, Jiashi
    Kalantidis, Yannis
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 433 - 442
  • [2] Gori M., 2005, P 2005 IEEE INT JOIN, DOI DOI 10.1109/IJCNN.2005.1555942
  • [3] Gualtieri M, 2016, 2016 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2016), P598, DOI 10.1109/IROS.2016.7759114
  • [4] Extraction of Physically Plausible Support Relations to Predict and Validate Manipulation Action Effects
    Kartmann, Rainer
    Paus, Fabian
    Grotz, Markus
    Asfour, Tamim
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04): : 3991 - 3998
  • [5] Kipf T. N., 2016, IEEE CVF C COMP VIS
  • [6] Learning task-oriented grasping for tool manipulation from simulated self-supervision
    Kuan Fang
    Zhu, Yuke
    Garg, Animesh
    Kurenkov, Andrey
    Mehta, Viraj
    Li Fei-Fei
    Savarese, Silvio
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2020, 39 (2-3) : 202 - 216
  • [7] Li Y., 2016, INT C LEARNING REPRE, P1
  • [8] Visual Relationship Detection with Language Priors
    Lu, Cewu
    Krishna, Ranjay
    Bernstein, Michael
    Li Fei-Fei
    [J]. COMPUTER VISION - ECCV 2016, PT I, 2016, 9905 : 852 - 869
  • [9] Mahler J, 2017, ROBOTICS: SCIENCE AND SYSTEMS XIII
  • [10] Mi L, 2020, PROC CVPR IEEE, P13883, DOI 10.1109/CVPR42600.2020.01390