Learning Affordance Space in Physical World for Vision-based Robotic Object Manipulation

被引:0
作者
Wu, Huadong [1 ]
Zhang, Zhanpeng [2 ]
Cheng, Hui [1 ]
Yang, Kai [1 ]
Liu, Jiaming [2 ]
Guo, Ziying [1 ]
机构
[1] Sun Yat Sen Univ, Guangzhou, Peoples R China
[2] SenseTime Grp Ltd, Hong Kong, Peoples R China
来源
2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA) | 2020年
关键词
D O I
10.1109/icra40945.2020.9196783
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
What is a proper representation for objects in manipulation? What would human try to perceive when manipulating a new object in a new environment? In fact, instead of focusing on the texture and illumination, human can infer the "affordance" [36] of the objects from vision. Here "affordance" describes the object's intrinsic property that affords a particular type of manipulation. In this work, we investigate whether such affordance can be learned by a deep neural network. In particular, we propose an Affordance Space Perception Network (ASPN) that takes an image as input and outputs an affordance map. Different from existing works that infer the pixel-wise probability affordance map in image space, our affordance is defined in the real world space, thus eliminates the need of hand-eye calibration. In addition, we extend the representation ability of affordance by defining it in a 3D affordance space and propose a novel training strategy to improve the performance. Trained purely with simulation data, ASPN can achieve significant performance in the real world. It is a task-agnostic framework and can handle different objects, scenes and viewpoints. Extensive real-world experiments demonstrate the accuracy and robustness of our approach. We achieve the success rates of 94.2% for singular-object pushing and 92.4% for multiple-object pushing. We also achieve the success rates of 97.2% for singular-object grasping and 95.4% for multiple-object grasping, which outperform current state-of-the-art methods.
引用
收藏
页码:4652 / 4658
页数:7
相关论文
共 38 条
[1]  
[Anonymous], 2011, Learning grasp affordance densities, DOI DOI 10.2478/S13230-011-0012-X
[2]  
[Anonymous], 1993, A Mathematical Introduction to Robotic Manipulation
[3]  
Bousmalis K, 2016, ADV NEUR IN, V29
[4]  
Bousmalis K, 2018, IEEE INT CONF ROBOT, P4243
[6]  
Cai JH, 2019, IEEE INT CONF ROBOT, P4960, DOI [10.1109/ICRA.2019.8793912, 10.1109/icra.2019.8793912]
[7]  
Dehban A, 2017, 2017 IEEE-RAS 17TH INTERNATIONAL CONFERENCE ON HUMANOID ROBOTICS (HUMANOIDS), P476, DOI 10.1109/HUMANOIDS.2017.8246915
[8]  
Detry R, 2010, STUD COMPUT INTELL, V264, P451
[9]  
Fang K, 2018, IEEE INT CONF ROBOT, P3516, DOI 10.1109/ICRA.2018.8461041
[10]  
Gibson J. J., 2014, ECOLOGICAL APPROACH