Affordance Detection of Tool Parts from Geometric Features

被引:0
作者
Myers, Austin [1 ]
Teo, Ching L. [1 ]
Fermiiller, Cornelia [1 ]
Aloimonos, Yiannis [1 ]
机构
[1] Univ Maryland, Dept Comp Sci, College Pk, MD 20742 USA
来源
2015 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA) | 2015年
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
As robots begin to collaborate with humans in everyday workspaces, they will need to understand the functions of tools and their parts. To cut an apple or hammer a nail, robots need to not just know the tool's name, but they must localize the tool's parts and identify their functions. Intuitively, the geometry of a part is closely related to its possible functions, or its affordances. Therefore, we propose two approaches for learning affordances from local shape and geometry primitives: 1) superpixel based hierarchical matching pursuit S-HMP); and 2) structured random forests (SRF). Moreover, since a part can be used in many ways, we introduce a large RGB-Depth dataset where tool parts are labeled with multiple affordances and their relative rankings. With ranked affordances, we evaluate the proposed methods on 3 cluttered scenes and over 105 kitchen, workshop and garden tools, using ranked correlation and a weighted F-measure score [26]. Experimental results over sequences containing clutter, occlusions, and viewpoint changes show that the approaches return precise predictions that could be used by a robot. S-HMP achieves high accuracy but at a significant computational cost, while SRF provides slightly less accurate predictions but in real-time. Finally, we validate the effectiveness of our approaches on the Cornell Grasping Dataset [25] for detecting graspable regions, and achieve state-of-the-art performance.
引用
收藏
页码:1374 / 1381
页数:8
相关论文
共 36 条
[31]  
Silberman Pushmeet Kohli Nathan, 2012, ECCV, DOI [DOI 10.1007/978-3-642-33715-4_54, 10.1007/978-3-642-33715-4_54]
[33]  
Stark M, 2008, LECT NOTES COMPUT SC, V5008, P435
[34]  
Sun YY, 2013, IEEE INT CONF ROBOT, P2096, DOI 10.1109/ICRA.2013.6630858
[35]  
Tin Kam Ho, 1995, Proceedings of the Third International Conference on Document Analysis and Recognition, P278, DOI 10.1109/ICDAR.1995.598994
[36]  
Yu XD, 2010, LECT NOTES COMPUT SC, V6315, P127