A Large-Scale Hierarchical Multi-View RGB-D Object Dataset

被引:0
作者
Lai, Kevin [1 ]
Bo, Liefeng [1 ]
Ren, Xiaofeng [2 ]
Fox, Dieter [1 ]
机构
[1] Univ Washington, Dept Comp Sci & Engn, Seattle, WA 98195 USA
[2] Intel Labs Seattle, Seattle, WA 98105 USA
来源
2011 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA) | 2011年
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Over the last decade, the availability of public image repositories and recognition benchmarks has enabled rapid progress in visual object category and instance detection. Today we are witnessing the birth of a new generation of sensing technologies capable of providing high quality synchronized videos of both color and depth, the RGB-D (Kinect-style) camera. With its advanced sensing capabilities and the potential for mass adoption, this technology represents an opportunity to dramatically increase robotic object recognition, manipulation, navigation, and interaction capabilities. In this paper, we introduce a large-scale, hierarchical multi-view object dataset collected using an RGB-D camera. The dataset contains 300 objects organized into 51 categories and has been made publicly available to the research community so as to enable rapid progress based on this promising technology. This paper describes the dataset collection procedure and introduces techniques for RGB-D based object recognition and detection, demonstrating that combining color and depth information substantially improves quality of results.
引用
收藏
页码:1817 / 1824
页数:8
相关论文
共 22 条
[1]  
[Anonymous], EUR WORKSH ADV VID B
[2]  
[Anonymous], 2003, The robotics data set repository (radish)
[3]  
Bo L., 2009, Advances in neural information processing systems, P135
[4]  
Bouguet JY, Camera calibration toolbox for matlab
[5]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[6]   OBJECT MODELING BY REGISTRATION OF MULTIPLE RANGE IMAGES [J].
CHEN, Y ;
MEDIONI, G .
IMAGE AND VISION COMPUTING, 1992, 10 (03) :145-155
[7]  
Dalal N., 2005, IEEE C COMP VIS PATT
[8]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[9]  
Fan RE, 2008, J MACH LEARN RES, V9, P1871
[10]  
Felzenszwalb P, 2008, PROC CVPR IEEE, P1984