Beat the MTurkers: Automatic Image Labeling from Weak 3D Supervision

被引:33
作者
Chen, Liang-Chieh [1 ]
Fidler, Sanja [2 ,3 ]
Yuille, Alan L. [1 ]
Urtasun, Raquel [2 ,3 ]
机构
[1] Univ Calif Los Angeles, Los Angeles, CA 90024 USA
[2] Univ Toronto, Toronto, ON M5S 1A1, Canada
[3] ITI Chicago, Chicago, IL USA
来源
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2014年
关键词
D O I
10.1109/CVPR.2014.409
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Labeling large-scale datasets with very accurate object segmentations is an elaborate task that requires a high degree of quality control and a budget of tens or hundreds of thousands of dollars. Thus, developing solutions that can automatically perform the labeling given only weak supervision is key to reduce this cost. In this paper, we show how to exploit 3D information to automatically generate very accurate object segmentations given annotated 3D bounding boxes. We formulate the problem as the one of inference in a binary Markov random field which exploits appearance models, stereo and/or noisy point clouds, a repository of 3D CAD models as well as topological constraints. We demonstrate the effectiveness of our approach in the context of autonomous driving, and show that we can segment cars with the accuracy of 86% intersection-over-union, performing as well as highly recommended MTurkers!
引用
收藏
页码:3198 / 3205
页数:8
相关论文
共 39 条