Exploring Data-Efficient 3D Scene Understanding with Contrastive Scene Contexts

被引:146
作者
Hou, Ji [1 ]
Graham, Benjamin [2 ]
Niesner, Matthias [1 ]
Xie, Saining [2 ]
机构
[1] Tech Univ Munich, Munich, Germany
[2] Facebook AI Res, Menlo Pk, CA USA
来源
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021 | 2021年
关键词
D O I
10.1109/CVPR46437.2021.01533
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The rapid progress in 3D scene understanding has come with growing demand for data; however, collecting and annotating 3D scenes (e.g. point clouds) are notoriously hard. For example, the number of scenes (e.g. indoor rooms) that can be accessed and scanned might be limited; even given sufficient data, acquiring 3D labels (e.g. instance masks) requires intensive human labor. In this paper, we explore data-efficient learning for 3D point cloud. As a first step towards this direction, we propose Contrastive Scene Contexts, a 3D pre-training method that makes use of both point-level correspondences and spatial contexts in a scene. Our method achieves state-of-the-art results on a suite of benchmarks where training data or labels are scarce. Our study reveals that exhaustive labelling of 3D point clouds might be unnecessary; and remarkably, on ScanNet, even using 0.1% of point labels, we still achieve 89% (instance segmentation) and 96% (semantic segmentation) of the baseline performance that uses full annotations.
引用
收藏
页码:15582 / 15592
页数:11
相关论文
共 71 条
[1]  
[Anonymous], 2015, CVPR
[2]  
[Anonymous], 2016 IEEE C COMPUTER, DOI DOI 10.1109/CVPR.2016.609
[3]  
[Anonymous], 2020, ICML
[4]  
[Anonymous], 2019, CVPR, DOI DOI 10.1109/CVPR.2019.00319
[5]  
[Anonymous], 2020, NeurIPS
[6]  
[Anonymous], 2018, CVPR, DOI DOI 10.1109/CVPR.2018.00030
[7]  
[Anonymous], 2018, CVPR, DOI DOI 10.1109/CVPR.2018.00256
[8]  
[Anonymous], 2019, CVPR, DOI DOI 10.1109/CVPR.2019.00455
[9]  
[Anonymous], 2020, ICML
[10]  
[Anonymous], 2019, CVPR, DOI DOI 10.1109/CVPR.2019.00100