Real-time 3D Reconstruction in Dynamic Scenes using Point-based Fusion

被引:187
作者
Keller, Maik [1 ]
Lefloch, Damien [2 ]
Lambers, Martin [2 ]
Izadi, Shahram [3 ]
Weyrich, Tim [4 ]
Kolb, Andreas [2 ]
机构
[1] Pmdtechnologies, Siegen, Germany
[2] Univ Siegen, D-57068 Siegen, Germany
[3] Microsoft Res, Mountain View, CA USA
[4] UCL, London WC1E 6BT, England
来源
2013 INTERNATIONAL CONFERENCE ON 3D VISION (3DV 2013) | 2013年
关键词
CAMERAS;
D O I
10.1109/3DV.2013.9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Real-time or online 3D reconstruction has wide applicability and receives further interest due to availability of consumer depth cameras. Typical approaches use a moving sensor to accumulate depth measurements into a single model which is continuously refined. Designing such systems is an intricate balance between reconstruction quality, speed, spatial scale, and scene assumptions. Existing online methods either trade scale to achieve higher quality reconstructions of small objects/scenes. Or handle larger scenes by trading real-time performance and/or quality, or by limiting the bounds of the active reconstruction. Additionally, many systems assume a static scene, and cannot robustly handle scene motion or reconstructions that evolve to reflect scene changes. We address these limitations with a new system for real-time dense reconstruction with equivalent quality to existing online methods, but with support for additional spatial scale and robustness in dynamic scenes. Our system is designed around a simple and flat point-based representation, which directly works with the input acquired from range/depth sensors, without the overhead of converting between representations. The use of points enables speed and memory efficiency, directly leveraging the standard graphics pipeline for all central operations; i.e., camera pose estimation, data association, outlier removal, fusion of depth maps into a single denoised model, and detection and update of dynamic objects. We conclude with qualitative and quantitative results that highlight robust tracking and high quality reconstructions of a diverse set of scenes at varying scales.
引用
收藏
页码:1 / 8
页数:8
相关论文
共 27 条
[1]  
[Anonymous], 1987, ACM siggraph computer graphics, DOI [10.1145/37401.37422, DOI 10.1145/37401.37422]
[2]  
[Anonymous], 2006, 2006 IEEE COMP SOC C
[3]  
[Anonymous], 2006, P 4 EUR S GEOM PROC
[4]   A METHOD FOR REGISTRATION OF 3-D SHAPES [J].
BESL, PJ ;
MCKAY, ND .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1992, 14 (02) :239-256
[5]   OBJECT MODELING BY REGISTRATION OF MULTIPLE RANGE IMAGES [J].
CHEN, Y ;
MEDIONI, G .
IMAGE AND VISION COMPUTING, 1992, 10 (03) :145-155
[6]  
Curless B., 1996, Computer Graphics Proceedings. SIGGRAPH '96, P303, DOI 10.1145/237170.237269
[7]  
Gallup D, 2010, LECT NOTES COMPUT SC, V6376, P1
[8]   RGB-D mapping: Using Kinect-style depth cameras for dense 3D modeling of indoor environments [J].
Henry, Peter ;
Krainin, Michael ;
Herbst, Evan ;
Ren, Xiaofeng ;
Fox, Dieter .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2012, 31 (05) :647-663
[9]  
HOPPE H, 1992, COMP GRAPH, V26, P71, DOI 10.1145/142920.134011
[10]  
Izadi S., 2011, P 24 ANN ACM S US IN, P559