Dense Hybrid Recurrent Multi-view Stereo Net with Dynamic Consistency Checking

被引:128
作者
Yan, Jianfeng [1 ]
Wei, Zizhuang [1 ]
Yi, Hongwei [1 ]
Ding, Mingyu [2 ]
Zhang, Runze [3 ]
Chen, Yisong [1 ]
Wang, Guoping [1 ]
Tai, Yu-Wing [4 ]
机构
[1] Peking Univ, Beijing, Peoples R China
[2] HKU, Pokfulam, Peoples R China
[3] Tencent, Shenzhen, Peoples R China
[4] Kwai Inc, Beijing, Peoples R China
来源
COMPUTER VISION - ECCV 2020, PT IV | 2020年 / 12349卷
基金
国家重点研发计划;
关键词
Multi-view stereo; Deep learning; Dense hybrid recurrent-MVSNet; Dynamic consistency checking;
D O I
10.1007/978-3-030-58548-8_39
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose an efficient and effective dense hybrid recurrent multi-view stereo net with dynamic consistency checking, namely (DHC)-H-2-RMVSNet, for accurate dense point cloud reconstruction. Our novel hybrid recurrent multi-view stereo net consists of two core modules: 1) a light DRENet (Dense Reception Expanded) module to extract dense feature maps of original size with multi-scale context information, 2) a HU-LSTM (Hybrid U-LSTM) to regularize 3D matching volume into predicted depth map, which efficiently aggregates different scale information by coupling LSTM and U-Net architecture. To further improve the accuracy and completeness of reconstructed point clouds, we leverage a dynamic consistency checking strategy instead of prefixed parameters and strategies widely adopted in existing methods for dense point cloud reconstruction. In doing so, we dynamically aggregate geometric consistency matching error among all the views. Our method ranks 1st on the complex outdoor Tanks and Temples benchmark over all the methods. Extensive experiments on the in-door DTU dataset show our method exhibits competitive performance to the state-of-the-art method while dramatically reduces memory consumption, which costs only 19.4% of R-MVSNet memory consumption. The codebase is available at https://github.com/yhw-yhw/D2HC-RMVSNet.
引用
收藏
页码:674 / 689
页数:16
相关论文
共 35 条
[1]   Large-Scale Data for Multiple-View Stereopsis [J].
Aanaes, Henrik ;
Jensen, Rasmus Ramsbol ;
Vogiatzis, George ;
Tola, Engin ;
Dahl, Anders Bjorholm .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2016, 120 (02) :153-168
[2]  
Alteryx, About us
[3]  
[Anonymous], 2014, OpenMVG. an open multiple view geometry library
[4]  
[Anonymous], about us
[5]  
Chen R, 2019, Arxiv, DOI arXiv:1908.04422
[6]   Point-Based Multi-View Stereo Network [J].
Chen, Rui ;
Han, Songfang ;
Xu, Jing ;
Su, Hao .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :1538-1547
[7]   FlowNet: Learning Optical Flow with Convolutional Networks [J].
Dosovitskiy, Alexey ;
Fischer, Philipp ;
Ilg, Eddy ;
Haeusser, Philip ;
Hazirbas, Caner ;
Golkov, Vladimir ;
van der Smagt, Patrick ;
Cremers, Daniel ;
Brox, Thomas .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :2758-2766
[8]   DeepStereo: Learning to Predict New Views from the World's Imagery [J].
Flynn, John ;
Neulander, Ivan ;
Philbin, James ;
Snavely, Noah .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :5515-5524
[9]   Massively Parallel Multiview Stereopsis by Surface Normal Diffusion [J].
Galliani, Silvano ;
Lasinger, Katrin ;
Schindler, Konrad .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :873-881
[10]  
Gu XD, 2020, Arxiv, DOI [arXiv:1912.06378, DOI 10.48550/ARXIV.1912.06378]