Depth map reconstruction method based on multi-scale cross-modal feature fusion

被引:0
作者
Yang J. [1 ]
Xie T. [1 ]
Yue H. [1 ]
机构
[1] School of Electrical and Information Engineering, Tianjin University, Tianjin
来源
Huazhong Keji Daxue Xuebao (Ziran Kexue Ban)/Journal of Huazhong University of Science and Technology (Natural Science Edition) | 2023年 / 51卷 / 03期
关键词
deep learning; depth map; depth map reconstruction; feature fusion; residual learning;
D O I
10.13245/j.hust.230603
中图分类号
学科分类号
摘要
Aiming at the reconstruction problem of sparse depth maps,a deep-learning based depth map reconstruction method with multi-level cross-modal feature fusion was proposed. RGB information and depth information were encoded respectively by convolution neural network and adaptive feature fusion was performed in multiple scales. The reconstruction results were further refined by residual learning in the decoding stage. The feature upsampling was guided by the skip connection from depth encoding branch,and the initial depth map was output. At the same time,the rich semantic features extracted by the RGB encoding branch were multiplexed,and the upsampling feature maps output at multiple scales were iteratively upsampled by the pyramid upsampling block. Then,the residual with the final reconstruction result was learned,which improved the quality of reconstruction results.Experimental results on both NYU-Depth-v2 and KITTI dataset show that the proposed method has better depth map reconstruction performance compared with the existing mainstream methods,and generates sharper depth boundaries in visual comparison. © 2023 Huazhong University of Science and Technology. All rights reserved.
引用
收藏
页码:52 / 59
页数:7
相关论文
共 28 条
[1]  
FU C, MERTZ C, DOLAN J M., LIDAR and monocular camera fusion:on-road depth completion for autonomous driving[C], Proceedings of the IEEE Intelligent Transportation Systems Conference, pp. 273-278, (2019)
[2]  
YANG J, YE X, LI K, Color-guided depth recovery from RGB-D data using an adaptive autoregres-sive model[J], IEEE Transactions on Image Processing, 23, 8, pp. 3443-3458, (2014)
[3]  
HAM B, CHO M, PONCE J., Robust image filtering using joint static and dynamic guidance[C], Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4823-4831, (2015)
[4]  
YANG J, YE X, Graph based non-uniform sampling and reconstruction of depth maps[C], Proceedings of the IEEE International Conference on Image Processing, pp. 123-129, (2019)
[5]  
MA F,, KARAMAN S., Sparse-to-dense:depth prediction from sparse depth samples and a single image[C], Proceedings of the IEEE International Conference on Robotics and Automation, pp. 4796-4803, (2018)
[6]  
CHENG X, WANG P,, YANG R., Depth estimation via affinity learned with convolutional spatial propagation network[C], Proceedings of the European Conference on Computer Vision, pp. 103-119, (2018)
[7]  
QIU J, CUI Z, ZHANG Y, DeepLiDAR:deep surface normal guided depth prediction for outdoor scene from sparse LiDAR data and single color image[C], Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3308-3317, (2019)
[8]  
ELDESOKEY A, FELSBERG M, KHAN F S., Confidence propagation through CNNs for guided sparse depth regression[J], IEEE Transactions on Pattern Analysis and Machine Intelligence, 42, 10, pp. 2423-2436, (2020)
[9]  
TANG J, TIAN F P,, FENG W, Learning guided convolutional network for depth completion[J], IEEE Transactions on Image Processing, 30, 4, pp. 1116-1129, (2021)
[10]  
SENUSHKIN D,, ROMANOV M, BELIKOV I, Decoder modulation for indoor depth completion[C], Proceedings of the IEEE International Conference on Intelligent Robots and Systems, pp. 2181-2188, (2021)