Hierarchical Features Driven Residual Learning for Depth Map Super-Resolution

被引:161
作者
Guo, Chunle [1 ]
Li, Chongyi [1 ]
Guo, Jichang [1 ]
Cong, Runmin [1 ]
Fu, Huazhu [2 ]
Han, Ping [3 ]
机构
[1] Tianjin Univ, Sch Elect & Informat Engn, Tianjin 300072, Peoples R China
[2] Incept Inst Artificial Intelligence, Abu Dhabi, U Arab Emirates
[3] Civil Aviat Univ China, Coll Elect Informat & Automat, Tianjin 300300, Peoples R China
基金
中国国家自然科学基金;
关键词
Convolutional neural network (CNN); depth map super-resolution (SR); residual learning; image reconstruction; RESOLUTION;
D O I
10.1109/TIP.2018.2887029
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Rapid development of affordable and portable consumer depth cameras facilitates the use of depth information in many computer vision tasks such as intelligent vehicles and 3D reconstruction. However, depth map captured by low-cost depth sensors (e.g., Kinect) usually suffers from low spatial resolution, which limits its potential applications. In this paper, we propose a novel deep network for depth map super-resolution (SR), called DepthSR-Net. The proposed DepthSR-Net automatically infers a high-resolution (HR) depth map from its low-resolution (LR) version by hierarchical features driven residual learning. Specifically, DepthSR-Net is built on residual U-Net deep network architecture. Given LR depth map, we first obtain the desired HR by bicubic interpolation upsampling and then construct an input pyramid to achieve multiple level receptive fields. Next, we extract hierarchical features from the input pyramid, intensity image, and encoder-decoder structure of U-Net. Finally, we learn the residual between the interpolated depth map and the corresponding HR one using the rich hierarchical features. The final HR depth map is achieved by adding the learned residual to the interpolated depth map. We conduct an ablation study to demonstrate the effectiveness of each component in the proposed network. Extensive experiments demonstrate that the proposed method outperforms the state-of-the-art methods. In addition, the potential usage of the proposed network in other low-level vision problems is discussed.
引用
收藏
页码:2545 / 2557
页数:13
相关论文
共 53 条
[1]  
[Anonymous], P EUR C COMP VIS ECC
[2]  
[Anonymous], PROC CVPR IEEE
[3]  
[Anonymous], 2018, 2018 IEEE 28 INT WOR
[4]  
[Anonymous], 2017, P IEEE C COMP VIS PA, DOI DOI 10.1109/CVPR.2017.19
[5]  
[Anonymous], 2017, IEEE T CYBERNETICS
[6]  
[Anonymous], 2017, P BMVC
[7]  
[Anonymous], 2017, P IEEE C COMP VIS PA
[8]   Co-Saliency Detection for RGBD Images Based on Multi-Constraint Feature Matching and Cross Label Propagation [J].
Cong, Runmin ;
Lei, Jianjun ;
Fu, Huazhu ;
Huang, Qingming ;
Cao, Xiaochun ;
Hou, Chunping .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (02) :568-579
[9]   Image Super-Resolution Using Deep Convolutional Networks [J].
Dong, Chao ;
Loy, Chen Change ;
He, Kaiming ;
Tang, Xiaoou .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2016, 38 (02) :295-307
[10]   Color-Guided Depth Recovery via Joint Local Structural and Nonlocal Low-Rank Regularization [J].
Dong, Weisheng ;
Shi, Guangming ;
Li, Xin ;
Peng, Kefan ;
Wu, Jinjian ;
Guo, Zhenhua .
IEEE TRANSACTIONS ON MULTIMEDIA, 2017, 19 (02) :293-301