Estimating Depth from Single Image Using Unsupervised Convolutional Network

被引:0
作者
Sun Y. [1 ]
Shi J. [1 ]
Sun Z. [2 ]
机构
[1] School of Computer Science and Engineering, Jiangsu University of Science and Technology, Zhenjiang
[2] State Key Laboratory for Novel Software Technology, Nanjing University, Nanjing
来源
Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao/Journal of Computer-Aided Design and Computer Graphics | 2020年 / 32卷 / 04期
关键词
Convolution neural network; Depth estimation;
D O I
10.3724/SP.J.1089.2020.17825
中图分类号
学科分类号
摘要
To improve the accuracy of monocular depth estimation by deep learning, this paper proposes a method of unsupervised convolutional neural network for depth estimation from one single image. By introducing residual structure, dense connection structure and short-cut connection in the encode-decode network structure, the single image depth estimation convolutional neural network is improved and the learning efficiency and performance of the network are improved, and the convergence speed of the network is accelerated. Secondly, combined with the loss metrics such as gray similarity, disparity smoothing and left and right disparities matching, a more efficient loss function is designed, which effectively reduces the influence of image illumination factors, suppresses the discontinuity of image depth and ensures the consistency of left and right disparities. Thereby, the robustness of the depth estimation is improved. Finally, this method realizes the end-to-end depth estimation from single image, where stereo image sequences are used as the training data and the depth information in the single image scene can be estimated without the target depth supervision information. Experiments and comparisons on the KITTI and Cityscapes datasets with TensorFlow framework prove the effectiveness of the proposed method with higher accuracy and faster convergence. © 2020, Beijing China Science Journal Publishing Co. Ltd. All right reserved.
引用
收藏
页码:643 / 651
页数:8
相关论文
共 40 条
[1]  
Saxena A, Schulte J, Ng A Y., Depth estimation using monocular and stereo cues, Proceedings of the 20th International Joint Conference on Artificial Intelligence, pp. 2197-2203, (2007)
[2]  
Saxena A, Chung S H, Ng A Y., Learning depth from single monocular images, Proceedings of the 18th International Conference on Neural Information Processing Systems, pp. 1161-1168, (2005)
[3]  
Zhou T H, Brown M, Snavely N, Et al., Unsupervised learning of depth and ego-motion from video, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1851-1858, (2017)
[4]  
Xie J Y, Girshick R, Farhadi A., Deep3D: fully automatic 2D-to-3D video conversion with deep convolutional neural networks, Proceedings of European Conference on Computer Vision, pp. 842-857, (2016)
[5]  
Stoyanov D, Scarzanella M V, Pratt P, Et al., Real-time stereo reconstruction in robotically assisted minimally invasive surgery, Proceedings of the 13th International Conference on Medical Image Computing and Computer-Assisted Intervention, Part I, pp. 275-282, (2010)
[6]  
Eigen D, Puhrsch C, Fergus R., Depth map prediction from a single image using a multi-scale deep network, Proceedings of the 27th International Conference on Neural Information Processing Systems, pp. 2366-2374, (2014)
[7]  
Chakrabarti A, Shao J, Shakhnarovich G., Depth from a single image by harmonizing overcomplete local network predictions, Proceedings of in Neural Information Processing Systems, pp. 2658-2666, (2016)
[8]  
Eigen D, Fergus R., Predicting depth, surface normals and semantic labels with a common multi-scale convolutional architecture, Proceedings of the IEEE International Conference on Computer Vision, pp. 2650-2658, (2015)
[9]  
Wang P, Shen X H, Lin Z, Et al., Towards unified depth and semantic prediction from a single image, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2800-2809, (2015)
[10]  
Jafari O H, Groth O, Kirillov A, Et al., Analyzing modular CNN architectures for joint depth prediction and semantic segmentation, Proceedings of the IEEE International Conference on Robotics and Automation, pp. 4620-4627, (2017)