DCL-depth: monocular depth estimation network based on iam and depth consistency loss

被引:0
作者
Han C. [1 ]
Lv C. [1 ]
Kou Q. [2 ]
Jiang H. [1 ]
Cheng D. [1 ]
机构
[1] The School of Information and Control Engineering, China University of Mining and Technology, Xuzhou
[2] The School of Computer Science and Technology, China University of Mining and Technology, Xuzhou
基金
中国国家自然科学基金;
关键词
Depth consistency loss; Depth estimation; Image activity measure; Self-Supervised learning;
D O I
10.1007/s11042-024-18877-7
中图分类号
学科分类号
摘要
The self-supervised monocular depth estimation algorithm obtains excellent results in outdoor environments. However, traditional self-supervised depth estimation methods often suffer from edge blurring in complex textured regions and the loss of depth information in pixels within weakly-textured areas. To enhance the perception ability of the deep network for complex textured areas and the accuracy of depth estimation in weakly-textured regions, the following methods are proposed in this paper. First of all, the image activity measure (IAM) is used to segment the image features. Based on the multi-directional distribution of image contours, the network's perception ability has been improved, resulting in effective enhancement of depth estimation in complex regions. Furthermore, a new loss function called depth consistency loss (DCL) is proposed, which is based on recursive recurrent networks. The DCL aims to measure the similarity between the output images of the first-order network and the second-order network, and the network's constraint on weak-texture regions has been strengthened. By employing this approach, the accuracy of estimating depth information in weakly-textured regions can be enhanced. Through adequate experimentation on the public indoor datasets, the results show that our network outperforms the compared algorithms in terms of accuracy and visualization of predicted depth. © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2024.
引用
收藏
页码:4773 / 4787
页数:14
相关论文
共 46 条
  • [1] Ali U., Bayramli B., Alsarhan T., Et al., A lightweight network for monocular depth estimation with decoupled body and edge supervision[J], Image Vis Comput, 113, (2021)
  • [2] Mathew A., Mathew J., Monocular depth estimation with SPN loss[J], Image Vis Comput, 100, (2020)
  • [3] Bay H., Tuytelaars T., Gool L.V., Surf: Speeded up robust features[C, ]. In: Proceedings of the European Conference on Computer Vision, pp. 404-417, (2006)
  • [4] Yu Z., Jin L., Gao S., P<sup>2</sup> net: Patch-match and plane-regularization for unsupervised indoor depth estimation, In: European Conference on Computer Vision, pp. 206-222, (2020)
  • [5] Li B., Huang Y., Liu Z., Zou D., Yu W., StructDepth: Leveraging the structural regularities for self-supervised indoor depth estimation, In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 12663-12673, (2021)
  • [6] Patil V., Sakaridis C., Liniger A., van Gool L., P3Depth: Monocular depth estimation with a piecewise planarity prior, In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1610-1621, (2022)
  • [7] Chen L., Guo L., Cheng D., Et al., Structure-preserving and color-restoring up-sampling for single low-light image[J], IEEE Trans Circuits Syst Video Technol, 32, 4, pp. 1889-1902, (2021)
  • [8] Yang H., Lin W., Deng C., Image activity measure (IAM) for screen image segmentation[C], In: Proc IEEE Int Conf Image Process, pp. 1569-1572, (2012)
  • [9] Silberman N., Hoiem D., Kohli P., Fergus R., Indoor segmentation and support inference from rgbd images[C, Proceedings of the European Conference on Computer Vision, pp. 746-760, (2012)
  • [10] Dai A., Chang A.X., Savva M., Halber M., Funkhouser T., Niessner M., Scannet: Richly-annotated 3d reconstructions of indoor scenes, In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5828-5839, (2017)