Using full-scale feature fusion for self-supervised indoor depth estimation

被引:0
|
作者
Cheng, Deqiang [1 ]
Chen, Junhui [1 ]
Lv, Chen [1 ]
Han, Chenggong [1 ]
Jiang, He [1 ]
机构
[1] China Univ Min & Technol, Sch Informat & Control Engn, Xuzhou 221116, Peoples R China
基金
中国国家自然科学基金;
关键词
Monocular depth estimation; Feature fusion; Self-supervised; Indoor scenes; ResNeSt;
D O I
10.1007/s11042-023-16581-6
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Monocular depth estimation is a crucial task in computer vision, and self-supervised algorithms are gaining popularity due to their independence from expensive ground truth supervision. However, current self-supervised algorithms may not provide accurate estimation and may suffer from distorted boundaries when applied to indoor scenes. Combining multi-scale features is an important research direction in image segmentation to achieve accurate estimation and resolve boundary distortion. However, there are few studies on indoor self-supervised algorithms in this regard. To solve this issue, we propose a novel full-scale feature information fusion approach that includes a full-scale skip-connection and a full-scale feature fusion block. This approach can aggregate the high-level and low-level information of all scale feature maps during the network's encoding and decoding process to compensate for the network's loss of cross-layer feature information. The proposed full-scale feature fusion improves accuracy and reduces the decoder parameters. To fully exploit the superiority of the full-scale feature fusion module, we replace the encoder backbone from ResNet with the more advanced ResNeSt. Combining these two methods results in a significant improvement in prediction accuracy. We have extensively evaluated our approach on the indoor benchmark datasets NYU Depth V2 and ScanNet. Our experimental results demonstrate that our method outperforms existing algorithms, particularly on NYU Depth V2, where our precision is raised to 83.8%.
引用
收藏
页码:28215 / 28233
页数:19
相关论文
共 50 条
  • [1] Using full-scale feature fusion for self-supervised indoor depth estimation
    Deqiang Cheng
    Junhui Chen
    Chen Lv
    Chenggong Han
    He Jiang
    Multimedia Tools and Applications, 2024, 83 : 28215 - 28233
  • [2] Self-Supervised Monocular Depth Estimation Based on Full Scale Feature Fusion
    Wang C.
    Chen Y.
    Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao/Journal of Computer-Aided Design and Computer Graphics, 2023, 35 (05): : 667 - 675
  • [3] Indoor self-supervised monocular depth estimation based on level feature fusion
    Cheng D.
    Zhang H.
    Kou Q.
    Lü C.
    Qian J.
    Guangxue Jingmi Gongcheng/Optics and Precision Engineering, 2023, 31 (20): : 2993 - 3009
  • [4] Multilevel feature fusion and edge optimization network for self-supervised monocular depth estimation
    Liu, Guohua
    Niu, Shuqing
    JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (03)
  • [5] Multi-Frame Self-Supervised Depth Estimation with Multi-Scale Feature Fusion in Dynamic Scenes
    Zhong, Jiquan
    Huang, Xiaolin
    Yu, Xiao
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 2553 - 2563
  • [6] Depth Estimation Using a Self-Supervised Network Based on Cross-Layer Feature Fusion and the Quadtree Constraint
    Tian, Fangzheng
    Gao, Yongbin
    Fang, Zhijun
    Fang, Yuming
    Gu, Jia
    Fujita, Hamido
    Hwang, Jenq-Neng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (04) : 1751 - 1766
  • [7] Self-distilled Feature Aggregation for Self-supervised Monocular Depth Estimation
    Zhou, Zhengming
    Dong, Qiulei
    COMPUTER VISION - ECCV 2022, PT I, 2022, 13661 : 709 - 726
  • [8] TFDEPTH: SELF-SUPERVISED MONOCULARDEPTH ESTIMATION WITH MULITI-SCALE SELECTIVE TRANSFORMER FEATURE FUSION
    Hu, Hongli
    Miao, Jun
    Zhu, Guanghu
    Yan, Je
    Chu, Jun
    IMAGE ANALYSIS & STEREOLOGY, 2024, 43 (02): : 139 - 149
  • [9] Underwater self-supervised depth estimation
    Yang, Xuewen
    Zhang, Xing
    Wang, Nan
    Xin, Guoling
    Hu, Wenjie
    NEUROCOMPUTING, 2022, 514 : 362 - 373
  • [10] F2Depth: Self-supervised indoor monocular depth estimation via optical flow consistency and feature map synthesis
    Guo, Xiaotong
    Zhao, Huijie
    Shao, Shuwei
    Li, Xudong
    Zhang, Baochang
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 133