Using full-scale feature fusion for self-supervised indoor depth estimation

被引:0
|
作者
Cheng, Deqiang [1 ]
Chen, Junhui [1 ]
Lv, Chen [1 ]
Han, Chenggong [1 ]
Jiang, He [1 ]
机构
[1] China Univ Min & Technol, Sch Informat & Control Engn, Xuzhou 221116, Peoples R China
基金
中国国家自然科学基金;
关键词
Monocular depth estimation; Feature fusion; Self-supervised; Indoor scenes; ResNeSt;
D O I
10.1007/s11042-023-16581-6
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Monocular depth estimation is a crucial task in computer vision, and self-supervised algorithms are gaining popularity due to their independence from expensive ground truth supervision. However, current self-supervised algorithms may not provide accurate estimation and may suffer from distorted boundaries when applied to indoor scenes. Combining multi-scale features is an important research direction in image segmentation to achieve accurate estimation and resolve boundary distortion. However, there are few studies on indoor self-supervised algorithms in this regard. To solve this issue, we propose a novel full-scale feature information fusion approach that includes a full-scale skip-connection and a full-scale feature fusion block. This approach can aggregate the high-level and low-level information of all scale feature maps during the network's encoding and decoding process to compensate for the network's loss of cross-layer feature information. The proposed full-scale feature fusion improves accuracy and reduces the decoder parameters. To fully exploit the superiority of the full-scale feature fusion module, we replace the encoder backbone from ResNet with the more advanced ResNeSt. Combining these two methods results in a significant improvement in prediction accuracy. We have extensively evaluated our approach on the indoor benchmark datasets NYU Depth V2 and ScanNet. Our experimental results demonstrate that our method outperforms existing algorithms, particularly on NYU Depth V2, where our precision is raised to 83.8%.
引用
收藏
页码:28215 / 28233
页数:19
相关论文
共 50 条
  • [21] Self-Supervised Monocular Depth Estimation With Multiscale Perception
    Zhang, Yourun
    Gong, Maoguo
    Li, Jianzhao
    Zhang, Mingyang
    Jiang, Fenlong
    Zhao, Hongyu
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 3251 - 3266
  • [22] Self-Supervised Monocular Depth Estimation With Extensive Pretraining
    Choi, Hyukdoo
    IEEE ACCESS, 2021, 9 : 157236 - 157246
  • [23] Self-supervised monocular depth estimation for gastrointestinal endoscopy
    Liu, Yuying
    Zuo, Siyang
    COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE, 2023, 238
  • [24] Hierarchical Multi-scale Architecture Search for Self-supervised Monocular Depth Estimation
    Ren, Jian
    Xie, Jin
    Jin, Zhong
    PATTERN RECOGNITION, ACPR 2021, PT II, 2022, 13189 : 447 - 461
  • [25] Self-supervised monocular Depth estimation with multi-scale structure similarity loss
    Han, Chenggong
    Cheng, Deqiang
    Kou, Qiqi
    Wang, Xiaoyi
    Chen, Liangliang
    Zhao, Jiamin
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 82 (24) : 38035 - 38050
  • [26] Self-supervised monocular Depth estimation with multi-scale structure similarity loss
    Chenggong Han
    Deqiang Cheng
    Qiqi Kou
    Xiaoyi Wang
    Liangliang Chen
    Jiamin Zhao
    Multimedia Tools and Applications, 2023, 82 : 38035 - 38050
  • [27] GlocalFuse-Depth: Fusing transformers and CNNs for all-day self-supervised monocular depth estimation
    Zhang, Zezheng
    Chan, Ryan K. Y.
    Wong, Kenneth K. Y.
    NEUROCOMPUTING, 2024, 569
  • [28] Self-supervised monocular depth and ego-motion estimation for CT-bronchoscopy fusion
    Chang, Qi
    Higgins, William E.
    IMAGE-GUIDED PROCEDURES, ROBOTIC INTERVENTIONS, AND MODELING, MEDICAL IMAGING 2024, 2024, 12928
  • [29] Scale-Aware Visual-Inertial Depth Estimation and Odometry Using Monocular Self-Supervised Learning
    Lee, Chungkeun
    Kim, Changhyeon
    Kim, Pyojin
    Lee, Hyeonbeom
    Kim, H. Jin
    IEEE ACCESS, 2023, 11 : 24087 - 24102
  • [30] Monocular Depth Estimation via Self-Supervised Self-Distillation
    Hu, Haifeng
    Feng, Yuyang
    Li, Dapeng
    Zhang, Suofei
    Zhao, Haitao
    SENSORS, 2024, 24 (13)