Self-Supervised Monocular Depth Estimation Using HOG Feature Prediction

被引:0
作者
He, Xin [1 ]
Zhao, Xiao [2 ]
机构
[1] Beijing Inst Technol, Beijing 100081, Peoples R China
[2] Shanghai Inst Satellite Engn, Shanghai 201109, Peoples R China
来源
PROCEEDINGS OF 2024 INTERNATIONAL CONFERENCE ON COMPUTER AND MULTIMEDIA TECHNOLOGY, ICCMT 2024 | 2024年
关键词
Image processing; Depth estimation; Self-supervised learning; HOG Prediction;
D O I
10.1145/3675249.3675316
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Accurate depth estimation from monocular images continues to pose a formidable challenge owing to boundary blurring, illumination variations, and occlusions encountered in traditional monocular depth estimation methods. In response, this paper introduces a fresh approach to self-supervised monocular depth estimation leveraging HOG (Histogram of Oriented Gradients) feature prediction. The method aims to mitigate the aforementioned challenges by preserving fine details at object boundaries and improving prediction accuracy. Central to our methodology is the introduction of a HOG feature prediction module. This module meticulously extracts HOG feature vectors from input image characteristics while emphasizing the retention of crucial boundary information. Furthermore, it strategically enhances the encoder's downward output features, thereby refining the depth estimation process. Our proposed approach has been thoroughly assessed through comprehensive experimental evaluations on the extensively used KITTI datasets, affirming its efficacy. The results underscore its superiority when compared to prevailing mainstream methodologies, particularly in accurately predicting fine-grained depth details along object edges. The proposed framework exhibits robustness against boundary blurring, illumination variations, and occlusions, offering promising advancements in monocular depth estimation techniques.
引用
收藏
页码:382 / 387
页数:6
相关论文
共 18 条
[1]  
Casser V, 2019, AAAI CONF ARTIF INTE, P8001
[2]   Digging Into Self-Supervised Monocular Depth Estimation [J].
Godard, Clement ;
Mac Aodha, Oisin ;
Firman, Michael ;
Brostow, Gabriel .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :3827-3837
[3]   Photon-Starved Scene Inference using Single Photon Cameras [J].
Goyal, Bhavya ;
Gupta, Mohit .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :2492-2501
[4]  
Kim D., 2022, arXiv
[5]   MINE: Towards Continuous Depth MPI with NeRF for Novel View Synthesis [J].
Li, Jiaxin ;
Feng, Zijian ;
She, Qi ;
Ding, Henghui ;
Wang, Changhu ;
Lee, Gim Hee .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :12558-12568
[6]   Adaptive Surface Normal Constraint for Depth Estimation [J].
Long, Xiaoxiao ;
Lin, Cheng ;
Liu, Lingjie ;
Li, Wei ;
Theobalt, Christian ;
Yang, Ruigang ;
Wang, Wenping .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :12829-12838
[7]   Trap Attention: Monocular Depth Estimation with Manual Traps [J].
Ning, Chao ;
Gan, Hongping .
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, :5033-5043
[8]   Is Pseudo-Lidar needed for Monocular 3D Object detection? [J].
Park, Dennis ;
Ambrus, Rares ;
Guizilini, Vitor ;
Li, Jie ;
Gaidon, Adrien .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :3122-3132
[9]   Vision Transformers for Dense Prediction [J].
Ranftl, Rene ;
Bochkovskiy, Alexey ;
Koltun, Vladlen .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :12159-12168
[10]   Competitive Collaboration: Joint Unsupervised Learning of Depth, Camera Motion, Optical Flow and Motion Segmentation [J].
Ranjan, Anurag ;
Jampani, Varun ;
Balles, Lukas ;
Kim, Kihwan ;
Sun, Deqing ;
Wulff, Jonas ;
Black, Michael J. .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :12232-12241