CoMoDA: Continuous Monocular Depth Adaptation Using Past Experiences

被引:37
作者
Kuznietsov, Yevhen [1 ]
Proesmans, Marc [1 ]
Van Gool, Luc [1 ,2 ]
机构
[1] Katholieke Univ Leuven, ESAT PSI, Leuven, Belgium
[2] Swiss Fed Inst Technol, CVL, Zurich, Switzerland
来源
2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WACV 2021 | 2021年
关键词
D O I
10.1109/WACV48630.2021.00295
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While ground truth depth data remains hard to obtain, self-supervised monocular depth estimation methods enjoy growing attention. Much research in this area aims at improving loss functions or network architectures. Most works, however, do not leverage self-supervision to its full potential. They stick to the standard closed world train-test pipeline, assuming the network parameters to be fixed after the training is finished. Such an assumption does not allow to adapt to new scenes, whereas with self-supervision this becomes possible without extra annotations. In this paper, we propose a novel self-supervised Continuous Monocular Depth Adaptation method (CoMoDA), which adapts the pretrained model on a test video on the fly. As opposed to existing test-time refinement methods that use isolated frame triplets, we opt for continuous adaptation, making use of the previous experience from the same scene. We additionally augment the proposed procedure with the experience from the distant past, preventing the model from overfitting and thus forgetting already learnt information. We demonstrate that our method can be used for both intra- and cross-dataset adaptation. By adapting the model from train to test set of the Eigen split of KITTI, we achieve state-of-the-art depth estimation performance and surpass all existing methods using standard architectures. We also show that our method runs 15 times faster than existing test-time refinement methods. The code is available at https://github.com/Yevkuzn/CoMoDA.
引用
收藏
页码:2906 / 2916
页数:11
相关论文
共 54 条
[1]   Generative Adversarial Networks for Unsupervised Monocular Depth Prediction [J].
Aleotti, Filippo ;
Tosi, Fabio ;
Poggi, Matteo ;
Mattoccia, Stefano .
COMPUTER VISION - ECCV 2018 WORKSHOPS, PT I, 2019, 11129 :337-354
[2]   Real-Time Monocular Depth Estimation using Synthetic Data with Domain Adaptation via Image Style Transfer [J].
Atapour-Abarghouei, Amir ;
Breckon, Toby P. .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :2800-2810
[3]  
Caesar H., P IEEE CVF C COMP VI
[4]  
Casser V., Depth prediction without the sensors: Leveraging structure for unsupervised learning from monocular videos
[5]  
Casser V, 2019, AAAI CONF ARTIF INTE, P8001
[6]  
Chen W, 2016, ADV NEUR IN, V29
[7]   Self-supervised Learning with Geometric Constraints in Monocular Video Connecting Flow, Depth, and Camera [J].
Chen, Yuhua ;
Schmid, Cordelia ;
Sminchisescu, Cristian .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :7062-7071
[8]  
Eigen D, 2014, ADV NEUR IN, V27
[9]   Deep Ordinal Regression Network for Monocular Depth Estimation [J].
Fu, Huan ;
Gong, Mingming ;
Wang, Chaohui ;
Batmanghelich, Kayhan ;
Tao, Dacheng .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :2002-2011
[10]   Unsupervised CNN for Single View Depth Estimation: Geometry to the Rescue [J].
Garg, Ravi ;
VijayKumar, B. G. ;
Carneiro, Gustavo ;
Reid, Ian .
COMPUTER VISION - ECCV 2016, PT VIII, 2016, 9912 :740-756