FF-GAN: Feature Fusion GAN for Monocular Depth Estimation

被引:1
作者
Jia, Ruiming [1 ]
Li, Tong [1 ]
Yuan, Fei [2 ]
机构
[1] North China Univ Technol, Sch Informat Sci & Technol, Beijing, Peoples R China
[2] Chinese Acad Sci, Inst Automat, Digital Content Technol & Media Serv Res Ctr, Beijing, Peoples R China
来源
PATTERN RECOGNITION AND COMPUTER VISION, PT I, PRCV 2020 | 2020年 / 12305卷
关键词
Conditional Generative Adversarial Network; Encoder-decoder; Monocular depth estimation; Receptive field;
D O I
10.1007/978-3-030-60633-6_14
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Since the results of CNN methods for monocular depth estimation generally suffer the problem of visual dissatisfaction, we propose Feature Fusion GAN (FF-GAN) to address this issue. First, an end-to-end network based on encoder-decoder structure is proposed as the generator of FF-GAN, which can exploit the information of different scales. The encoder of our generator fuse features in different levels with a feature fusion module. The component which can obtain the information of multi-scale receptive field is the main part of the decoder of our generator. Second, in order to match the generator, the discriminator of FF-GANis designed to efficiently learn the information of different scales by applying pyramid structure. Experiments on public datasets demonstrate the effectiveness of our generator and discriminator. Compared with the CNN methods, the results predicted by FF-GAN are significantly improved in terms of texture loss and edge blur while ensuring accuracy, and the visual effect is better.
引用
收藏
页码:167 / 179
页数:13
相关论文
共 24 条
  • [1] The 2018 PIRM Challenge on Perceptual Image Super-Resolution
    Blau, Yochai
    Mechrez, Roey
    Timofte, Radu
    Michaeli, Tomer
    Zelnik-Manor, Lihi
    [J]. COMPUTER VISION - ECCV 2018 WORKSHOPS, PT V, 2019, 11133 : 334 - 355
  • [2] Chen LC, 2017, Arxiv, DOI arXiv:1706.05587
  • [3] Eigen D, 2014, ADV NEUR IN, V27
  • [4] Geiger A, 2012, PROC CVPR IEEE, P3354, DOI 10.1109/CVPR.2012.6248074
  • [5] Digging Into Self-Supervised Monocular Depth Estimation
    Godard, Clement
    Mac Aodha, Oisin
    Firman, Michael
    Brostow, Gabriel
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 3827 - 3837
  • [6] Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672
  • [7] High-quality Depth from Uncalibrated Small Motion Clip
    Ha, Hyowon
    Im, Sunghoon
    Park, Jaesik
    Jeon, Hae-Gon
    Kweon, In So
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 5413 - 5421
  • [8] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [9] Image-to-Image Translation with Conditional Adversarial Networks
    Isola, Phillip
    Zhu, Jun-Yan
    Zhou, Tinghui
    Efros, Alexei A.
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 5967 - 5976
  • [10] Karsch K., 2016, Dense Image Correspondences for Computer Vision, P173, DOI [10.1007/978-3-319-23048-1_9, DOI 10.1007/978-3-319-23048-1_9]