Deep Convolutional Neural Network for Multi-Modal Image Restoration and Fusion

被引:147
作者
Deng, Xin [1 ]
Dragotti, Pier Luigi [1 ]
机构
[1] Imperial Coll London, Dept Elect & Elect Engn, London SW7 2BU, England
关键词
Image fusion; Task analysis; Image restoration; Convolutional codes; Image reconstruction; Convolutional neural networks; Image coding; Multi-modal image restoration; image fusion; multi-modal convolutional sparse coding; SUPERRESOLUTION;
D O I
10.1109/TPAMI.2020.2984244
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose a novel deep convolutional neural network to solve the general multi-modal image restoration (MIR) and multi-modal image fusion (MIF) problems. Different from other methods based on deep learning, our network architecture is designed by drawing inspirations from a new proposed multi-modal convolutional sparse coding (MCSC) model. The key feature of the proposed network is that it can automatically split the common information shared among different modalities, from the unique information that belongs to each single modality, and is therefore denoted with CU-Net, i.e., common and unique information splitting network. Specifically, the CU-Net is composed of three modules, i.e., the unique feature extraction module (UFEM), common feature preservation module (CFPM), and image reconstruction module (IRM). The architecture of each module is derived from the corresponding part in the MCSC model, which consists of several learned convolutional sparse coding (LCSC) blocks. Extensive numerical results verify the effectiveness of our method on a variety of MIR and MIF tasks, including RGB guided depth image super-resolution, flash guided non-flash image denoising, multi-focus and multi-exposure image fusion.
引用
收藏
页码:3333 / 3348
页数:16
相关论文
共 81 条
  • [1] A Dataset of Flash and Ambient Illumination Pairs from the Crowd
    Aksoy, Yagiz
    Kim, Changil
    Kellnhofer, Petr
    Paris, Sylvain
    Elgharib, Mohamed
    Pollefeys, Marc
    Matusik, Wojciech
    [J]. COMPUTER VISION - ECCV 2018, PT IX, 2018, 11213 : 644 - 660
  • [2] The Perception-Distortion Tradeoff
    Blau, Yochai
    Michaeli, Tomer
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 6228 - 6237
  • [3] Efficient nonlocal means for denoising of textural patterns
    Brox, Thomas
    Kleinschmidt, Oliver
    Cremers, Daniel
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2008, 17 (07) : 1083 - 1092
  • [4] A Naturalistic Open Source Movie for Optical Flow Evaluation
    Butler, Daniel J.
    Wulff, Jonas
    Stanley, Garrett B.
    Black, Michael J.
    [J]. COMPUTER VISION - ECCV 2012, PT VI, 2012, 7577 : 611 - 625
  • [5] Learning a Deep Single Image Contrast Enhancer from Multi-Exposure Images
    Cai, Jianrui
    Gu, Shuhang
    Zhang, Lei
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (04) : 2049 - 2062
  • [6] Multi-Focus Image Fusion Based on Spatial Frequency in Discrete Cosine Transform Domain
    Cao, Liu
    Jin, Longxu
    Tao, Hongjiang
    Li, Guoning
    Zhuang, Zhuang
    Zhang, Yanfu
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2015, 22 (02) : 220 - 224
  • [7] Image restoration by sparse 3D transform-domain collaborative filtering
    Dabov, Kostadin
    Foi, Alessandro
    Katkovnik, Vladimir
    Egiazarian, Karen
    [J]. IMAGE PROCESSING: ALGORITHMS AND SYSTEMS VI, 2008, 6812
  • [8] Optimum Wavelet-Based Homomorphic Medical Image Fusion Using Hybrid Genetic-Grey Wolf Optimization Algorithm
    Daniel, Ebenezer
    [J]. IEEE SENSORS JOURNAL, 2018, 18 (16) : 6804 - 6811
  • [9] An iterative thresholding algorithm for linear inverse problems with a sparsity constraint
    Daubechies, I
    Defrise, M
    De Mol, C
    [J]. COMMUNICATIONS ON PURE AND APPLIED MATHEMATICS, 2004, 57 (11) : 1413 - 1457
  • [10] RADAR: Robust Algorithm for Depth Image Super Resolution Based on FRI Theory and Multimodal Dictionary Learning
    Deng, Xin
    Song, Pingfan
    Rodrigues, Miguel R. D.
    Dragotti, Pier Luigi
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (08) : 2447 - 2462