Unsupervised Deep Image Fusion With Structure Tensor Representations

被引:123
作者
Jung, Hyungjoo [1 ]
Kim, Youngjung [2 ]
Jang, Hyunsung [1 ,3 ]
Ha, Namkoo [3 ]
Sohn, Kwanghoon [1 ]
机构
[1] Yonsei Univ, Sch Elect & Elect Engn, Seoul 120749, South Korea
[2] Agcy Def Dev, Inst Def Adv Technol Res, Daejeon 34060, South Korea
[3] LIG Nex1 Co Ltd, EO IR Res & Dev Lab, Yongin 16911, South Korea
基金
新加坡国家研究基金会;
关键词
Image fusion; image contrast; structure tensor; convolutional neural network; and unsupervised learning; PERFORMANCE; GRADIENT; INFORMATION; FRAMEWORK; CURVELET; NETWORK;
D O I
10.1109/TIP.2020.2966075
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Convolutional neural networks (CNNs) have facilitated substantial progress on various problems in computer vision and image processing. However, applying them to image fusion has remained challenging due to the lack of the labelled data for supervised learning. This paper introduces a deep image fusion network (DIF-Net), an unsupervised deep learning framework for image fusion. The DIF-Net parameterizes the entire processes of image fusion, comprising of feature extraction, feature fusion, and image reconstruction, using a CNN. The purpose of DIF-Net is to generate an output image which has an identical contrast to high-dimensional input images. To realize this, we propose an unsupervised loss function using the structure tensor representation of the multi-channel image contrasts. Different from traditional fusion methods that involve time-consuming optimization or iterative procedures to obtain the results, our loss function is minimized by a stochastic deep learning solver with large-scale examples. Consequently, the proposed method can produce fused images that preserve source image details through a single forward network trained without reference ground-truth labels. The proposed method has broad applicability to various image fusion problems, including multi-spectral, multi-focus, and multi-exposure image fusions. Quantitative and qualitative evaluations show that the proposed technique outperforms existing state-of-the-art approaches for various applications.
引用
收藏
页码:3845 / 3858
页数:14
相关论文
共 68 条
  • [1] [Anonymous], 2015, 32 INT C MACH LEARN
  • [2] [Anonymous], 2015, PROC CVPR IEEE
  • [3] [Anonymous], ELECT IMAGING
  • [4] [Anonymous], P INT C IM PROC JUN
  • [5] A new image quality metric for image fusion: The sum of the correlations of differences
    Aslantas, V.
    Bendes, E.
    [J]. AEU-INTERNATIONAL JOURNAL OF ELECTRONICS AND COMMUNICATIONS, 2015, 69 (12) : 160 - 166
  • [6] Brown M., 2011, P CVPR
  • [7] Chambolle A, 2004, J MATH IMAGING VIS, V20, P89
  • [8] Connah D, 2014, LECT NOTES COMPUT SC, V8693, P65, DOI 10.1007/978-3-319-10602-1_5
  • [9] A NOTE ON THE GRADIENT OF A MULTIIMAGE
    DIZENZO, S
    [J]. COMPUTER VISION GRAPHICS AND IMAGE PROCESSING, 1986, 33 (01): : 116 - 125
  • [10] Learning a Deep Convolutional Network for Image Super-Resolution
    Dong, Chao
    Loy, Chen Change
    He, Kaiming
    Tang, Xiaoou
    [J]. COMPUTER VISION - ECCV 2014, PT IV, 2014, 8692 : 184 - 199