VIF-Net: An Unsupervised Framework for Infrared and Visible Image Fusion

被引:201
作者
Hou, Ruichao [1 ]
Zhou, Dongming [2 ]
Nie, Rencan [2 ]
Liu, Dong [2 ]
Xiong, Lei [3 ]
Guo, Yanbu [2 ]
Yu, Chuanbo [4 ]
机构
[1] Nanjing Univ, Dept Comp Sci & Technol, Nanjing 210023, Peoples R China
[2] Yunnan Univ, Sch Informat, Kunming 650504, Yunnan, Peoples R China
[3] Chinese Acad Sci, Shenzhen Inst Adv Technol, Shenzhen 518055, Peoples R China
[4] Tianjin Univ, Sch Elect & Informat Engn, Tianjin 300072, Peoples R China
基金
中国国家自然科学基金;
关键词
Unsupervised learning; image fusion; convolutional neural networks; infrared images; visible images; CONTOURLET TRANSFORM; PERFORMANCE;
D O I
10.1109/TCI.2020.2965304
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Visible images provide abundant texture details and environmental information, while infrared images benefit from night-time visibility and suppression of highly dynamic regions; it is a meaningful task to fuse these two types of features from different sensors to generate an informative image. In this article, we propose an unsupervised end-to-end learning framework for infrared and visible image fusion. We first construct enough benchmark training datasets using the visible and infrared frames, which can address the limitation of the training dataset. Additionally, due to the lack of labeled datasets, our architecture is derived from a robust mixed loss function that consists of the modified structural similarity (M-SSIM) metric and the total variation (TV) by designing an unsupervised learning process that can adaptively fuse thermal radiation and texture details and suppress noise interference. In addition, our method is an end to end model, which avoids setting hand-crafted fusion rules and reducing computational cost. Furthermore, extensive experimental results demonstrate that the proposed architecture performs better than state-of-the-art methods in both subjective and objective evaluations.
引用
收藏
页码:640 / 651
页数:12
相关论文
共 53 条
[1]  
Abadi M, 2016, ACM SIGPLAN NOTICES, V51, P1, DOI [10.1145/2951913.2976746, 10.1145/3022670.2976746]
[2]   DISCRETE COSINE TRANSFORM [J].
AHMED, N ;
NATARAJAN, T ;
RAO, KR .
IEEE TRANSACTIONS ON COMPUTERS, 1974, C 23 (01) :90-93
[3]   Convolutional Autoencoder-Based Multispectral Image Fusion [J].
Azarang, Arian ;
Manoochehri, Hafez E. ;
Kehtarnavaz, Nasser .
IEEE ACCESS, 2019, 7 :35673-35683
[4]  
Bavirisetti DP, 2017, 2017 20TH INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION), P701
[5]   Fusion of Infrared and Visible Sensor Images Based on Anisotropic Diffusion and Karhunen-Loeve Transform [J].
Bavirisetti, Durga Prasad ;
Dhuli, Ravindra .
IEEE SENSORS JOURNAL, 2016, 16 (01) :203-209
[6]   THE LAPLACIAN PYRAMID AS A COMPACT IMAGE CODE [J].
BURT, PJ ;
ADELSON, EH .
IEEE TRANSACTIONS ON COMMUNICATIONS, 1983, 31 (04) :532-540
[7]   Image recovery via total variation minimization and related problems [J].
Chambolle, A ;
Lions, PL .
NUMERISCHE MATHEMATIK, 1997, 76 (02) :167-188
[8]   The nonsubsampled contourlet transform: Theory, design, and applications [J].
da Cunha, Arthur L. ;
Zhou, Jianping ;
Do, Minh N. .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2006, 15 (10) :3089-3101
[9]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[10]   The contourlet transform: An efficient directional multiresolution image representation [J].
Do, MN ;
Vetterli, M .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2005, 14 (12) :2091-2106