CAMF: An Interpretable Infrared and Visible Image Fusion Network Based on Class Activation Mapping

被引:13
作者
Tang, Linfeng [1 ]
Chen, Ziang [1 ]
Huang, Jun [1 ]
Ma, Jiayi [1 ]
机构
[1] Wuhan Univ, Elect Informat Sch, Wuhan 430072, Peoples R China
基金
中国国家自然科学基金;
关键词
Image fusion; Feature extraction; Transforms; Image reconstruction; Deep learning; Task analysis; Pollution measurement; learnable fusion rule; class activation mapping; deep learning; FRAMEWORK; INPUT;
D O I
10.1109/TMM.2023.3326296
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Image fusion aims to integrate the complementary information of source images and synthesize a single fused image. Existing image fusion algorithms apply hand-crafted fusion rules to merge deep features which cause information loss and limit the fusion performance of methods since the uninterpretability of deep learning. To overcome the above shortcomings, we propose a learnable fusion rule for infrared and visible image fusion based on class activation mapping. Our proposed fusion rule can selectively preserve meaningful information and reduce distortion. More specifically, we first train an encoder-decoder network and an auxiliary classifier based on the shared encoder. Then, the class activation weights are taken out from the auxiliary classifier, which indicates the importance of each channel. Finally, the deep features extracted by the encoder are adaptively fused according to the class activation weights and the fused image is reconstructed from the fused features via the pre-trained decoder. Note that our learnable fusion rule can automatically measure the importance of each deep feature without human participation. Moreover, it fully preserves the significant features of source images such as salient targets and texture details. Extensive experiments manifest our superiority over state-of-the-art algorithms. Visualization of feature maps and their corresponding weights reveals the high interpretability of our method.
引用
收藏
页码:4776 / 4791
页数:16
相关论文
共 64 条
[1]   A new image quality metric for image fusion: The sum of the correlations of differences [J].
Aslantas, V. ;
Bendes, E. .
AEU-INTERNATIONAL JOURNAL OF ELECTRONICS AND COMMUNICATIONS, 2015, 69 (12) :160-166
[2]   Multi-Focus Image Fusion Based on Multi-Scale Gradients and Image Matting [J].
Chen, Jun ;
Li, Xuejiao ;
Luo, Linbo ;
Ma, Jiayi .
IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 :655-667
[3]   Fusion of multispectral and panchromatic satellite images using the curvelet transform [J].
Choi, M ;
Kim, RY ;
Nam, MR ;
Kim, HO .
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2005, 2 (02) :136-140
[4]   Detail preserved fusion of visible and infrared images using regional saliency extraction and multi-scale image decomposition [J].
Cui, Guangmang ;
Feng, Huajun ;
Xu, Zhihai ;
Li, Qi ;
Chen, Yueting .
OPTICS COMMUNICATIONS, 2015, 341 :199-209
[5]   Region-based multimodal image fusion using ICA bases [J].
Cvejic, Nedeljko ;
Bull, David ;
Canagarajah, Nishan .
IEEE SENSORS JOURNAL, 2007, 7 (5-6) :743-751
[6]  
Deshmukh M., 2010, Int. J. Image Process., V4, P484
[7]   Infrared and visible images fusion based on RPCA and NSCT [J].
Fu, Zhizhong ;
Wang, Xue ;
Xu, Jin ;
Zhou, Ning ;
Zhao, Yufei .
INFRARED PHYSICS & TECHNOLOGY, 2016, 77 :114-123
[8]   A new image fusion performance metric based on visual information fidelity [J].
Han, Yu ;
Cai, Yunze ;
Cao, Yin ;
Xu, Xiaoming .
INFORMATION FUSION, 2013, 14 (02) :127-135
[9]   Image compression techniques: A survey in lossless and lossy algorithms [J].
Hussain, A. J. ;
Al-Fayadh, Ali ;
Radi, Naeem .
NEUROCOMPUTING, 2018, 300 :44-69
[10]   LLVIP: A Visible-infrared Paired Dataset for Low-light Vision [J].
Jia, Xinyu ;
Zhu, Chuang ;
Li, Minzhen ;
Tang, Wenqi ;
Zhou, Wenli .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, :3489-3497