Multigrained Attention Network for Infrared and Visible Image Fusion

被引:95
作者
Li, Jing [1 ]
Huo, Hongtao [1 ]
Li, Chang [2 ]
Wang, Renhua [1 ]
Sui, Chenhong [3 ]
Liu, Zhao [4 ]
机构
[1] Peoples Publ Secur Univ China, Dept Informat Technol & Cyber Secur, Beijing 100038, Peoples R China
[2] Hefei Univ Technol, Dept Biomed Engn, Hefei 230009, Peoples R China
[3] Yantai Univ, Sch Optoelect Informat Sci & Technol, Yantai 264000, Peoples R China
[4] Peoples Publ Secur Univ China, Grad Sch, Beijing 100038, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature loss; generative adversarial network (GAN); image fusion; multigrained attention mechanism; PERFORMANCE; TRANSFORM; MODEL;
D O I
10.1109/TIM.2020.3029360
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Methods based on generative adversarial network (GAN) have been widely used in infrared and visible images fusion. However, these methods cannot perceive the discriminative parts of an image. Therefore, we introduce a multigrained attention module into encoder-decoder network to fuse infrared and visible images (MgAN-Fuse). The infrared and visible images are encoded by two independent encoder networks due to their diverse modalities. Then, the results of the two encoders are concatenated to calculate the fused result by the decoder. To exploit the features of multiscale layers fully and force the model focus on the discriminative regions, we integrate attention modules into multiscale layers of the encoder to obtain multigrained attention maps, and then, the multigrained attention maps are concatenated with the corresponding multiscale features of the decoder network. Thus, the proposed method can preserve the foreground target information of the infrared image and capture the context information of the visible image. Furthermore, we design an additional feature loss in the training process to preserve the important features of the visible image, and a dual adversarial architecture is employed to help the model capture enough infrared intensity information and visible details simultaneously. The ablation studies illustrate the validity of the multigrained attention network and feature loss function. Extensive experiments on two infrared and visible image data sets demonstrate that the proposed MgAN-Fuse has a better performance than state-of-the-art methods.
引用
收藏
页数:12
相关论文
共 46 条
[1]  
[Anonymous], Conf. Comput. Vis. (ICCV)
[2]  
Arjovsky M, 2017, PR MACH LEARN RES, V70
[3]   An Attentive Survey of Attention Models [J].
Chaudhari, Sneha ;
Mithal, Varun ;
Polatkan, Gungor ;
Ramanath, Rohan .
ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2021, 12 (05)
[4]  
Chipman LJ, 1995, INTERNATIONAL CONFERENCE ON IMAGE PROCESSING - PROCEEDINGS, VOLS I-III, pC248
[5]   Detail preserved fusion of visible and infrared images using regional saliency extraction and multi-scale image decomposition [J].
Cui, Guangmang ;
Feng, Huajun ;
Xu, Zhihai ;
Li, Qi ;
Chen, Yueting .
OPTICS COMMUNICATIONS, 2015, 341 :199-209
[6]  
Durugkar I., 2016, ARXIV161101673
[7]   Fast saliency-aware multi-modality image fusion [J].
Han, Jungong ;
Pauwels, Eric J. ;
de Zeeuw, Paul .
NEUROCOMPUTING, 2013, 111 :70-80
[8]   A new image fusion performance metric based on visual information fidelity [J].
Han, Yu ;
Cai, Yunze ;
Cao, Yin ;
Xu, Xiaoming .
INFORMATION FUSION, 2013, 14 (02) :127-135
[9]  
Gulrajani I, 2017, ADV NEUR IN, V30
[10]   Perceptual Losses for Real-Time Style Transfer and Super-Resolution [J].
Johnson, Justin ;
Alahi, Alexandre ;
Li Fei-Fei .
COMPUTER VISION - ECCV 2016, PT II, 2016, 9906 :694-711