Two-Stream Maximal Feature Attention-Guided Contrastive-Learning GAN for Image Fusion

被引:0
作者
Yang, Danqing [1 ]
Zhu, Naibo [2 ]
Wang, Xiaorui [1 ]
Li, Shuang [2 ]
机构
[1] Xidian Univ, Sch Optoelect Engn, Xian 710071, Peoples R China
[2] PLA Acad Mil Sci, Res Inst Syst Engn, Beijing 100091, Peoples R China
关键词
Attention mechanism; contrastive learning; generative adversarial network (GAN); infrared and visible image (VI) fusion; maximal feature map; residual-dense block; GENERATIVE ADVERSARIAL NETWORK; INFORMATION; TRANSFORM;
D O I
10.1109/JSEN.2024.3394874
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Off-the-shelf generative adversarial network (GAN)-based image fusion methods solely utilize the diverse modal images as reference samples, aiming to make the distribution of the generated image closely resemble that of the source images through the adversarial process. This leads to the resultant images that resemble the compromised outcomes of visible images (VIs) and infrared counterparts. This article reports a seminal study on fusing infrared and VIs, which ingeniously introduces the maximal feature constraints built upon contrastive learning into a GAN-based image fusion framework, facilitating more effective semantic knowledge extraction from both image types. The generator is devised as a lateral feature aggregation codec structure, aiming to enhance interaction among multilevel features and compensate for feature loss during transmission. Unlike methods that focused only on edge information, a novel two-stream maximal feature-guided attention network upon the multilevel exclusive features is designed to strengthen the attention toward prominent information during fusion. This allows finer texture detail restoration and maintaining highlighted targets while attenuating noise. A hybrid loss function, defined as a combination of content loss and adversarial-perceptual-contrastive constraint, is minimized to generate fusion results with source imagery statistics by employing an objective that concentrates on feature distribution rather than simply evaluating pixel intensity difference. Extensive trials demonstrate that the proposed algorithm performs better than previous top-performing methods in both visual and objective aspects on four benchmark datasets.
引用
收藏
页码:21533 / 21548
页数:16
相关论文
共 59 条
[1]   Infrared and visible image fusion based on target-enhanced multiscale transform decomposition [J].
Chen, Jun ;
Li, Xuejiao ;
Luo, Linbo ;
Mei, Xiaoguang ;
Ma, Jiayi .
INFORMATION SCIENCES, 2020, 508 (508) :64-78
[2]   MUFusion: A general unsupervised image fusion network based on memory unit [J].
Cheng, Chunyang ;
Xu, Tianyang ;
Wu, Xiao-Jun .
INFORMATION FUSION, 2023, 92 :80-92
[3]   Image quality measures and their performance [J].
Eskicioglu, AM ;
Fisher, PS .
IEEE TRANSACTIONS ON COMMUNICATIONS, 1995, 43 (12) :2959-2965
[4]   MDRANet: A multiscale dense residual attention network for magnetic resonance and nuclear medicine image fusion [J].
Fu, Jun ;
Li, Weisheng ;
Peng, Xiuxiu ;
Du, Jiao ;
Ouyang, Aijia ;
Wang, Qian ;
Chen, Xin .
BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2023, 80
[5]   A Dual-branch Network for Infrared and Visible Image Fusion [J].
Fu, Yu ;
Wu, Xiao-Jun .
2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, :10675-10680
[6]   Image fusion based on generative adversarial network consistent with perception [J].
Fu, Yu ;
Wu, Xiao-Jun ;
Durrani, Tariq .
INFORMATION FUSION, 2021, 72 :110-125
[7]   SaReGAN: a salient regional generative adversarial network for visible and infrared image fusion [J].
Gao, Mingliang ;
Zhou, Yi'nan ;
Zhai, Wenzhe ;
Zeng, Shuai ;
Li, Qilei .
MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (22) :61659-61671
[8]   Efficient Neural Network Compression Inspired by Compressive Sensing [J].
Gao, Wei ;
Guo, Yang ;
Ma, Siwei ;
Li, Ge ;
Kwong, Sam .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (02) :1965-1979
[9]   DCDR-GAN: A Densely Connected Disentangled Representation Generative Adversarial Network for Infrared and Visible Image Fusion [J].
Gao, Yuan ;
Ma, Shiwei ;
Liu, Jingjing .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (02) :549-561
[10]   Variational gated autoencoder-based feature extraction model for inferring disease-miRNA associations based on multiview features [J].
Guo, Yanbu ;
Zhou, Dongming ;
Ruan, Xiaoli ;
Cao, Jinde .
NEURAL NETWORKS, 2023, 165 :491-505