Multi-feature decomposition and transformer-fusion: an infrared and visible image fusion network based on multi-feature decomposition and transformer

被引:0
作者
Li, Xujun [1 ]
Duan, Zhicheng [1 ]
Chang, Jia [1 ]
机构
[1] Xiangtan Univ, Sch Phys & Optoelectron, Xiangtan, Peoples R China
关键词
image fusion; infrared image; visible image; transformer; convolutional neural network; GENERATIVE ADVERSARIAL NETWORK; PERFORMANCE; FRAMEWORK;
D O I
10.1117/1.JEI.33.6.063053
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Most of the existing infrared and visible image fusion methods focus only on the extraction of detail and basic features of both images. However, these methods ignore the extraction of shallow common features, which causes the problem of weakening the ability of detail and basic feature extraction. Consequently, we propose an infrared and visible image fusion network based on multi-feature decomposition and transformer (MFDT-Fusion). The multi-feature decomposition is processed by three encoders. First, to fully extract the shallow common features of the images, multi-head gated block based on transformer improvement is proposed as the common feature encoder (CFE). Multi-head dilated attention is designed in CFE to aggregate information from different receptive fields of infrared and visible images. Then, the long-range information-capturing ability of the transformer is used to compensate for the limited receptive field problem of the convolutional neural network. The detail feature encoder and basic feature encoder are designed to decompose the two features of the images. Finally, the features are separately fused by detail and basic fusion layers. Experimental results on three public datasets show that MFDT-Fusion achieves better performance than state-of-the-art methods. (c) 2024 SPIE and IS&T
引用
收藏
页数:18
相关论文
共 46 条
[1]   A new image quality metric for image fusion: The sum of the correlations of differences [J].
Aslantas, V. ;
Bendes, E. .
AEU-INTERNATIONAL JOURNAL OF ELECTRONICS AND COMMUNICATIONS, 2015, 69 (12) :160-166
[2]  
Bavirisetti DP, 2017, 2017 20TH INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION), P701
[3]   Two-scale image fusion of visible and infrared images using saliency detection [J].
Bavirisetti, Durga Prasad ;
Dhuli, Ravindra .
INFRARED PHYSICS & TECHNOLOGY, 2016, 76 :52-64
[4]  
Dosovitskiy A, 2021, Arxiv, DOI arXiv:2010.11929
[5]   Image quality measures and their performance [J].
Eskicioglu, AM ;
Fisher, PS .
IEEE TRANSACTIONS ON COMMUNICATIONS, 1995, 43 (12) :2959-2965
[6]   A Dual-branch Network for Infrared and Visible Image Fusion [J].
Fu, Yu ;
Wu, Xiao-Jun .
2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, :10675-10680
[7]   Image fusion based on generative adversarial network consistent with perception [J].
Fu, Yu ;
Wu, Xiao-Jun ;
Durrani, Tariq .
INFORMATION FUSION, 2021, 72 :110-125
[8]   A new image fusion performance metric based on visual information fidelity [J].
Han, Yu ;
Cai, Yunze ;
Cao, Yin ;
Xu, Xiaoming .
INFORMATION FUSION, 2013, 14 (02) :127-135
[9]   Comments on 'Information measure for performance of image fusion' [J].
Hossny, M. ;
Nahavandi, S. ;
Creighton, D. .
ELECTRONICS LETTERS, 2008, 44 (18) :1066-U28
[10]  
Huang H., 2023, P IEEE CVF C COMP VI, P22690