FAFusion: Learning for Infrared and Visible Image Fusion via Frequency Awareness

被引:0
作者
Xiao, Guobao [1 ]
Tang, Zhimin [2 ]
Guo, Hanlin [3 ]
Yu, Jun [4 ,5 ]
Shen, Heng Tao [1 ]
机构
[1] Tongji Univ, Sch Elect & Informat Engn, Shanghai 201804, Peoples R China
[2] Fuzhou Univ, Coll Comp & Data Sci, Fuzhou 350108, Peoples R China
[3] Xiamen Univ Technol, Sch Elect Engn & Automat, Xiamen 361024, Peoples R China
[4] Hangzhou Dianzi Univ, Sch Comp Sci & Technol, Hangzhou 310018, Peoples R China
[5] Harbin Inst Technol, Dept Comp Sci & Technol, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep learning; image fusion; frequency-aware; TRANSFORM; NETWORK;
D O I
10.1109/TIM.2024.3374294
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In this article, we introduce a novel Frequency-Aware Infrared and Visible Image Fusion Network (FAFusion) designed to explore both low and high-frequency information present in infrared and visible images. Our approach involves leveraging the low-frequency information, which preserves the contour and overall brightness of the source images, in the encoder. This enables FAFusion to prioritize salient targets. Simultaneously, we feed the high-frequency information, preserving object edges and texture details, into the decoder, enhancing the fused images with richer details. To maintain the frequency information from the source images, we propose a maximum frequency loss function. This function considers the frequency components between the fused image and the source images, ensuring the preservation of critical frequency details. Additionally, we introduce a Residual Multiscale Feature Extraction (RMFE) module to capture diverse contextual information from the source images. The resulting FAFusion demonstrates the capability to generate fused images that exhibit both rich texture details and emphasize salient targets. We validate the effectiveness of our approach through extensive experiments on three publicly available datasets (TNO, MSRS, and M3FD). Comparative analyses against nine state-of-the-art methods highlight the superior visual effects, quantitative metrics, and generalization performance achieved by the proposed FAFusion.The source codes will be available at https://github.com/guobaoxiao/FAFusion.
引用
收藏
页码:1 / 11
页数:11
相关论文
共 47 条
[1]   Two-scale image fusion of visible and infrared images using saliency detection [J].
Bavirisetti, Durga Prasad ;
Dhuli, Ravindra .
INFRARED PHYSICS & TECHNOLOGY, 2016, 76 :52-64
[2]  
Bavirisetti G., 2017, P 20 INT C INF FUS, P1
[3]   THE WAVELET TRANSFORM, TIME-FREQUENCY LOCALIZATION AND SIGNAL ANALYSIS [J].
DAUBECHIES, I .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1990, 36 (05) :961-1005
[4]  
Fu Y, 2022, Arxiv, DOI arXiv:2107.13967
[5]  
Ha Q, 2017, IEEE INT C INT ROBOT, P5108, DOI 10.1109/IROS.2017.8206396
[6]   Focal Frequency Loss for Image Reconstruction and Synthesis [J].
Jiang, Liming ;
Dai, Bo ;
Wu, Wayne ;
Loy, Chen Change .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :13899-13909
[7]   RFN-Nest: An end-to-end residual fusion network for infrared and visible images [J].
Li, Hui ;
Wu, Xiao-Jun ;
Kittler, Josef .
INFORMATION FUSION, 2021, 73 :72-86
[8]   CGTF: Convolution-Guided Transformer for Infrared and Visible Image Fusion [J].
Li, Jing ;
Zhu, Jianming ;
Li, Chang ;
Chen, Xun ;
Yang, Bin .
IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2022, 71
[9]   Image Fusion with Guided Filtering [J].
Li, Shutao ;
Kang, Xudong ;
Hu, Jianwen .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2013, 22 (07) :2864-2875
[10]   Group-Sparse Representation With Dictionary Learning for Medical Image Denoising and Fusion [J].
Li, Shutao ;
Yin, Haitao ;
Fang, Leyuan .
IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, 2012, 59 (12) :3450-3459