CAEFusion: A New Convolutional Autoencoder-Based Infrared and Visible Light Image Fusion Algorithm

被引:0
作者
Wu, Chun-Ming [1 ]
Ren, Mei-Ling [2 ]
Lei, Jin [2 ]
Jiang, Zi-Mu [3 ]
机构
[1] Northeast Elect Power Univ, Key Lab Modern Power Syst Simulat & Control & Rene, Sch Elect Engn, Minist Educ, Jilin 132012, Peoples R China
[2] Northeast Elect Power Univ, Sch Elect Engn, Jilin 132012, Peoples R China
[3] Bozhou Univ, Sch Elect Informat Engn, Bozhou 236800, Peoples R China
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2024年 / 80卷 / 02期
关键词
Image fusion; deep learning; auto-encoder (AE); infrared; visible light;
D O I
10.32604/cmc.2024.053708
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To address the issues of incomplete information, blurred details, loss of details, and insufficient contrast in infrared and visible image fusion, an image fusion algorithm based on a convolutional autoencoder is proposed. The region attention module is meant to extract the background feature map based on the distinct properties of the background feature map and the detail feature map. A multi-scale convolution attention module is suggested to enhance the communication of feature information. At the same time, the feature transformation module is introduced to learn more robust feature representations, aiming to preserve the integrity of image information. This study uses three available datasets from TNO, FLIR, and NIR to perform thorough quantitative and qualitative trials with five additional algorithms. The methods are assessed based on four indicators: information entropy (EN), standard deviation (SD), spatial frequency (SF), and average gradient (AG). Object detection experiments were done on the M3FD dataset to further verify the algorithm's performance in comparison with five other algorithms. The algorithm's accuracy was evaluated using the mean average precision at a threshold of 0.5 (mAP@0.5) index. Comprehensive experimental findings show that CAEFusion performs well in subjective visual and objective evaluation criteria and has promising potential in downstream object detection tasks.
引用
收藏
页码:2857 / 2872
页数:16
相关论文
共 22 条
[1]   Multi-scale Guided Image and Video Fusion: A Fast and Efficient Approach [J].
Bavirisetti, Durga Prasad ;
Xiao, Gang ;
Zhao, Junhao ;
Dhuli, Ravindra ;
Liu, Gang .
CIRCUITS SYSTEMS AND SIGNAL PROCESSING, 2019, 38 (12) :5576-5605
[2]   Fusion of Infrared and Visible Images Using Fuzzy Based Siamese Convolutional Network [J].
Bhalla, Kanika ;
Koundal, Deepika ;
Bhatia, Surbhi ;
Rahmani, Mohammad Khalid Imam ;
Tahir, Muhammad .
CMC-COMPUTERS MATERIALS & CONTINUA, 2022, 70 (03) :5503-5518
[3]  
Brown M, 2011, PROC CVPR IEEE, P177, DOI 10.1109/CVPR.2011.5995637
[4]   Sparse mixed attention aggregation network for multimodal images fusion tracking [J].
Feng, Mingzheng ;
Su, Jianbo .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 127
[5]   Infrared and Visible Image Fusion with Deep Neural Network in Enhanced Flight Vision System [J].
Gao, Xuyang ;
Shi, Yibing ;
Zhu, Qi ;
Fu, Qiang ;
Wu, Yuezhou .
REMOTE SENSING, 2022, 14 (12)
[6]   Visible and Infrared Image Fusion of Forest Fire Scenes Based on Generative Adversarial Networks with Multi-Classification and Multi-Level Constraints [J].
Jin, Qi ;
Tan, Sanqing ;
Zhang, Gui ;
Yang, Zhigao ;
Wen, Yijun ;
Xiao, Huashun ;
Wu, Xin .
FORESTS, 2023, 14 (10)
[7]  
Lahoud F, 2019, Arxiv, DOI arXiv:1905.03590
[8]  
Li C, 2022, Arxiv, DOI [arXiv:2209.07947, 10.48550/arXiv.2209.07947, DOI 10.48550/ARXIV.2209.07947]
[9]   Robust registration for infrared and visible images based on salient gradient mutual information and local search [J].
Li, Jiawen ;
Wu, Xiaoyan ;
Liao, Penghao ;
Song, Honghong ;
Yang, Xiaomin ;
Zhang, Rongzhu .
INFRARED PHYSICS & TECHNOLOGY, 2023, 131
[10]   Coupled GAN With Relativistic Discriminators for Infrared and Visible Images Fusion [J].
Li, Qilei ;
Lu, Lu ;
Li, Zhen ;
Wu, Wei ;
Liu, Zheng ;
Jeon, Gwanggil ;
Yang, Xiaomin .
IEEE SENSORS JOURNAL, 2021, 21 (06) :7458-7467