MEFuse: end-to-end infrared and visible image fusion method based on multibranch encoder

被引:5
作者
Hong, Yulu [1 ]
Wu, Xiao-Jun [1 ]
Xu, Tianyang [1 ]
机构
[1] Jiangnan Univ, Sch Artificial Intelligence & Comp Sci, Jiangsu Prov Engn Lab Pattern Recognit & Computat, Wuxi, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
image fusion; multibranch encoder; convolutional neural network; PERFORMANCE; INFORMATION; FRAMEWORK;
D O I
10.1117/1.JEI.31.3.033043
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In recent years, deep learning-based image fusion methods have achieved promising fusion performance. However, existing fusion approaches always focus on increasing the network depth to enhance the feature extraction capacity. In addition, fusion is performed in a fixed feature space, which further degrades the performance. To this end, we proposed an end-to-end infrared and visible image fusion method based on a multibranch encoder, coined MEFuse. In principle, MEFuse increased the number of branches of the encoder, thus mapping the source images into different feature spaces, increasing the diversity of the extracted feature maps, and enhancing the learning capacity of the network. In addition, a multistage fusion strategy was designed to achieve multigranularity image fusion. First, the cascaded multimodal images were merged in the input stage; thus the primary visual fusion was realized in the coding process. Second, the feature maps from multiple coding branches were fused to fuse multimodal information comprehensively. Finally, the long-range lateral connection sent the input appearance of each encoding branch to the decoder to guide the image reconstruction process. Experiments on the benchmark data sets demonstrate that, compared with the existing methods, the proposed MEFuse achieved superior performance in both human perception and objective evaluation. (C) 2022 SPIE and IS&T
引用
收藏
页数:14
相关论文
共 45 条
[1]  
Ben Hamza A, 2005, INTEGR COMPUT-AID E, V12, P135
[2]   Image quality measures and their performance [J].
Eskicioglu, AM ;
Fisher, PS .
IEEE TRANSACTIONS ON COMMUNICATIONS, 1995, 43 (12) :2959-2965
[3]  
Haghighat M, 2014, I C APPL INF COMM TE, P424
[4]  
Huang GL, 2017, IEEE ICC
[5]  
Kristan M, 2020, P EUR C COMP VIS WOR, V12539, P547, DOI 10.1007/978-3-030-68238-5_39
[6]  
Kumar P, 2006, LECT NOTES COMPUT SC, V4338, P528
[7]   A new thermal infrared and visible spectrum images-based pedestrian detection system [J].
Lahmyed, Redouan ;
El Ansari, Mohamed ;
Ellahyani, Ayoub .
MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (12) :15861-15885
[8]   RGB-T object tracking: Benchmark and baseline [J].
Li, Chenglong ;
Liang, Xinyan ;
Lu, Yijuan ;
Zhao, Nan ;
Tang, Jin .
PATTERN RECOGNITION, 2019, 96
[9]   NestFuse: An Infrared and Visible Image Fusion Architecture Based on Nest Connection and Spatial/Channel Attention Models [J].
Li, Hui ;
Wu, Xiao-Jun ;
Durrani, Tariq .
IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2020, 69 (12) :9645-9656
[10]  
Li H, 2018, INT C PATT RECOG, P2705, DOI 10.1109/ICPR.2018.8546006