Infrared and Visible Image Fusion Based on Mask and Cross-Dynamic Fusion

被引:3
作者
Fu, Qiang [1 ]
Fu, Hanxiang [1 ]
Wu, Yuezhou [1 ]
机构
[1] Civil Aviat Flight Univ China, Sch Comp Sci, Guanghan 618307, Peoples R China
关键词
dynamic convolution; image fusion; infrared image; mask; visible image; GENERATIVE ADVERSARIAL NETWORK; PERFORMANCE; FRAMEWORK; WAVELET; NEST;
D O I
10.3390/electronics12204342
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Both single infrared and visible images have respective limitations. Fusion technology has been developed to conquer these restrictions. It is designed to generate a fused image with infrared information and texture details. Most traditional fusion methods use hand-designed fusion strategies, but some are too rough and have limited fusion performance. Recently, some researchers have proposed fusion methods based on deep learning, but some early fusion networks cannot adaptively fuse images due to unreasonable design. Therefore, we propose a mask and cross-dynamic fusion-based network called MCDFN. This network adaptively preserves the salient features of infrared images and the texture details of visible images through an end-to-end fusion process. Specifically, we designed a two-stage fusion network. In the first stage, we train the autoencoder network so that the encoder and decoder learn feature extraction and reconstruction capabilities. In the second stage, the autoencoder is fixed, and we employ a fusion strategy combining mask and cross-dynamic fusion to train the entire fusion network. This strategy is conducive to the adaptive fusion of image information between infrared images and visible images in multiple dimensions. On the public TNO dataset and the RoadScene dataset, we selected nine different fusion methods to compare with our proposed method. Experimental results show that our proposed fusion method achieves good results on both datasets.
引用
收藏
页数:22
相关论文
共 48 条
[1]  
Bavirisetti DP, 2017, 2017 20TH INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION), P701
[2]  
Ben Hamza A, 2005, INTEGR COMPUT-AID E, V12, P135
[3]   THE LAPLACIAN PYRAMID AS A COMPACT IMAGE CODE [J].
BURT, PJ ;
ADELSON, EH .
IEEE TRANSACTIONS ON COMMUNICATIONS, 1983, 31 (04) :532-540
[4]   Dynamic Convolution: Attention over Convolution Kernels [J].
Chen, Yinpeng ;
Dai, Xiyang ;
Liu, Mengchen ;
Chen, Dongdong ;
Yuan, Lu ;
Liu, Zicheng .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :11027-11036
[5]   Fusion of multispectral and panchromatic satellite images using the curvelet transform [J].
Choi, M ;
Kim, RY ;
Nam, MR ;
Kim, HO .
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2005, 2 (02) :136-140
[6]   A multiscale residual pyramid attention network for medical image fusion [J].
Fu, Jun ;
Li, Weisheng ;
Du, Jiao ;
Huang, Yuping .
BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2021, 66
[7]   MDFN: Mask deep fusion network for visible and infrared image fusion without reference ground-truth? [J].
Guo, Chaoxun ;
Fan, Dandan ;
Jiang, Zhixing ;
Zhang, David .
EXPERT SYSTEMS WITH APPLICATIONS, 2023, 211
[8]  
Haghighat M, 2014, I C APPL INF COMM TE, P424
[9]   A new image fusion performance metric based on visual information fidelity [J].
Han, Yu ;
Cai, Yunze ;
Cao, Yin ;
Xu, Xiaoming .
INFORMATION FUSION, 2013, 14 (02) :127-135
[10]   HYPERSPECTRAL IMAGE CLASSIFICATION AND DIMENSIONALITY REDUCTION - AN ORTHOGONAL SUBSPACE PROJECTION APPROACH [J].
HARSANYI, JC ;
CHANG, CI .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 1994, 32 (04) :779-785