DRF: Disentangled Representation for Visible and Infrared Image Fusion

被引:154
作者
Xu, Han [1 ]
Wang, Xinya [1 ]
Ma, Jiayi [1 ]
机构
[1] Wuhan Univ, Elect Informat Sch, Wuhan 430072, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep learning; disentangled representation; image fusion; infrared; visible; TRANSFORM;
D O I
10.1109/TIM.2021.3056645
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In this article, we propose a novel decomposition method by applying disentangled representation for visible and infrared image fusion (DRF). According to the imaging principle, we perform the decomposition depending on the source of information in the visible and infrared images. More concretely, we disentangle the images into the scene- and sensor modality (attribute)-related representations through the corresponding encoders, respectively. In this way, the unique information defined by the attribute-related representation is closer to the information captured by each type of sensor individually. Thus, the problem of inappropriate extraction of unique information can be alleviated. Then, different strategies are applied for the fusion of these different types of representations. Finally, the fused representations are fed into the pretrained generator to generate the fusion result. The qualitative and quantitative experiments on the publicly available TNO and RoadScene data sets demonstrate the comparable performance of our DRF over the state of the art in terms of both visual effect and quantitative metrics.
引用
收藏
页数:13
相关论文
共 44 条
[1]  
[Anonymous], 2018, ARXIV180408992
[2]  
Bavirisetti DP, 2017, 2017 20TH INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION), P701
[3]   Infrared and visible image fusion based on target-enhanced multiscale transform decomposition [J].
Chen, Jun ;
Li, Xuejiao ;
Luo, Linbo ;
Mei, Xiaoguang ;
Ma, Jiayi .
INFORMATION SCIENCES, 2020, 508 :64-78
[4]  
Chen R.T.Q., 2018, Advances in neural information processing systems, VVolume 31, P6571, DOI DOI 10.48550/ARXIV.1806.07366
[5]   General fusion method for infrared and visual images via latent low-rank representation and local non-subsampled shearlet transform [J].
Cheng, Boyang ;
Jin, Longxu ;
Li, Guoning .
INFRARED PHYSICS & TECHNOLOGY, 2018, 92 :68-77
[6]   Fusing Infrared and Visible Images of Different Resolutions via Total Variation Model [J].
Du, Qinglei ;
Xu, Han ;
Ma, Yong ;
Huang, Jun ;
Fan, Fan .
SENSORS, 2018, 18 (11)
[7]   Infrared and visible image fusion with the use of multi-scale edge-preserving decomposition and guided image filter [J].
Gan, Wei ;
Wu, Xiaohong ;
Wu, Wei ;
Yang, Xiaomin ;
Ren, Chao ;
He, Xiaohai ;
Liu, Kai .
INFRARED PHYSICS & TECHNOLOGY, 2015, 72 :37-51
[8]   A non-reference image fusion metric based on mutual information of image features [J].
Haghighat, Mohammad Bagher Akbari ;
Aghagolzadeh, Ali ;
Seyedarabi, Hadi .
COMPUTERS & ELECTRICAL ENGINEERING, 2011, 37 (05) :744-756
[9]   Fast saliency-aware multi-modality image fusion [J].
Han, Jungong ;
Pauwels, Eric J. ;
de Zeeuw, Paul .
NEUROCOMPUTING, 2013, 111 :70-80
[10]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778