Semantic Region Adaptive Fusion of Infrared and Visible Images via Dual-DeepLab Guidance

被引:4
作者
Cao, Wenzi [1 ]
Zheng, Minghui [1 ]
Liao, Qing [1 ]
机构
[1] Wuhan Inst Technol, Hubei Key Lab Opt Informat & Pattern Recognit, Wuhan 430205, Peoples R China
关键词
Semantics; Task analysis; Feature extraction; Lighting; Training; Semantic segmentation; Image reconstruction; High-level semantic perception; image fusion; infrared; prioritized preservation scheme; region adaptation; visible; GENERATIVE ADVERSARIAL NETWORK; CLASSIFICATION; PERFORMANCE; NEST;
D O I
10.1109/TIM.2023.3318709
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Given the potential diminishment of semantic information caused by previous nonregion-specific maximum or weighted intensity losses under different lighting conditions, we propose a novel high-level semantic perception fusion framework, termed semantic region adaptive fusion (SRAFusion). We establish a prioritized preservation scheme for high-level semantic information (HLSI), gradient information, and intensity information, ranked in descending order of priority. Based on this prioritization scheme and prior knowledge of semantic distribution in the source images, we construct a ground truth for the fusion task. Specifically, we capture the HLSI distribution of the source images using two independent semantic segmentation networks. Subsequently, we introduce semantic region decision block (SRDB) to partition the original scene into region with bimodal HLSI, region with unimodal HLSI, and region lacking HLSI. We then design specific loss functions to constrain the aforementioned regions, facilitating the integration of complete semantic information. Furthermore, taking into account the susceptibility of the visible segmentation network to lighting conditions, we use a two-stage training strategy involving coarse-tuning and fine-tuning. This method aims to optimize one-stage training strategy and achieve a more accurate region delineation. Finally, qualitative and quantitative experiments conducted on publicly available datasets such as MFNet, RoadScene, and TNO demonstrate the superiority of our SRAFusion over state-of-the-art methods. Our code will be available: https://github.com/WenziCao/SRAFusion.
引用
收藏
页数:16
相关论文
共 71 条
[21]   Fast saliency-aware multi-modality image fusion [J].
Han, Jungong ;
Pauwels, Eric J. ;
de Zeeuw, Paul .
NEUROCOMPUTING, 2013, 111 :70-80
[22]   A new image fusion performance metric based on visual information fidelity [J].
Han, Yu ;
Cai, Yunze ;
Cao, Yin ;
Xu, Xiaoming .
INFORMATION FUSION, 2013, 14 (02) :127-135
[23]  
Haris M., 2021, NEURAL INFORM PROCES, V1516, P387, DOI [10.1007/978-3-030-92307-5_45, 10.1007/978-3-030-92307-545, DOI 10.1007/978-3-030-92307-5_45]
[24]   HYPERSPECTRAL IMAGE CLASSIFICATION AND DIMENSIONALITY REDUCTION - AN ORTHOGONAL SUBSPACE PROJECTION APPROACH [J].
HARSANYI, JC ;
CHANG, CI .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 1994, 32 (04) :779-785
[25]  
HE KM, 2016, PROC CVPR IEEE, P770, DOI [DOI 10.1109/CVPR.2016.90, 10.1109/CVPR.2016.90]
[26]  
HE Y, 2021, INFR PHYS TECHNOL, V116
[27]   Exploring modality-shared appearance features and modality-invariant relation features for cross-modality person Re-IDentification [J].
Huang, Nianchang ;
Liu, Jianan ;
Luo, Yongjiang ;
Zhang, Qiang ;
Han, Jungong .
PATTERN RECOGNITION, 2023, 135
[28]   SEDRFuse: A Symmetric Encoder-Decoder With Residual Block Network for Infrared and Visible Image Fusion [J].
Jian, Lihua ;
Yang, Xiaomin ;
Liu, Zheng ;
Jeon, Gwanggil ;
Gao, Mingliang ;
Chisholm, David .
IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2021, 70
[29]   Improving Multispectral Pedestrian Detection by Addressing Modality Imbalance Problems [J].
Zhou, Kailai ;
Chen, Linsen ;
Cao, Xun .
COMPUTER VISION - ECCV 2020, PT XVIII, 2020, 12363 :787-803
[30]  
KOTABE H, 2016, P COGSCI