An Infrared and Visible Image Fusion Method Based on Semantic-Sensitive Mask Selection and Bidirectional-Collaboration Region Fusion

被引:0
作者
Li, Xuan [1 ]
Zhang, Guomin [1 ]
Chen, Weiwei [2 ]
Cheng, Li [1 ]
Xie, Yining [3 ]
Ma, Jiayi [4 ]
机构
[1] Wuhan Inst Technol, Sch Elect & Informat Engn, Wuhan 430205, Peoples R China
[2] Fiberhome Telecommun Technol Co Ltd, Wuhan 430205, Peoples R China
[3] Northeast Forestry Univ, Coll Mech & Elect Engn, Harbin 150040, Peoples R China
[4] Wuhan Univ, Elect Informat Sch, Wuhan 430072, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature extraction; Image fusion; Circuits and systems; Deep learning; Generators; Network architecture; Measurement; Image reconstruction; Generative adversarial networks; Visual effects; semantic-sensitive mask selection; complementary-mask map; bidirectional-collaboration fusion;
D O I
10.1109/TCSVT.2024.3520252
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Mask is considered as an important prior for fusion, which could selectively enhance specific regions to generate ideal fused images. However, masks used in the existing methods exhibit limitations in the precise representation of targets, and more importantly, these masks are generated from a single modality, which restricts the effective integration of multi-modal information. To address this issue, we propose a competitive mask-guidance fusion method for infrared and visible images. A multi-modal semantic-sensitive mask selection network is proposed to generate complementary-mask maps, which organically integrate advantageous target regions of different modalities by competitively comparing the qualities of masks. In this network, a pseudosiamese architecture is designed to obtain respective target masks, and specifically, a spatial-aligned-based feature aggregation module is devised to produce high-quality pseudo-labels which are served as references for the generation of the complementary-mask maps. Furthermore, we propose a bidirectional-collaboration region fusion strategy, which enhances the expression of advantageous target regions from each modality inforeground while suppressing the contribution of corresponding regions from the other modality in background. Compared to methods on public datasets, the results show that our method significantly enhances the description of semantic-sensitive targets in fused images, including the saliency and the integrity of structural information. Code are available at https://github.com/xbsj-cool/MSCRFusion.
引用
收藏
页码:5518 / 5532
页数:15
相关论文
共 51 条
[1]   A new image quality metric for image fusion: The sum of the correlations of differences [J].
Aslantas, V. ;
Bendes, E. .
AEU-INTERNATIONAL JOURNAL OF ELECTRONICS AND COMMUNICATIONS, 2015, 69 (12) :160-166
[2]   HitFusion: Infrared and Visible Image Fusion for High-Level Vision Tasks Using Transformer [J].
Chen, Jun ;
Ding, Jianfeng ;
Ma, Jiayi .
IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 :10145-10159
[3]   LENFusion: A Joint Low-Light Enhancement and Fusion Network for Nighttime Infrared and Visible Image Fusion [J].
Chen, Jun ;
Yang, Liling ;
Liu, Wei ;
Tian, Xin ;
Ma, Jiayi .
IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 :1-15
[4]   Infrared and visible image fusion based on target-enhanced multiscale transform decomposition [J].
Chen, Jun ;
Li, Xuejiao ;
Luo, Linbo ;
Mei, Xiaoguang ;
Ma, Jiayi .
INFORMATION SCIENCES, 2020, 508 :64-78
[5]   Region-based multimodal image fusion using ICA bases [J].
Cvejic, Nedeljko ;
Bull, David ;
Canagarajah, Nishan .
IEEE SENSORS JOURNAL, 2007, 7 (5-6) :743-751
[6]  
Desa SM, 2008, INT C PATT RECOG, P2837
[7]  
Deshmukh M., 2010, International Journal of Image Processing (IJIP), V4, P484
[8]   Infrared and visible images fusion based on RPCA and NSCT [J].
Fu, Zhizhong ;
Wang, Xue ;
Xu, Jin ;
Zhou, Ning ;
Zhao, Yufei .
INFRARED PHYSICS & TECHNOLOGY, 2016, 77 :114-123
[9]   DCDR-GAN: A Densely Connected Disentangled Representation Generative Adversarial Network for Infrared and Visible Image Fusion [J].
Gao, Yuan ;
Ma, Shiwei ;
Liu, Jingjing .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (02) :549-561
[10]  
Haghighat M, 2014, I C APPL INF COMM TE, P424