Rethinking the approach to lightweight multi-branch heterogeneous image fusion frameworks: Infrared and visible image fusion via the parallel Mamba-KAN framework

被引:0
作者
Sun, Yichen [1 ,2 ,3 ]
Dong, Mingli [1 ,2 ,3 ]
Zhu, Lianqing [2 ,3 ]
机构
[1] Beijing Univ Technol, Coll Mech & Energy Engn, Beijing 100124, Peoples R China
[2] Beijing Informat Sci & Technol Univ, Key Lab, Minist Educ Optoelect Measurement Technol & Instru, Beijing 100016, Peoples R China
[3] Guangzhou Nansha Intelligent Photon Sensing Res In, Guangzhou 511462, Guangdong, Peoples R China
关键词
Image fusion; Mamba; KAN; Deep learning; Infrared image; NETWORK;
D O I
10.1016/j.optlastec.2025.112612
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
The infrared and visible image fusion (IVIF) technique, as an important branch of image processing, has garnered significant attention due to its ability to capture thermal radiation features in low-light conditions and combine them with the rich detail and color information of visible images. However, challenges remain in this field, including the difficulty of balancing information from both modalities using homogenous fusion strategies and the increasing model complexity, which affects computational efficiency. To address these issues, we present an innovative lightweight IVIF framework based on the parallel Mamba-KAN (PMKFuse) model, featuring a multibranch heterogeneous model design. By integrating our proposed multi-channel parallel cross-vision Mamba (PCVM) modules with parallel KAGtention (PKAGN) modules, our approach effectively extracts features at both global and local levels. This not only ensures high performance in image fusion tasks but also significantly reduces the number of model parameters. Additionally, a composite loss function is developed, integrating intensity, gradient, and feature decomposition losses to optimize the training process. The experimental results demonstrate that PMKFuse not only minimizes the number of model parameters but also outperforms the current state-of-the-art (SOTA) methods in both subjective visual evaluation and objective performance metrics for IVIF. These findings highlight the model's effectiveness in improving fusion quality and its potential for wide-ranging practical applications, advancing the field of image processing. The codes are available at https://github. com/sunyichen1994/PMKFuse.
引用
收藏
页数:20
相关论文
共 55 条
[1]  
Bhattacharjee S. S.., 2024, TorchKAN: Simplified KAN Model with Variations
[2]  
Bodner AD, 2024, Arxiv, DOI [arXiv:2406.13155, DOI 10.48550/ARXIV.2406.13155]
[3]  
Bozorgasl Z, 2024, Arxiv, DOI [arXiv:2405.12832, DOI 10.48550/ARXIV.2405.12832]
[4]   Conversational Agents in Therapeutic Interventions for Neurodevelopmental Disorders: A Survey [J].
Catania, Fabio ;
Spitale, Micol ;
Garzotto, Franca .
ACM COMPUTING SURVEYS, 2023, 55 (10)
[5]  
Gu A., 2021, arXiv, DOI [arXiv:2111.00396, DOI 10.48550/ARXIV.2111.00396]
[6]  
Gu A., 2023, arXiv, DOI arXiv:2312.00752
[7]   AUTOREGRESSIVE CONDITIONAL HETEROSKEDASTICITY AND CHANGES IN REGIME [J].
HAMILTON, JD ;
SUSMEL, R .
JOURNAL OF ECONOMETRICS, 1994, 64 (1-2) :307-333
[8]   LLVIP: A Visible-infrared Paired Dataset for Low-light Vision [J].
Jia, Xinyu ;
Zhu, Chuang ;
Li, Minzhen ;
Tang, Wenqi ;
Zhou, Wenli .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, :3489-3497
[9]  
Li CX, 2024, Arxiv, DOI arXiv:2406.02918
[10]   Conti-Fuse: A novel continuous decomposition-based fusion framework for infrared and visible images [J].
Li, Hui ;
Ma, Haolong ;
Cheng, Chunyang ;
Shen, Zhongwei ;
Song, Xiaoning ;
Wu, Xiao-Jun .
INFORMATION FUSION, 2025, 117