SIE: infrared and visible image fusion based on scene information embedding

被引:0
作者
Geng Y. [1 ]
Diao W. [1 ,2 ]
Zhao Y. [3 ]
机构
[1] Electrical and Information Engineering, Changchun Institute of Technology, Kuanping Road, Jilin, Changchun
[2] Information and Control Engineering, Jilin Institute of Chemical Technology, Chengde Street, Jilin, Jilin
[3] College of Communication Engineering, Jilin University, Qianjin Street, Jilin, Changchun
关键词
Convolutional Neural Network (CNN); Infrared and visible image fusion; Multi-scale fusion; Scene information;
D O I
10.1007/s11042-024-19105-y
中图分类号
学科分类号
摘要
In this article, we proposed an infrared and visible image fusion method based on scene information embedding, which is to obtain an fused image with high-quality target information such as pedestrian crossing, vehicles, signs and rich background texture information. First, scene information is described, so as to effectively represent highlight salient target for infrared image and texture detail for visible image. Second, we embed this scene information into multi-scale fusion framework to fuse these information and reconstruct the desired fused image. It is worth noting that scene information is constructed by pre-trained convolutional neural network, where not only information extracted by convolutional neural network is fully utilized, but also training complexity is not increased. In datasets LLVIP, MSRS, M3FD,TNO, RoadScene and elctricity1, the superiority of our proposed algorithm SIE over state-of-the-arts is demonstrated by subjective evaluation and objective evaluation including information entropy (EN), standard deviation (SD), spatial frequency (SF), average gradient (AG), sum of the correlations of differences (SCD) and mutual Information (MI) metrics. © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2024.
引用
收藏
页码:1463 / 1488
页数:25
相关论文
共 59 条
  • [41] Zeng K., Wang Y., Mao J., Et al., Deep stereo matching with hysteresis attention and supervised cost volume construction, IEEE Trans Image Process, 31, pp. 812-822, (2021)
  • [42] Tete X., Mannat S., Eric M., Et al., Early Convolutions Help Transformers See Better, Conf Neural Inf Process Sys, 2021, pp. 30392-30400, (2021)
  • [43] Jia X., Zhu C., Li M., Zhou W., T, LLVIP: A Visible-Infrared Paired Dataset for Low-Light Vision. 2021 IEEE/CVF Int Conf. Comp Vis Workshops (ICCVW), Montreal, BC, Canada, 2021, pp. 3489-3497, (2021)
  • [44] Alexander T., Maarten H., Tno Image Fusion Dataset, (2014)
  • [45] Liu J., Fan X., Huang Z., Et al., Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection, Computer Vis Pattern Reco, pp. 5792-5801, (2022)
  • [46] Tang L., Yuan J., Ma J., Et al., Piafusion: A progressive infrared and visible image fusion network based on illumination aware, Inf Fusion, 83-84, pp. 79-92, (2022)
  • [47] Xu H., Ma J., Le Z., Et al., Fusiondn: a unified densely connected network for image fusion, AAAI Conf Artif Intel, 34, pp. 12484-12491, (2020)
  • [48] Bavirisetti D., Dhuli R., Fusion of infrared and visible sensor images based on anisotropic diffusion and karhunen-loeve transform, IEEE Sensors J, 16, 1, pp. 203-209, (2016)
  • [49] Ma J., Zhou Z., Wang B., Et al., Infrared and visible image fusion based on visual saliency map and weighted least square optimization, Infrared Phy Technol, 82, pp. 8-17, (2017)
  • [50] Naidu V., Image fusion technique using multi-resolution singular value decomposition, Defence Sci J, 61, 5, pp. 479-484, (2011)