Improved Multi-modal Image Fusion with Attention and Dense Networks: Visual and Quantitative Evaluation

被引:0
|
作者
Banerjee, Ankan [1 ]
Patra, Dipti [1 ]
Roy, Pradipta [2 ]
机构
[1] Natl Inst Technol, Rourkela, India
[2] DRDO, Integrated Test Range, Candipur, India
来源
COMPUTER VISION AND IMAGE PROCESSING, CVIP 2023, PT III | 2024年 / 2011卷
关键词
image fusion; attention; human perception; Convolutional Block Attention Module;
D O I
10.1007/978-3-031-58535-7_20
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This article introduces a novel multi-modal image fusion approach based on Convolutional Block Attention Module and dense networks to enhance human perceptual quality and information content in the fused images. The proposed model preserves the edges of the infrared images and enhances the contrast of the visible image as a pre-processing part. Consequently, the use of Convolutional Block Attention Module has resulted in the extraction of more refined features from the source images. The visual results demonstrate that the fused images produced by the proposed method are visually superior to those generated by most standard fusion techniques. To substantiate the findings, quantitative analysis is conducted using various metrics. The proposed method exhibits the best Naturalness Image Quality Evaluator and Chen-Varshney metric values, which are human perception-based parameters. Moreover, the fused images exhibit the highest Standard Deviation value, signifying enhanced contrast. These results justify the proposed multi-modal image fusion technique outperforms standard methods both qualitatively and quantitatively, resulting in superior fused images with improved human perception quality.
引用
收藏
页码:237 / 248
页数:12
相关论文
共 50 条
  • [1] The multi-modal fusion in visual question answering: a review of attention mechanisms
    Lu, Siyu
    Liu, Mingzhe
    Yin, Lirong
    Yin, Zhengtong
    Liu, Xuan
    Zheng, Wenfeng
    PEERJ COMPUTER SCIENCE, 2023, 9
  • [2] Multi-modal Image Fusion with KNN Matting
    Zhang, Xia
    Lin, Hui
    Kang, Xudong
    Li, Shutao
    PATTERN RECOGNITION (CCPR 2014), PT II, 2014, 484 : 89 - 96
  • [3] An overview of multi-modal medical image fusion
    Du, Jiao
    Li, Weisheng
    Lu, Ke
    Xiao, Bin
    NEUROCOMPUTING, 2016, 215 : 3 - 20
  • [4] Guided Image Deblurring by Deep Multi-Modal Image Fusion
    Liu, Yuqi
    Sheng, Zehua
    Shen, Hui-Liang
    IEEE ACCESS, 2022, 10 : 130708 - 130718
  • [5] Attention-based multi-modal fusion sarcasm detection
    Liu, Jing
    Tian, Shengwei
    Yu, Long
    Long, Jun
    Zhou, Tiejun
    Wang, Bo
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 44 (02) : 2097 - 2108
  • [6] Multi-Stage Fusion and Multi-Source Attention Network for Multi-Modal Remote Sensing Image Segmentation
    Zhao, Jiaqi
    Zhou, Yong
    Shi, Boyu
    Yang, Jingsong
    Zhang, Di
    Yao, Rui
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2021, 12 (06)
  • [7] Multi-modal brain image fusion using multi feature guided fusion network
    Shibu, Tom Michael
    Madan, Niranjan
    Paramanandham, Nirmala
    Kumar, Aakash
    Santosh, Ashwin
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2025, 100
  • [8] HATF: Multi-Modal Feature Learning for Infrared and Visible Image Fusion via Hybrid Attention Transformer
    Liu, Xiangzeng
    Wang, Ziyao
    Gao, Haojie
    Li, Xiang
    Wang, Lei
    Miao, Qiguang
    REMOTE SENSING, 2024, 16 (05)
  • [9] MURF: Mutually Reinforcing Multi-Modal Image Registration and Fusion
    Xu, Han
    Yuan, Jiteng
    Ma, Jiayi
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (10) : 12148 - 12166
  • [10] SWT and PCA image fusion methods for multi-modal imagery
    Bashir, Rabia
    Junejo, Riaz
    Qadri, Nadia N.
    Fleury, Martin
    Qadri, Muhammad Yasir
    MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (02) : 1235 - 1263