Deep adaptive fusion with cross-modality feature transition and modality quaternion learning for medical image fusion

被引:1
|
作者
Srivastava, Somya [1 ]
Bhatia, Shaveta [2 ]
Agrawal, Arun Prakash [3 ]
Jayswal, Anant Kumar [4 ]
Godara, Jyoti [5 ]
Dubey, Gaurav [6 ]
机构
[1] ABES Engn Coll, Dept Comp Sci, Ghaziabad, UP, India
[2] Manav Rachna Int Inst Res & Studies, Faridabad, India
[3] Bennett Univ, Sch Comp Sci Engn & Technol, Greater Noida, India
[4] Amity Univ, Amity Sch Engn & Technol, Noida, UP, India
[5] Shree Guru Gobind Singh Tricentenary Univ, Dept Comp Sci Engn, Gurugram 122505, Haryana, India
[6] KIET Grp Inst, Dept Comp Sci, Ghaziabad, UP, India
关键词
Image fusion; Multimodal imaging; Attention network; Imaging data integration; Deep sparse coding; MODEL;
D O I
10.1007/s12530-024-09648-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In today's rapidly advancing medical landscape, the integration of information from multiple imaging modalities, known as medical fusion, stands at the forefront of diagnostic innovation. This approach combines the strengths of diverse techniques such as magnetic resonance imaging (MRI), computed tomography (CT), positron emission tomography (PET), and single-photon emission computed tomography (SPECT) to offer a more comprehensive view of a patient's condition. Issues such as data heterogeneity, where varied resolutions and contrasts must be harmonized, complicate the seamless integration of imaging data. The complexity of interpreting fused images demands specialized training for clinicians and raises concerns about potential diagnostic errors. This work presents the deep adaptive fusion (Deep-AF) model for image fusion in multimodal biomedical scans includes MRI, CT, PET, and SPECT. This Deep-AF model integrates convolutional neural network (CNN)-based decision maps, deep sparse coding, cross-modality feature transition, and fusion techniques. Three pre-processing steps, including intensity normalization, noise reduction, and spatial registration, are initially applied to enhance alignment and quality in fused images. Non-subsampled contourlet thresholding (NSCTT) is employed to address challenges related to intensity, resolution, and contrast differences among modalities, facilitating multi-scale and multidirectional representation. Despite challenges in spatial alignment, interpretation across modalities, and model generalization, the proposed gradient-weighted class activation mapping with CNN (GradCAM-CNN) enhances interpretability by visualizing crucial regions for CNN predictions. Deep sparse coding fusion (DSCF) overcomes challenges through the adaptive learning of complex features, capturing high-level features while enforcing sparsity. The cross-modality feature transition mechanism (CMFTM) addresses variations in modality characteristics. The attention weighted averaging network (AtWANet) addresses challenges in multimodal feature fusion by dynamically assigning weights based on relevance, providing a flexible approach despite misalignment and scale variations. AtWANet's model training optimizes the fusion process by dynamically assigning attention weights to each modality, ensuring effective integration of varied representations. Simulation results obtains that the proposed Deep-AF model obtains robust fusion results in terms of statistical and accuracy metrics.
引用
收藏
页数:26
相关论文
共 50 条
  • [31] Deep feature learning with attributes for cross-modality person re-identification
    Zhang, Shikun
    Chen, Changhong
    Song, Wanru
    Gan, Zongliang
    JOURNAL OF ELECTRONIC IMAGING, 2020, 29 (03)
  • [32] Cross-modality Attention Method for Medical Image Enhancement
    Hu, Zebin
    Liu, Hao
    Li, Zhendong
    Yu, Zekuan
    PATTERN RECOGNITION AND COMPUTER VISION,, PT III, 2021, 13021 : 411 - 423
  • [33] Fusion of Textural and Visual Information for Medical Image Modality Retrieval Using Deep Learning-Based Feature Engineering
    Iqbal, Saeed
    Qureshi, Adnan N.
    Alhussein, Musaed
    Choudhry, Imran Arshad
    Aurangzeb, Khursheed
    Khan, Tariq M.
    IEEE ACCESS, 2023, 11 : 93238 - 93253
  • [34] Correlative techniques for cross-modality medical image registration
    Richardson, DB
    Bury, EA
    MEDICAL IMAGING 1996: IMAGE PROCESSING, 1996, 2710 : 368 - 375
  • [35] A Survey on Multi - Modality Medical Image Fusion
    Bhavana, V
    Krishnappa, H. K.
    PROCEEDINGS OF THE 2016 IEEE INTERNATIONAL CONFERENCE ON WIRELESS COMMUNICATIONS, SIGNAL PROCESSING AND NETWORKING (WISPNET), 2016, : 1326 - 1329
  • [36] Synergistic Image and Feature Adaptation: Towards Cross-Modality Domain Adaptation for Medical Image Segmentation
    Chen, Cheng
    Dou, Qi
    Chen, Hao
    Qin, Jing
    Heng, Pheng-Ann
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 865 - 872
  • [37] Cross-Modality Multi-Atlas Segmentation via Deep Registration and Label Fusion
    Ding, Wangbin
    Li, Lei
    Zhuang, Xiahai
    Huang, Liqin
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2022, 26 (07) : 3104 - 3115
  • [38] Facial Expression Recognition Through Cross-Modality Attention Fusion
    Ni, Rongrong
    Yang, Biao
    Zhou, Xu
    Cangelosi, Angelo
    Liu, Xiaofeng
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2023, 15 (01) : 175 - 185
  • [39] Cross-modality complementary information fusion for multispectral pedestrian detection
    Yan, Chaoqi
    Zhang, Hong
    Li, Xuliang
    Yang, Yifan
    Yuan, Ding
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (14): : 10361 - 10386
  • [40] Hybrid Fusion with Intra- and Cross-Modality Attention for Image-Recipe Retrieval
    Li, Jiao
    Xu, Xing
    Yu, Wei
    Shen, Fumin
    Cao, Zuo
    Zuo, Kai
    Shen, Heng Tao
    SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, : 244 - 254