AFDFusion: An adaptive frequency decoupling fusion network for multi-modality image

被引:2
|
作者
Wang, Chengchao [1 ]
Zhao, Zhengpeng [1 ]
Yang, Qiuxia [1 ]
Nie, Rencan [1 ]
Cao, Jinde [2 ,3 ]
Pu, Yuanyuan [1 ,4 ]
机构
[1] Yunnan Univ, Coll Informat Sci & Engn, Kunming 650500, Peoples R China
[2] Southeast Univ, Sch Math, Nanjing 210096, Peoples R China
[3] Ahlia Univ, Manama, Bahrain
[4] Univ Key Lab Internet Things Technol & Applicat Yu, Kunming 650500, Yunnan, Peoples R China
基金
中国国家自然科学基金;
关键词
Image fusion; Adaptive frequency decoupling; Contrastive learning; Associative invariant; Intrinsic specific; ARCHITECTURE; PERFORMANCE; FRAMEWORK; NEST;
D O I
10.1016/j.eswa.2024.125694
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The multi-modality image fusion goal is to create a single image that provides a comprehensive scene description and conforms to visual perception by integrating complementary information about the merits of the different modalities, e.g ., salient intensities of infrared images and detail textures of visible images. Although some works explore decoupled representations of multi-modality images, they struggle with complex nonlinear relationships, fine modal decoupling, and noise handling. To cope with this issue, we propose an adaptive frequency decoupling module to perceive the associative invariant and inherent specific among cross- modality by dynamically adjusting the learnable low frequency weight of the kernel. Specifically, we utilize a contrastive learning loss for restricting the solution space of feature decoupling to learn representations of both the invariant and specific in the multi-modality images. The underlying idea is that: in decoupling, low frequency features, which are similar in the representation space, should be pulled closer to each other, signifying the associative invariant, while high frequencies are pushed farther away, also indicating the intrinsic specific. Additionally, a multi-stage training manner is introduced into our framework to achieve decoupling and fusion. Stage I, MixEncoder and MixDecoder with the same architecture but different parameters are trained to perform decoupling and reconstruction supervised by the contrastive self-supervised mechanism. Stage II, two feature fusion modules are added to integrate the invariant and specific features and output the fused image. Extensive experiments demonstrated the proposed method superiority over the state-of-the-art methods in both qualitative and quantitative evaluation on two multi-modal image fusion tasks.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Image retrieval with a multi-modality ontology
    Wang, Huan
    Liu, Song
    Chia, Liang-Tien
    MULTIMEDIA SYSTEMS, 2008, 13 (5-6) : 379 - 390
  • [42] Domain Adaptive Multi-Modality Neural Attention Network for Financial Forecasting
    Zhou, Dawei
    Zheng, Lecheng
    Zhu, Yada
    Li, Jianbo
    He, Jingrui
    WEB CONFERENCE 2020: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2020), 2020, : 2230 - 2240
  • [43] Multi-modality image fusion via generalized Riesz-wavelet transformation
    Jin, Bo
    Jing, Zhongliang
    Pan, Han
    KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS, 2014, 8 (11): : 4118 - 4136
  • [44] Multi-Modality Medical Image Fusion Based on Wavelet Analysis and Quality Evaluation
    Yu Lifeng
    & Zu Donglin Institute of Heavy Ion Physics
    JournalofSystemsEngineeringandElectronics, 2001, (01) : 42 - 48
  • [45] Multi-modality medical image, fusion method based on wavelet packet transform
    Li Wei
    Zhu Xue-feng
    PROCEEDINGS OF 2005 CHINESE CONTROL AND DECISION CONFERENCE, VOLS 1 AND 2, 2005, : 737 - +
  • [46] A multi-modality paradigm for CT and MRI fusion with applications of quantum image processing
    Dogra, Ayush
    Ahuja, Chirag Kamal
    Kumar, Sanjeev
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2022, 34 (20):
  • [47] Multi-Modality Imaging: A Software Fusion and Image-Guided Therapy Perspective
    Birkfellner, Wolfgang
    Figl, Michael
    Furtado, Hugo
    Renner, Andreas
    Hatamikia, Sepideh
    Hummel, Johann
    FRONTIERS IN PHYSICS, 2018, 6
  • [48] A review: Deep learning for medical image segmentation using multi-modality fusion
    Zhou, Tongxue
    Ruan, Su
    Canu, Stephane
    ARRAY, 2019, 3-4
  • [49] Classification of Mineral Foam Flotation Conditions Based on Multi-Modality Image Fusion
    Jiang, Xiaoping
    Zhao, Huilin
    Liu, Junwei
    APPLIED SCIENCES-BASEL, 2023, 13 (06):
  • [50] The application of multi-modality medical image fusion based method to cerebral infarction
    Dai, Yin
    Zhou, Zixia
    Xu, Lu
    EURASIP JOURNAL ON IMAGE AND VIDEO PROCESSING, 2017,