DDFM: Denoising Diffusion Model for Multi-Modality Image Fusion

被引:69
|
作者
Zhao, Zixiang [1 ,2 ]
Bai, Haowen [1 ]
Zhu, Yuanzhi [2 ]
Zhang, Jiangshe [1 ]
Xu, Shuang
Zhang, Yulun [2 ]
Zhang, Kai [2 ]
Meng, Deyu
Timofte, Radu [2 ]
Van Gool, Luc [2 ]
机构
[1] Xi An Jiao Tong Univ, Xian, Peoples R China
[2] Swiss Fed Inst Technol, Comp Vis Lab, Zurich, Switzerland
基金
中国国家自然科学基金;
关键词
NETWORK; NEST;
D O I
10.1109/ICCV51070.2023.00742
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-modality image fusion aims to combine different modalities to produce fused images that retain the complementary features of each modality, such as functional highlights and texture details. To leverage strong generative priors and address challenges such as unstable training and lack of interpretability for GAN-based generative methods, we propose a novel fusion algorithm based on the denoising diffusion probabilistic model (DDPM). The fusion task is formulated as a conditional generation problem under the DDPM sampling framework, which is further divided into an unconditional generation subproblem and a maximum likelihood subproblem. The latter is modeled in a hierarchical Bayesian manner with latent variables and inferred by the expectation-maximization (EM) algorithm. By integrating the inference solution into the diffusion sampling iteration, our method can generate high-quality fused images with natural image generative priors and cross-modality information from source images. Note that all we required is an unconditional pre-trained generative model, and no fine-tuning is needed. Our extensive experiments indicate that our approach yields promising fusion results in infrared-visible image fusion and medical image fusion. The code is available at https://github. com/Zhaozixiang1228/MMIF-DDFM.
引用
收藏
页码:8048 / 8059
页数:12
相关论文
共 50 条
  • [31] A multi-modality paradigm for CT and MRI fusion with applications of quantum image processing
    Dogra, Ayush
    Ahuja, Chirag Kamal
    Kumar, Sanjeev
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2022, 34 (20):
  • [32] ReCoNet: Recurrent Correction Network for Fast and Efficient Multi-modality Image Fusion
    Huang, Zhanbo
    Liu, Jinyuan
    Fan, Xin
    Liu, Risheng
    Zhong, Wei
    Luo, Zhongxuan
    COMPUTER VISION - ECCV 2022, PT XVIII, 2022, 13678 : 539 - 555
  • [33] Multi-Modality Imaging: A Software Fusion and Image-Guided Therapy Perspective
    Birkfellner, Wolfgang
    Figl, Michael
    Furtado, Hugo
    Renner, Andreas
    Hatamikia, Sepideh
    Hummel, Johann
    FRONTIERS IN PHYSICS, 2018, 6
  • [34] A review: Deep learning for medical image segmentation using multi-modality fusion
    Zhou, Tongxue
    Ruan, Su
    Canu, Stephane
    ARRAY, 2019, 3-4
  • [35] Classification of Mineral Foam Flotation Conditions Based on Multi-Modality Image Fusion
    Jiang, Xiaoping
    Zhao, Huilin
    Liu, Junwei
    APPLIED SCIENCES-BASEL, 2023, 13 (06):
  • [36] The application of multi-modality medical image fusion based method to cerebral infarction
    Dai, Yin
    Zhou, Zixia
    Xu, Lu
    EURASIP JOURNAL ON IMAGE AND VIDEO PROCESSING, 2017,
  • [37] Diagnosis and treatment of navigation technology based on the multi-modality image fusion for angioneoplasm
    Gao, Baofeng
    Guo, Shuxiang
    Hu, Kangqi
    2014 IEEE INTERNATIONAL CONFERENCE ON MECHATRONICS AND AUTOMATION (IEEE ICMA 2014), 2014, : 565 - 570
  • [38] The application of multi-modality medical image fusion based method to cerebral infarction
    Yin Dai
    Zixia Zhou
    Lu Xu
    EURASIP Journal on Image and Video Processing, 2017
  • [39] Multi-modality Fusion Network for Action Recognition
    Huang, Kai
    Qin, Zheng
    Xu, Kaiping
    Ye, Shuxiong
    Wang, Guolong
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING - PCM 2017, PT II, 2018, 10736 : 139 - 149
  • [40] Fusion of multi-modality volumetric medical imagery
    Aguilar, M
    New, JR
    PROCEEDINGS OF THE FIFTH INTERNATIONAL CONFERENCE ON INFORMATION FUSION, VOL II, 2002, : 1206 - 1212