DDFM: Denoising Diffusion Model for Multi-Modality Image Fusion

被引:69
|
作者
Zhao, Zixiang [1 ,2 ]
Bai, Haowen [1 ]
Zhu, Yuanzhi [2 ]
Zhang, Jiangshe [1 ]
Xu, Shuang
Zhang, Yulun [2 ]
Zhang, Kai [2 ]
Meng, Deyu
Timofte, Radu [2 ]
Van Gool, Luc [2 ]
机构
[1] Xi An Jiao Tong Univ, Xian, Peoples R China
[2] Swiss Fed Inst Technol, Comp Vis Lab, Zurich, Switzerland
基金
中国国家自然科学基金;
关键词
NETWORK; NEST;
D O I
10.1109/ICCV51070.2023.00742
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-modality image fusion aims to combine different modalities to produce fused images that retain the complementary features of each modality, such as functional highlights and texture details. To leverage strong generative priors and address challenges such as unstable training and lack of interpretability for GAN-based generative methods, we propose a novel fusion algorithm based on the denoising diffusion probabilistic model (DDPM). The fusion task is formulated as a conditional generation problem under the DDPM sampling framework, which is further divided into an unconditional generation subproblem and a maximum likelihood subproblem. The latter is modeled in a hierarchical Bayesian manner with latent variables and inferred by the expectation-maximization (EM) algorithm. By integrating the inference solution into the diffusion sampling iteration, our method can generate high-quality fused images with natural image generative priors and cross-modality information from source images. Note that all we required is an unconditional pre-trained generative model, and no fine-tuning is needed. Our extensive experiments indicate that our approach yields promising fusion results in infrared-visible image fusion and medical image fusion. The code is available at https://github. com/Zhaozixiang1228/MMIF-DDFM.
引用
收藏
页码:8048 / 8059
页数:12
相关论文
共 50 条
  • [21] A Bilevel Integrated Model With Data-Driven Layer Ensemble for Multi-Modality Image Fusion
    Liu, Risheng
    Liu, Jinyuan
    Jiang, Zhiying
    Fan, Xin
    Luo, Zhongxuan
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 1261 - 1274
  • [22] MMGInpainting: Multi-Modality Guided Image Inpainting Based on Diffusion Models
    Zhang, Cong
    Yang, Wenxia
    Li, Xin
    Han, Huan
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 8811 - 8823
  • [23] DiffMoCa: Diffusion Model Based Multi-modality Cut and Paste
    Zhang, Junjie
    Wu, Shaojin
    Gao, Junbin
    Yu, Fusheng
    Xu, Hao
    Zeng, Zhigang
    ADVANCES IN NEURAL NETWORKS-ISNN 2024, 2024, 14827 : 153 - 162
  • [24] A novel multi-modality image fusion method based on image decomposition and sparse representation
    Zhu, Zhiqin
    Yin, Hongpeng
    Chai, Yi
    Li, Yanxia
    Qi, Guanqiu
    INFORMATION SCIENCES, 2018, 432 : 516 - 529
  • [25] Multi-modality Paraphrase Generation Model Integrating Image Information
    Ma C.
    Wan Z.
    Zhang Y.
    Xu J.
    Chen Y.
    Beijing Daxue Xuebao (Ziran Kexue Ban)/Acta Scientiarum Naturalium Universitatis Pekinensis, 2022, 58 (01): : 45 - 53
  • [26] Image retrieval with a multi-modality ontology
    Huan Wang
    Song Liu
    Liang-Tien Chia
    Multimedia Systems, 2008, 13 : 379 - 390
  • [27] Image retrieval with a multi-modality ontology
    Wang, Huan
    Liu, Song
    Chia, Liang-Tien
    MULTIMEDIA SYSTEMS, 2008, 13 (5-6) : 379 - 390
  • [28] Multi-modality image fusion via generalized Riesz-wavelet transformation
    Jin, Bo
    Jing, Zhongliang
    Pan, Han
    KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS, 2014, 8 (11): : 4118 - 4136
  • [29] Multi-Modality Medical Image Fusion Based on Wavelet Analysis and Quality Evaluation
    Yu Lifeng
    & Zu Donglin Institute of Heavy Ion Physics
    JournalofSystemsEngineeringandElectronics, 2001, (01) : 42 - 48
  • [30] Multi-modality medical image, fusion method based on wavelet packet transform
    Li Wei
    Zhu Xue-feng
    PROCEEDINGS OF 2005 CHINESE CONTROL AND DECISION CONFERENCE, VOLS 1 AND 2, 2005, : 737 - +