DDFM: Denoising Diffusion Model for Multi-Modality Image Fusion

被引:69
|
作者
Zhao, Zixiang [1 ,2 ]
Bai, Haowen [1 ]
Zhu, Yuanzhi [2 ]
Zhang, Jiangshe [1 ]
Xu, Shuang
Zhang, Yulun [2 ]
Zhang, Kai [2 ]
Meng, Deyu
Timofte, Radu [2 ]
Van Gool, Luc [2 ]
机构
[1] Xi An Jiao Tong Univ, Xian, Peoples R China
[2] Swiss Fed Inst Technol, Comp Vis Lab, Zurich, Switzerland
基金
中国国家自然科学基金;
关键词
NETWORK; NEST;
D O I
10.1109/ICCV51070.2023.00742
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-modality image fusion aims to combine different modalities to produce fused images that retain the complementary features of each modality, such as functional highlights and texture details. To leverage strong generative priors and address challenges such as unstable training and lack of interpretability for GAN-based generative methods, we propose a novel fusion algorithm based on the denoising diffusion probabilistic model (DDPM). The fusion task is formulated as a conditional generation problem under the DDPM sampling framework, which is further divided into an unconditional generation subproblem and a maximum likelihood subproblem. The latter is modeled in a hierarchical Bayesian manner with latent variables and inferred by the expectation-maximization (EM) algorithm. By integrating the inference solution into the diffusion sampling iteration, our method can generate high-quality fused images with natural image generative priors and cross-modality information from source images. Note that all we required is an unconditional pre-trained generative model, and no fine-tuning is needed. Our extensive experiments indicate that our approach yields promising fusion results in infrared-visible image fusion and medical image fusion. The code is available at https://github. com/Zhaozixiang1228/MMIF-DDFM.
引用
收藏
页码:8048 / 8059
页数:12
相关论文
共 50 条
  • [1] Diff-IF: Multi-modality image fusion via diffusion model with fusion knowledge prior
    Yi, Xunpeng
    Tang, Linfeng
    Zhang, Hao
    Xu, Han
    Ma, Jiayi
    INFORMATION FUSION, 2024, 110
  • [2] Diffusion-driven multi-modality medical image fusion
    Qu, Jiantao
    Huang, Dongjin
    Shi, Yongsheng
    Liu, Jinhua
    Tang, Wen
    MEDICAL & BIOLOGICAL ENGINEERING & COMPUTING, 2025,
  • [3] Equivariant Multi-Modality Image Fusion
    Zhao, Zixiang
    Hai, Haowen
    Zhang, Jiangshe
    Zhang, Yulun
    Zhane, Kai
    Xu, Shuang
    Chen, Dongdong
    Timofte, Radu
    Van Gool, Luc
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 25912 - 25921
  • [4] A Novel Multi-Modality Image Simultaneous Denoising and Fusion Method Based on Sparse Representation
    Qi, Guanqiu
    Hu, Gang
    Mazur, Neal
    Liang, Huahua
    Haner, Matthew
    COMPUTERS, 2021, 10 (10)
  • [5] Multi-modality image fusion for image-guided neurosurgery
    Haller, JW
    Ryken, T
    Madsen, M
    Edwards, A
    Bolinger, L
    Vannier, MW
    CARS '99: COMPUTER ASSISTED RADIOLOGY AND SURGERY, 1999, 1191 : 681 - 685
  • [6] Fast saliency-aware multi-modality image fusion
    Han, Jungong
    Pauwels, Eric J.
    de Zeeuw, Paul
    NEUROCOMPUTING, 2013, 111 : 70 - 80
  • [7] Lymphatic flow mapping utilizing multi-modality image fusion
    Vicic, M
    Thorstad, W
    Low, D
    Deasy, J
    MEDICAL PHYSICS, 2004, 31 (06) : 1900 - 1900
  • [8] Multi-Modality Image Fusion Using the Nonsubsampled Contourlet Transform
    Liu, Cuiyin
    Chen, Shu-qing
    Fu, Qiao
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2013, E96D (10): : 2215 - 2223
  • [9] Multi-modality gaze-contingent displays for image fusion
    Nikolov, SG
    Bull, DR
    Canagarajah, CN
    Jones, MG
    Gilchrist, ID
    PROCEEDINGS OF THE FIFTH INTERNATIONAL CONFERENCE ON INFORMATION FUSION, VOL II, 2002, : 1213 - 1220
  • [10] The application of wavelet transform to multi-modality medical image fusion
    Wang, Anna
    Sun, Haijing
    Guan, Yueyang
    PROCEEDINGS OF THE 2006 IEEE INTERNATIONAL CONFERENCE ON NETWORKING, SENSING AND CONTROL, 2006, : 270 - 274