DDFM: Denoising Diffusion Model for Multi-Modality Image Fusion

被引:72
作者
Zhao, Zixiang [1 ,2 ]
Bai, Haowen [1 ]
Zhu, Yuanzhi [2 ]
Zhang, Jiangshe [1 ]
Xu, Shuang
Zhang, Yulun [2 ]
Zhang, Kai [2 ]
Meng, Deyu
Timofte, Radu [2 ]
Van Gool, Luc [2 ]
机构
[1] Xi An Jiao Tong Univ, Xian, Peoples R China
[2] Swiss Fed Inst Technol, Comp Vis Lab, Zurich, Switzerland
来源
2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023) | 2023年
基金
中国国家自然科学基金;
关键词
NETWORK; NEST;
D O I
10.1109/ICCV51070.2023.00742
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-modality image fusion aims to combine different modalities to produce fused images that retain the complementary features of each modality, such as functional highlights and texture details. To leverage strong generative priors and address challenges such as unstable training and lack of interpretability for GAN-based generative methods, we propose a novel fusion algorithm based on the denoising diffusion probabilistic model (DDPM). The fusion task is formulated as a conditional generation problem under the DDPM sampling framework, which is further divided into an unconditional generation subproblem and a maximum likelihood subproblem. The latter is modeled in a hierarchical Bayesian manner with latent variables and inferred by the expectation-maximization (EM) algorithm. By integrating the inference solution into the diffusion sampling iteration, our method can generate high-quality fused images with natural image generative priors and cross-modality information from source images. Note that all we required is an unconditional pre-trained generative model, and no fine-tuning is needed. Our extensive experiments indicate that our approach yields promising fusion results in infrared-visible image fusion and medical image fusion. The code is available at https://github. com/Zhaozixiang1228/MMIF-DDFM.
引用
收藏
页码:8048 / 8059
页数:12
相关论文
共 76 条
  • [31] DDcGAN: A Dual-Discriminator Conditional Generative Adversarial Network for Multi-Resolution Image Fusion
    Ma, Jiayi
    Xu, Han
    Jiang, Junjun
    Mei, Xiaoguang
    Zhang, Xiao-Ping
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 4980 - 4995
  • [32] Infrared and visible image fusion via detail preserving adversarial learning
    Ma, Jiayi
    Liang, Pengwei
    Yu, Wei
    Chen, Chen
    Guo, Xiaojie
    Wu, Jia
    Jiang, Junjun
    [J]. INFORMATION FUSION, 2020, 54 : 85 - 98
  • [33] FusionGAN: A generative adversarial network for infrared and visible image fusion
    Ma, Jiayi
    Yu, Wei
    Liang, Pengwei
    Li, Chang
    Jiang, Junjun
    [J]. INFORMATION FUSION, 2019, 48 : 11 - 26
  • [34] Infrared and visible image fusion methods and applications: A survey
    Ma, Jiayi
    Ma, Yong
    Li, Chang
    [J]. INFORMATION FUSION, 2019, 45 : 153 - 178
  • [35] Infrared and visible image fusion via gradient transfer and total variation minimization
    Ma, Jiayi
    Chen, Chen
    Li, Chang
    Huang, Jun
    [J]. INFORMATION FUSION, 2016, 31 : 100 - 109
  • [36] Least Squares Generative Adversarial Networks
    Mao, Xudong
    Li, Qing
    Xie, Haoran
    Lau, Raymond Y. K.
    Wang, Zhen
    Smolley, Stephen Paul
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 2813 - 2821
  • [37] A survey on region based image fusion methods
    Meher, Bikash
    Agrawal, Sanjay
    Panda, Rutuparna
    Abraham, Ajith
    [J]. INFORMATION FUSION, 2019, 48 : 119 - 132
  • [38] Mirza M, 2014, Arxiv, DOI [arXiv:1411.1784, 10.48550/arXiv.1411.1784]
  • [39] Nichol A, 2021, PR MACH LEARN RES, V139
  • [40] Diverse Sample Generation: Pushing the Limit of Generative Data-Free Quantization
    Qin, Haotong
    Ding, Yifu
    Zhang, Xiangguo
    Wang, Jiakai
    Liu, Xianglong
    Lu, Jiwen
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (10) : 11689 - 11706