Patch Diffusion: Faster and More Data-Efficient Training of Diffusion Models

被引:0
作者
Wang, Zhendong [1 ,2 ]
Jiang, Yifan [1 ]
Zheng, Huangjie [1 ,2 ]
Wang, Peihao [1 ]
He, Pengcheng [2 ]
Wang, Zhangyang [1 ]
Chen, Weizhu [2 ]
Zhou, Mingyuan [1 ]
机构
[1] Univ Texas Austin, Austin, TX 78712 USA
[2] Microsoft Azure AI, Austin, TX 78759 USA
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023) | 2023年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Diffusion models are powerful, but they require a lot of time and data to train. We propose Patch Diffusion, a generic patch-wise training framework, to significantly reduce the training time costs while improving data efficiency, which thus helps democratize diffusion model training to broader users. At the core of our innovations is a new conditional score function at the patch level, where the patch location in the original image is included as additional coordinate channels, while the patch size is randomized and diversified throughout training to encode the cross-region dependency at multiple scales. Sampling with our method is as easy as in the original diffusion model. Through Patch Diffusion, we could achieve >= 2x faster training, while maintaining comparable or better generation quality. Patch Diffusion meanwhile improves the performance of diffusion models trained on relatively small datasets, e.g., as few as 5,000 images to train from scratch. We achieve outstanding FID scores in line with state-of-the-art benchmarks: 1.77 on CelebA-64x64, 1.93 on AFHQv2-Wild-64x64, and 2.72 on ImageNet-256x256. We share our code and pre-trained models at https://github.com/Zhendong- Wang/Patch- Diffusion.
引用
收藏
页数:18
相关论文
共 50 条
  • [21] Faster Is More Different: Mean-Field Dynamics of Innovation Diffusion
    Baek, Seung Ki
    Durang, Xavier
    Kim, Mina
    [J]. PLOS ONE, 2013, 8 (07):
  • [22] Faster, More Accurate Quantification of Diffusion and Correlated Motions in Lipid Bilayers
    Urner, Tara M.
    Claflin, Gwendolyn A.
    Lerner, Michael G.
    Kyvelou-Kokkaliaris, Rodoula
    [J]. BIOPHYSICAL JOURNAL, 2016, 110 (03) : 568A - 568A
  • [23] TwinLab: a framework for data-efficient training of non-intrusive reduced-order models for digital twins
    Kannapinn, Maximilian
    Schaefer, Michael
    Weeger, Oliver
    [J]. ENGINEERING COMPUTATIONS, 2024,
  • [24] Post Training Quantization Strategies for Diffusion Models
    Vora, Jayneel
    [J]. PUBLICATION OF THE 26TH ACM INTERNATIONAL CONFERENCE ON MOBILE HUMAN-COMPUTER INTERACTION, MOBILEHCI 2024 ADJUNCT PROCEEDINGS, 2024,
  • [25] Post-training Quantization on Diffusion Models
    Shang, Yuzhang
    Yuan, Zhihang
    Xie, Bin
    Wu, Bingzhe
    Yan, Yan
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 1972 - 1981
  • [26] Analyzing and Improving the Training Dynamics of Diffusion Models
    Karras, Tero
    Aittala, Miika
    Lehtinen, Jaakko
    Hellsten, Janne
    Aila, Timo
    Laine, Samuli
    [J]. 2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 24174 - 24184
  • [27] QCore: Data-Efficient, On-Device Continual Calibration for Quantized Models
    Campos, David
    Yang, Bin
    Kieu, Tung
    Zhang, Miao
    Guo, Chenjuan
    Jensen, Christian S.
    [J]. PROCEEDINGS OF THE VLDB ENDOWMENT, 2024, 17 (11): : 2708 - 2721
  • [28] Less is more: Data-efficient complex question answering over knowledge bases
    Hua, Yuncheng
    Li, Yuan-Fang
    Qi, Guilin
    Wu, Wei
    Zhang, Jingyao
    Qi, Daiqing
    [J]. JOURNAL OF WEB SEMANTICS, 2020, 65
  • [29] Data-efficient Machine Learning Methods for Electric Motor Surrogate Models
    Wang, Bingnan
    Sakamoto, Yusuke
    [J]. 2024 IEEE 21ST BIENNIAL CONFERENCE ON ELECTROMAGNETIC FIELD COMPUTATION, CEFC 2024, 2024,
  • [30] Synthetic data generation by diffusion models
    Jun Zhu
    [J]. National Science Review, 2024, 11 (08) : 19 - 21