Patch Diffusion: Faster and More Data-Efficient Training of Diffusion Models

被引:0
作者
Wang, Zhendong [1 ,2 ]
Jiang, Yifan [1 ]
Zheng, Huangjie [1 ,2 ]
Wang, Peihao [1 ]
He, Pengcheng [2 ]
Wang, Zhangyang [1 ]
Chen, Weizhu [2 ]
Zhou, Mingyuan [1 ]
机构
[1] Univ Texas Austin, Austin, TX 78712 USA
[2] Microsoft Azure AI, Austin, TX 78759 USA
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023) | 2023年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Diffusion models are powerful, but they require a lot of time and data to train. We propose Patch Diffusion, a generic patch-wise training framework, to significantly reduce the training time costs while improving data efficiency, which thus helps democratize diffusion model training to broader users. At the core of our innovations is a new conditional score function at the patch level, where the patch location in the original image is included as additional coordinate channels, while the patch size is randomized and diversified throughout training to encode the cross-region dependency at multiple scales. Sampling with our method is as easy as in the original diffusion model. Through Patch Diffusion, we could achieve >= 2x faster training, while maintaining comparable or better generation quality. Patch Diffusion meanwhile improves the performance of diffusion models trained on relatively small datasets, e.g., as few as 5,000 images to train from scratch. We achieve outstanding FID scores in line with state-of-the-art benchmarks: 1.77 on CelebA-64x64, 1.93 on AFHQv2-Wild-64x64, and 2.72 on ImageNet-256x256. We share our code and pre-trained models at https://github.com/Zhendong- Wang/Patch- Diffusion.
引用
收藏
页数:18
相关论文
共 50 条
[41]   Data-efficient Machine Learning Methods for Electric Motor Surrogate Models [J].
Wang, Bingnan ;
Sakamoto, Yusuke .
2024 IEEE 21ST BIENNIAL CONFERENCE ON ELECTROMAGNETIC FIELD COMPUTATION, CEFC 2024, 2024,
[42]   Efficient computation of the likelihood expansions for diffusion models [J].
Li, Chenxu ;
An, Yu ;
Chen, Dachuan ;
Lin, Qi ;
Si, Nian .
IIE TRANSACTIONS, 2016, 48 (12) :1156-1171
[43]   Data sketching: The approximate approach is often faster and more efficient [J].
Cormode, Graham .
Queue, 2017, 15 (02) :1-28
[44]   When less is more: a more efficient vapour-diffusion protocol [J].
Dunlop, KV ;
Hazes, B .
ACTA CRYSTALLOGRAPHICA SECTION D-BIOLOGICAL CRYSTALLOGRAPHY, 2003, 59 :1797-1800
[45]   A self-supervised deep learning method for data-efficient training in genomics [J].
Guenduez, Hueseyin Anil ;
Binder, Martin ;
To, Xiao-Yin ;
Mreches, Rene ;
Bischl, Bernd ;
McHardy, Alice C. ;
Muench, Philipp C. ;
Rezaei, Mina .
COMMUNICATIONS BIOLOGY, 2023, 6 (01)
[46]   A self-supervised deep learning method for data-efficient training in genomics [J].
Hüseyin Anil Gündüz ;
Martin Binder ;
Xiao-Yin To ;
René Mreches ;
Bernd Bischl ;
Alice C. McHardy ;
Philipp C. Münch ;
Mina Rezaei .
Communications Biology, 6
[47]   A Data-Efficient Framework for Training and Sim-to-Real Transfer of Navigation Policies [J].
Bharadhwaj, Homanga ;
Wang, Zihan ;
Bengio, Yoshua ;
Paull, Liam .
2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, :782-788
[48]   A Data-Efficient Training Model for Signal Integrity Analysis based on Transfer Learning [J].
Zhang, Tingrui ;
Chen, Siyu ;
Wei, Shuwu ;
Chen, Jienan .
2019 IEEE ASIA PACIFIC CONFERENCE ON CIRCUITS AND SYSTEMS (APCCAS 2019), 2019, :182-185
[49]   Effect of Training Epoch Number on Patient Data Memorization in Unconditional Latent Diffusion Models [J].
Dar, Salman U. Hassan ;
Ayx, Isabelle ;
Kapusta, Marie ;
Papavassiliu, Theano ;
Schoenberg, Stefan O. ;
Engelhardt, Sandy .
BILDVERARBEITUNG FUR DIE MEDIZIN 2024, 2024, :88-93
[50]   Unleashing Registration: Diffusion Models for Synthetic Paired 3D Training Data [J].
Falta, Fenja ;
Heyer, Wiebke ;
Grossbroehmer, Christoph ;
Heinrich, Mattias P. .
BIOMEDICAL IMAGE REGISTRATION, WBIR 2024, 2025, 15249 :45-59