Patch Diffusion: Faster and More Data-Efficient Training of Diffusion Models

被引:0
|
作者
Wang, Zhendong [1 ,2 ]
Jiang, Yifan [1 ]
Zheng, Huangjie [1 ,2 ]
Wang, Peihao [1 ]
He, Pengcheng [2 ]
Wang, Zhangyang [1 ]
Chen, Weizhu [2 ]
Zhou, Mingyuan [1 ]
机构
[1] Univ Texas Austin, Austin, TX 78712 USA
[2] Microsoft Azure AI, Austin, TX 78759 USA
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023) | 2023年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Diffusion models are powerful, but they require a lot of time and data to train. We propose Patch Diffusion, a generic patch-wise training framework, to significantly reduce the training time costs while improving data efficiency, which thus helps democratize diffusion model training to broader users. At the core of our innovations is a new conditional score function at the patch level, where the patch location in the original image is included as additional coordinate channels, while the patch size is randomized and diversified throughout training to encode the cross-region dependency at multiple scales. Sampling with our method is as easy as in the original diffusion model. Through Patch Diffusion, we could achieve >= 2x faster training, while maintaining comparable or better generation quality. Patch Diffusion meanwhile improves the performance of diffusion models trained on relatively small datasets, e.g., as few as 5,000 images to train from scratch. We achieve outstanding FID scores in line with state-of-the-art benchmarks: 1.77 on CelebA-64x64, 1.93 on AFHQv2-Wild-64x64, and 2.72 on ImageNet-256x256. We share our code and pre-trained models at https://github.com/Zhendong- Wang/Patch- Diffusion.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Extracting Training Data from Diffusion Models
    Carlini, Nicholas
    Hayes, Jamie
    Nasr, Milad
    Jagielski, Matthew
    Sehwag, Vikash
    Tramer, Florian
    Balle, Borja
    Ippolito, Daphne
    Wallace, Eric
    PROCEEDINGS OF THE 32ND USENIX SECURITY SYMPOSIUM, 2023, : 5253 - 5270
  • [2] Data-Efficient Augmentation for Training Neural Networks
    Liu, Tian Yu
    Mirzasoleiman, Baharan
    Advances in Neural Information Processing Systems, 2022, 35
  • [3] Data-Efficient Augmentation for Training Neural Networks
    Liu, Tian Yu
    Mirzasoleiman, Baharan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [4] ACT-Diffusion: Efficient Adversarial Consistency Training for One-step Diffusion Models
    Kong, Fei
    Duan, Jinhao
    Sun, Lichao
    Cheng, Hao
    Xu, Renjing
    Shen, Hengtao
    Zhu, Xiaofeng
    Shi, Xiaoshuang
    Xu, Kaidi
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2024, 2024, : 8890 - 8899
  • [5] Synthesizing efficient data with diffusion models for person re-identification pre-training
    Niu, Ke
    Yu, Haiyang
    Qian, Xuelin
    Fu, Teng
    Li, Bin
    Xue, Xiangyang
    MACHINE LEARNING, 2025, 114 (03)
  • [6] Unsupervised Salient Patch Selection for Data-Efficient Reinforcement Learning
    Jiang, Zhaohui
    Weng, Paul
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT IV, 2023, 14172 : 556 - 572
  • [7] Data-Efficient Training Strategies for Neural TTS Systems
    Prajwal, K. R.
    Jawahar, C., V
    CODS-COMAD 2021: PROCEEDINGS OF THE 3RD ACM INDIA JOINT INTERNATIONAL CONFERENCE ON DATA SCIENCE & MANAGEMENT OF DATA (8TH ACM IKDD CODS & 26TH COMAD), 2021, : 223 - 227
  • [8] A Data-Efficient Training Method for Deep Reinforcement Learning
    Feng, Wenhui
    Han, Chongzhao
    Lian, Feng
    Liu, Xia
    ELECTRONICS, 2022, 11 (24)
  • [9] Sobolev Training for Data-efficient Approximate Nonlinear MPC
    Lueken, Lukas
    Brandner, Dean
    Lucia, Sergio
    IFAC PAPERSONLINE, 2023, 56 (02): : 5765 - 5772
  • [10] A hierarchical and data-efficient network based on patch-based representation
    Dazhi Yao
    Yunxue Shao
    Signal, Image and Video Processing, 2023, 17 : 2713 - 2719