KD-DLGAN: Data Limited Image Generation via Knowledge Distillation

被引:16
作者
Cui, Kaiwen [1 ]
Yu, Yingchen [1 ]
Zhan, Fangneng [2 ]
Liao, Shengcai [3 ]
Lu, Shijian [1 ]
Xing, Eric [4 ]
机构
[1] Nanyang Technol Univ, Singapore, Singapore
[2] Max Planck Inst Informat, Saarbrucken, Germany
[3] Incept Inst Artificial Intelligence, Abu Dhabi, U Arab Emirates
[4] Mohamed bin Zayed Univ Artificial Intelligence, Abu Dhabi, U Arab Emirates
来源
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR | 2023年
关键词
D O I
10.1109/CVPR52729.2023.00377
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Generative Adversarial Networks (GANs) rely heavily on large-scale training data for training high-quality image generation models. With limited training data, the GAN discriminator often suffers from severe overfitting which directly leads to degraded generation especially in generation diversity. Inspired by the recent advances in knowledge distillation (KD), we propose KD-DLGAN, a knowledge-distillation based generation framework that introduces pre-trained vision-language models for training effective data-limited generation models. KD-DLGAN consists of two innovative designs. The first is aggregated generative KD that mitigates the discriminator overfitting by challenging the discriminator with harder learning tasks and distilling more generalizable knowledge from the pre-trained models. The second is correlated generative KD that improves the generation diversity by distilling and preserving the diverse image-text correlation within the pre-trained models. Extensive experiments over multiple benchmarks show that KD-DLGAN achieves superior image generation with limited training data. In addition, KD-DLGAN complements the state-of-the-art with consistent and substantial performance gains. Note that codes will be released.
引用
收藏
页码:3872 / 3882
页数:11
相关论文
共 54 条
[41]  
Romero AdrianaBallas., 2014, FITNETS HINTS THIN D
[42]  
Salimans T, 2016, ADV NEUR IN, V29
[43]   Learning Hybrid Image Templates (HIT) by Information Projection [J].
Si, Zhangzhang ;
Zhu, Song-Chun .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2012, 34 (07) :1354-1367
[44]  
Tseng Hung-Yu, 2021, REGULARIZING GENERAT
[45]   Similarity-Preserving Knowledge Distillation [J].
Tung, Frederick ;
Mori, Greg .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :1365-1374
[46]   Detecting Overfitting of Deep Generative Networks via Latent Recovery [J].
Webster, Ryan ;
Rabin, Julien ;
Simon, Loic ;
Jurie, Frederic .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :11265-11274
[47]   Cross-Image Relational Knowledge Distillation for Semantic Segmentation [J].
Yang, Chuanguang ;
Zhou, Helong ;
An, Zhulin ;
Jiang, Xue ;
Xu, Yongjun ;
Zhang, Qian .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, :12309-12318
[48]   Towards Counterfactual Image Manipulation via CLIP [J].
Yu, Yingchen ;
Zhan, Fangneng ;
Wu, Rongliang ;
Zhang, Jiahui ;
Lu, Shijian ;
Cui, Miaomiao ;
Xie, Xuansong ;
Hua, Xian-Sheng ;
Miao, Chunyan .
PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, :3637-3645
[49]   Regularizing Class-wise Predictions via Self-knowledge Distillation [J].
Yun, Sukmin ;
Park, Jongjin ;
Lee, Kimin ;
Shin, Jinwoo .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :13873-13882
[50]   Unbalanced Feature Transport for Exemplar-based Image Translation [J].
Zhan, Fangneng ;
Yu, Yingchen ;
Cui, Kaiwen ;
Zhang, Gongjie ;
Lu, Shijian ;
Pan, Jianxiong ;
Zhang, Changgong ;
Ma, Feiying ;
Xie, Xuansong ;
Miao, Chunyan .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :15023-15033