On Data Augmentation for GAN Training

被引:230
作者
Tran, Ngoc-Trung [1 ]
Tran, Viet-Hung [1 ]
Nguyen, Ngoc-Bao [1 ]
Nguyen, Trung-Kien [1 ]
Cheung, Ngai-Man [1 ]
机构
[1] Singapore Univ Technol & Design SUTD, Informat Syst Technol & Design ISTD, Singapore 487372, Singapore
基金
新加坡国家研究基金会;
关键词
Gallium nitride; Generative adversarial networks; Generators; Training; Task analysis; Standards; Data models; GAN; data augmentation; limited data; conditional GAN; self-supervised GAN; CycleGAN;
D O I
10.1109/TIP.2021.3049346
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent successes in Generative Adversarial Networks (GAN) have affirmed the importance of using more data in GAN training. Yet it is expensive to collect data in many domains such as medical applications. Data Augmentation (DA) has been applied in these applications. In this work, we first argue that the classical DA approach could mislead the generator to learn the distribution of the augmented data, which could be different from that of the original data. We then propose a principled framework, termed Data Augmentation Optimized for GAN (DAG), to enable the use of augmented data in GAN training to improve the learning of the original distribution. We provide theoretical analysis to show that using our proposed DAG aligns with the original GAN in minimizing the Jensen-Shannon (JS) divergence between the original distribution and model distribution. Importantly, the proposed DAG effectively leverages the augmented data to improve the learning of discriminator and generator. We conduct experiments to apply DAG to different GAN models: unconditional GAN, conditional GAN, self-supervised GAN and CycleGAN using datasets of natural images and medical images. The results show that DAG achieves consistent and considerable improvements across these models. Furthermore, when DAG is used in some GAN models, the system establishes state-of-the-art Frechet Inception Distance (FID) scores. Our code is available (https://github.com/tntrung/dag-gans).
引用
收藏
页码:1882 / 1897
页数:16
相关论文
共 67 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]  
[Anonymous], 2019, PR MACH LEARN RES
[3]  
[Anonymous], 2018, ARXIV180306107
[4]  
Arjovsky M, 2017, PR MACH LEARN RES, V70
[5]  
Brock A., 2018, 7 INT C LEARN REPR I, P1, DOI DOI 10.48550/ARXIV.1809.11096
[6]   Self-Supervised GANs via Auxiliary Rotation Loss [J].
Chen, Ting ;
Zhai, Xiaohua ;
Ritter, Marvin ;
Lucic, Mario ;
Houlsby, Neil .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :12146-12155
[7]  
Chen X, 2016, 30 C NEURAL INFORM P, V29
[8]  
Cheung, 2020, ARXIV200704589
[9]  
Dao Tri, 2019, P MACHINE LEARNING R, V97
[10]  
Donahue J., 2016, ARXIV PREPRINT ARXIV