Generative Adversarial Networks Based on Penalty of Conditional Entropy Distance

被引:0
作者
Tan H.-W. [1 ,2 ]
Wang G.-D. [1 ]
Zhou L.-Y. [2 ]
Zhang Z.-L. [1 ,3 ]
机构
[1] College of Computer and Information Science, Southwest University, Chongqing
[2] School of Mathematics and Statistics, Guizhou University of Finance and Economics, Guiyang
[3] School of Information Technology, Deakin University, Locked Bag 20000, Geelong, 3220, VIC
来源
Zhang, Zi-Li (zhangzl@swu.edu.cn) | 1600年 / Chinese Academy of Sciences卷 / 32期
基金
中国国家自然科学基金;
关键词
Conditional entropy distance; Generative adversarial networks; Image generation; Network structure; Sample diversity;
D O I
10.13328/j.cnki.jos.006156
中图分类号
学科分类号
摘要
Generating high-quality samples is always one of the main challenges in generative adversarial networks (GANs) field. To this end, in this study, a GANs penalty algorithm is proposed, which leverages a constructed conditional entropy distance to penalize its generator. Under the condition of keeping the entropy invariant, the algorithm makes the generated distribution as close to the target distribution as possible and greatly improves the quality of the generated samples. In addition, to improve the training efficiency of GANs, the network structure of GANs is optimized and the initialization strategy of the two networks is changed. The experimental results on several datasets show that the penalty algorithm significantly improves the quality of generated samples. Especially, on the CIFAR10, STL10, and CelebA datasets, the best FID value is reduced from 16.19, 14.10, 4.65 to 14.02, 12.83, and 3.22, respectively. © Copyright 2021, Institute of Software, the Chinese Academy of Sciences. All rights reserved.
引用
收藏
页码:1116 / 1128
页数:12
相关论文
共 44 条
[31]  
Krizhevsky A., Learning multiple layers of features from tiny images, (2009)
[32]  
Netzer Y, wang TJ, Coates A, Bissacco A, Wu BL, Ng AY., Reading digits in natural images with unsupervised feature learning, Proc. of the 25th Conf. on Neural Information Processing Systems Workshop on Deep Learning and Unsupervised Feature Learning, (2011)
[33]  
Coates A, Lee H, Ng AY., An analysis of single layer networks in unsupervised feature learning, Proc. of the Machine Learning Research (PMLR), 15, pp. 215-223, (2011)
[34]  
Liu ZW, Luo P, Wang XG, Tang XO., Deep learning face attributes in the wild, Proc. of the 2015 IEEE Int'l Conf. on Computer Vision, pp. 3730-3738, (2015)
[35]  
Kramberger T, Potocnik B., LSUN-Stanford car dataset: Enhancing large-scale car image datasets using deep learning for usage in GAN training, Applied Sciences, 10, 14, (2020)
[36]  
Salimans T, Goodfellow I, Zaremba W, Cheung V, Radford A, Chen X., Improved techniques for training GANs, Advances in Neural Information Processing Systems, 29, pp. 2234-2242, (2016)
[37]  
Heusel M, Ramsauer H, Unterthiner T, Nessler B, Hochreiter S., GANs trained by a two time-scale update rule converge to a local nash equilibrium, Advances in Neural Information Processing Systems 30, pp. 6626-6637, (2017)
[38]  
Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z., Rethinking the inception architecture for computer vision, Proc. of the 29th IEEE Conf. on Computer Vision and Pattern Recognition, pp. 2818-2826, (2016)
[39]  
Barratt S, Sharma R., A note on the inception score, Proc. of the Int'l Conf. on Machine Learning Workshop on Theoretical Foundations and Applications of Deep Generative Models, (2018)
[40]  
Vaserstein LN., Markov processes over denumerable products of spaces describing large systems of automata, Problems of Information Transmission, 5, 3, pp. 64-72, (1969)