PHom-GeM: Persistent Homology for Generative Models

被引:1
作者
Charlier, Jeremy [1 ]
State, Radu [1 ]
Hilger, Jean [2 ]
机构
[1] Univ Luxembourg, Luxembourg, Luxembourg
[2] BCEE, Luxembourg, Luxembourg
来源
2019 6TH SWISS CONFERENCE ON DATA SCIENCE (SDS) | 2019年
关键词
Neural Networks; Optimal Transport; Algebraic Topology;
D O I
10.1109/SDS.2019.000-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Generative neural network models, including Generative Adversarial Network (GAN) and Auto-Encoders (AE), are among the most popular neural network models to generate adversarial data. The GAN model is composed of a generator that produces synthetic data and of a discriminator that discriminates between the generator's output and the true data. AE consist of an encoder which maps the model distribution to a latent manifold and of a decoder which maps the latent manifold to a reconstructed distribution. However, generative models are known to provoke chaotically scattered reconstructed distribution during their training, and consequently, incomplete generated adversarial distributions. Current distance measures fail to address this problem because they are not able to acknowledge the shape of the data manifold, i.e. its topological features, and the scale at which the manifold should be analyzed. We propose Persistent Homology for Generative Models, PHom-GeM, a new methodology to assess and measure the distribution of a generative model. PHom-GeM minimizes an objective function between the true and the reconstructed distributions and uses persistent homology, the study of the topological features of a space at different spatial resolutions, to compare the nature of the true and the generated distributions. Our experiments underline the potential of persistent homology for Wasserstein GAN in comparison to Wasserstein AE and Variational AE. The experiments are conducted on a real-world data set particularly challenging for traditional distance measures and generative neural network models. PHom-GeM is the first methodology to propose a topological distance measure, the bottleneck distance, for generative models used to compare adversarial samples in the context of credit card transactions.
引用
收藏
页码:87 / 92
页数:6
相关论文
共 15 条
[1]  
[Anonymous], 2012, NEUR NETW MACH LEARN
[2]  
[Anonymous], 2003, TOPICS OPTIMAL TRANS
[3]  
[Anonymous], 2015, MATH SURVEYS MONOGRA
[4]  
[Anonymous], 2016, Deep learning. vol
[5]  
Arjovsky M, 2017, PR MACH LEARN RES, V70
[6]  
Bengio Y., 2013, Advances in neural information processing systems, V26, P899
[7]  
Bousquet Olivier, 2017, ARXIV170507642
[8]  
Chazal F., 2017, An introduction to topological data analysis: fundamental and practical aspects for data scientists
[9]   Topological persistence and simplification [J].
Edelsbrunner, H ;
Letscher, D ;
Zomorodian, A .
DISCRETE & COMPUTATIONAL GEOMETRY, 2002, 28 (04) :511-533
[10]  
Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672, DOI DOI 10.1145/3422622