Training Variational Autoencoders with Discrete Latent Variables Using Importance Sampling

被引:0
作者
Bartler, Alexander [1 ]
Wiewel, Felix [1 ]
Mauch, Lukas [1 ]
Yang, Bin [1 ]
机构
[1] Univ Stuttgart, Inst Signal Proc & Syst Theory, Stuttgart, Germany
来源
2019 27TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO) | 2019年
关键词
variational autoencoder; discrete latent variables; importance sampling;
D O I
10.23919/eusipco.2019.8902811
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The Variational Autoencoder (VAE) is a popular generative latent variable model that is often used for representation learning. Standard VAEs assume continuous-valued latent variables and are trained by maximization of the evidence lower bound (ELBO). Conventional methods obtain a differentiable estimate of the ELBO with reparametrized sampling and optimize it with Stochastic Gradient Descend (SGD). However, this is not possible if we want to train VAEs with discrete-valued latent variables, since reparametrized sampling is not possible. In this paper, we propose an easy method to train VAEs with binary or categorically valued latent representations. Therefore, we use a differentiable estimator for the ELBO which is based on importance sampling. In experiments, we verify the approach and train two different VAEs architectures with Bernoulli and categorically distributed latent representations on two different benchmark datasets.
引用
收藏
页数:5
相关论文
共 17 条
  • [1] An J., 2015, Special Lecture on IE, V1, P1
  • [2] [Anonymous], 2013, CoRR abs/1308.3432
  • [3] Chen Xi, 2016, Advances in Neural Information Processing Systems, V29
  • [4] Dilokthanakul N., 2016, DEEP UNSUPERVISED CL
  • [5] Grathwohl Will, 2018, INT C LEARNING REPRE
  • [6] Jang E., 2017, CATEGORICAL REPARAME
  • [7] Jiang Z., 2016, ARXIV161105148
  • [8] Kingma D. P, 2013, Auto-Encoding Variational Bayes
  • [9] Kingma DP, 2016, 30 C NEURAL INFORM P, V29
  • [10] LeCun Y, 1998, MNIST handwritten digit database