VARIATIONAL AND HIERARCHICAL RECURRENT AUTOENCODER

被引:0
作者
Chien, Jen-Tzung [1 ]
Wang, Chun-Wei [1 ]
机构
[1] Natl Chiao Tung Univ, Dept Elect & Comp Engn, Hsinchu, Taiwan
来源
2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP) | 2019年
关键词
Sequence generation; recurrent neural network; variational autoencoder; hierarchical model;
D O I
10.1109/icassp.2019.8683771
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Despite a great success in learning representation for image data, it is challenging to learn the stochastic latent features from natural language based on variational inference. The difficulty in stochastic sequential learning is due to the posterior collapse caused by an autoregressive decoder which is prone to be too strong to learn sufficient latent information during optimization. To compensate this weakness in learning procedure, a sophisticated latent structure is required to assure good convergence so that random features are sufficiently captured for sequential decoding. This study presents a new variational recurrent autoencoder (VRAE) for sequence reconstruction. There are two complementary encoders consisting of a long short-term memory (LSTM) and a pyramid bidirectional LSTM which are merged to discover the global and local dependencies in a hierarchical latent variable model, respectively. Experiments on Penn Treebank and Yelp 2013 demonstrate that the proposed hierarchical VRAE is able to learn the complementary representation as well as tackle the posterior collapse in stochastic sequential learning. The performance of recurrent autoencoder is substantially improved in terms of perplexity.
引用
收藏
页码:3202 / 3206
页数:5
相关论文
共 31 条
[1]  
[Anonymous], 2015, Bayesian Speech and Language Processing
[2]  
[Anonymous], 2018, P IEEE INT WORKSH MA
[3]  
[Anonymous], P IEEE INT C AC SPEE
[4]  
[Anonymous], 2017, ADV NEURAL INFORM PR
[5]  
[Anonymous], 2014, ARXIV14126581
[6]  
[Anonymous], 2016, P 9 ISCA WORKSH SPEE
[7]  
[Anonymous], 2017, ADV NEURAL INFORM PR
[8]  
[Anonymous], 2017, P INT C LEARN REPR
[9]  
[Anonymous], 2018, ARXIV180704863
[10]  
[Anonymous], 2017, CoRR