Self Attention in Variational Sequential Learning for Summarization

被引:18
作者
Chien, Jen-Tzung [1 ]
Wang, Chun-Wei [1 ]
机构
[1] Natl Chiao Tung Univ, Dept Elect & Comp Engn, Hsinchu, Taiwan
来源
INTERSPEECH 2019 | 2019年
关键词
sequence generation; variational autoencoder; sequence-to-sequence learning; attention mechanism;
D O I
10.21437/Interspeech.2019-1548
中图分类号
R36 [病理学]; R76 [耳鼻咽喉科学];
学科分类号
100104 ; 100213 ;
摘要
Attention mechanism plays a crucial role in sequential learning for many speech and language applications. However, it is challenging to develop a stochastic attention in a sequence-to-sequence model which consists of two recurrent neural networks (RNNs) as the encoder and decoder. The problem of posterior collapse happens in variational inference and results in the estimated latent variables close to a standard Gaussian prior so that the information from input sequence is disregarded in learning process. This paper presents a new recurrent autoencoder for sentence representation where a self attention scheme is incorporated to activate the interaction between inference and generation in training procedure. In particular, a stochastic RNN decoder is implemented to provide additional latent variable to fulfill self attention for sentence reconstruction. The posterior collapse is alleviated. The latent information is sufficiently attended in variational sequential learning. During test phase, the estimated prior distribution of decoder is sampled for stochastic attention and generation. Experiments on Penn Treebank and Yelp 2013 show the desirable generation performance in terms of perplexity. The visualization of attention weights also illustrates the usefulness of self attention. The evaluation on DUC 2007 demonstrates the merit of variational recurrent autoencoder for document summarization.
引用
收藏
页码:1318 / 1322
页数:5
相关论文
共 31 条
[1]  
[Anonymous], 2014, ARXIV14126581
[2]  
[Anonymous], 2017, ARXIV171105717
[3]  
[Anonymous], 2017, P INT C MACH LEARN
[4]  
Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
[5]  
Bowman Samuel R, 2015, P 20 SIGNLL C COMP N, DOI DOI 10.18653/V1/K16-1002
[6]  
Chan W, 2016, INT CONF ACOUST SPEE, P4960, DOI 10.1109/ICASSP.2016.7472621
[7]  
Chang YL, 2009, INT CONF ACOUST SPEE, P1689, DOI 10.1109/ICASSP.2009.4959927
[8]  
Chien J.-T., 2019, P ANN M ASS COMP LIN
[9]  
Chien JT, 2019, INT CONF ACOUST SPEE, P3202, DOI [10.1109/icassp.2019.8683771, 10.1109/ICASSP.2019.8683771]
[10]   Hierarchical Theme and Topic Modeling [J].
Chien, Jen-Tzung .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2016, 27 (03) :565-578