Variational Memory Encoder-Decoder

被引:0
作者
Hung Le [1 ]
Truyen Tran [1 ]
Thin Nguyen [1 ]
Venkatesh, Svetha [1 ]
机构
[1] Deakin Univ, Appl AI Inst, Geelong, Vic, Australia
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018) | 2018年 / 31卷
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Introducing variability while maintaining coherence is a core task in learning to generate utterances in conversation. Standard neural encoder-decoder models and their extensions using conditional variational autoencoder often result in either trivial or digressive responses. To overcome this, we explore a novel approach that injects variability into neural encoder-decoder via the use of external memory as a mixture model, namely Variational Memory Encoder-Decoder (VMED). By associating each memory read with a mode in the latent mixture distribution at each timestep, our model can capture the variability observed in sequential data such as natural conversations. We empirically compare the proposed model against other recent approaches on various conversational datasets. The results show that VMED consistently achieves significant improvement over others in both metric-based and qualitative evaluations.
引用
收藏
页数:11
相关论文
共 40 条
[1]  
[Anonymous], 2014, Generating sequences with recurrent neural networks
[2]  
[Anonymous], 2014, Advances in neural information processing systems
[3]  
[Anonymous], 2013, P 2013 C EMPIRICAL M
[4]  
[Anonymous], 2015, P 3 INT C LEARN REPR
[5]  
[Anonymous], 2007, IEEE INT C AC SPEECH
[6]  
[Anonymous], 2016, NIPS WORKSH BAYES DE
[7]  
[Anonymous], 2017, ARXIV170204649
[8]  
[Anonymous], 2017, P 2017 C EMP METH NA, DOI DOI 10.18653/V1/D17-1040
[9]  
[Anonymous], 2016, ARXIV
[10]  
[Anonymous], 2014, 9 WORKSH STAT MACH T, DOI DOI 10.3115/V1/W14-3346