Investigation of Recurrent-Neural-Network Architectures and Learning Methods for Spoken Language Understanding

被引:0
作者
Mesnil, Gregoire [1 ,3 ]
He, Xiaodong [2 ]
Deng, Li [2 ]
Bengio, Yoshua [1 ]
机构
[1] Univ Montreal, Montreal, PQ H3C 3J7, Canada
[2] Microsoft Res, Redmond, WA USA
[3] Univ Rouen, Rouen, France
来源
14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5 | 2013年
关键词
spoken language understanding; word embeddings; recurrent neural network; slot filling;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
One of the key problems in spoken language understanding (SLU) is the task of slot filling. In light of the recent success of applying deep neural network technologies in domain detection and intent identification, we carried out an in-depth investigation on the use of recurrent neural networks for the more difficult task of slot filling involving sequence discrimination. In this work, we implemented and compared several important recurrent-neural-network architectures, including the Elman-type and Jordan-type recurrent networks and their variants. To make the results easy to reproduce and compare, we implemented these networks on the common Theano neural network toolkit, and evaluated them on the ATIS benchmark. We also compared our results to a conditional random fields (CRF) baseline. Our results show that on this task, both types of recurrent networks outperform the CRF baseline substantially, and a bi-directional Jordan type network that takes into account both past and future dependencies among slots works best, outperforming a CRFbased baseline by 14% in relative error reduction.
引用
收藏
页码:3738 / 3742
页数:5
相关论文
共 26 条
[1]  
[Anonymous], 2011, P ICML
[2]  
[Anonymous], 1997, IEEE T SIGNAL PROCES
[3]  
[Anonymous], 2011, J MACHINE LEARNING R
[4]  
[Anonymous], IEEE WORKSH SPOK LAN
[5]  
[Anonymous], P PYTH SCI COMP C SC
[6]  
[Anonymous], 2012, ICASSP
[7]  
[Anonymous], 2008, Neural Inf. Process. Syst.
[8]  
[Anonymous], 2012, ARXIV12070580V1
[9]  
Bengio Y., NIPS 2000
[10]  
Collobert Ronan., ICML 2008