Enhanced LSTM for Natural Language Inference

被引:637
作者
Chen, Qian [1 ]
Zhu, Xiaodan [2 ]
Ling, Zhenhua [1 ]
Wei, Si [3 ]
Jiang, Hui [4 ]
Inkpen, Diana [5 ]
机构
[1] Univ Sci & Technol China, Hefei, Anhui, Peoples R China
[2] Natl Res Council Canada, Ottawa, ON, Canada
[3] IFLYTEK Res, Beijing, Peoples R China
[4] York Univ, N York, ON, Canada
[5] Univ Ottawa, Ottawa, ON, Canada
来源
PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1 | 2017年
关键词
D O I
10.18653/v1/P17-1152
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Reasoning and inference are central to human and artificial intelligence. Modeling inference in human language is very challenging. With the availability of large annotated data (Bowman et al., 2015), it has recently become feasible to train neural network based inference models, which have shown to be very effective. In this paper, we present a new state-of-the-art result, achieving the accuracy of 88.6% on the Stanford Natural Language Inference Dataset. Unlike the previous top models that use very complicated network architectures, we first demonstrate that carefully designing sequential inference models based on chain LSTMs can outperform all previous models. Based on this, we further show that by explicitly considering recursive architectures in both local inference modeling and inference composition, we achieve additional improvement. Particularly, incorporating syntactic parsing information contributes to our best result-it further improves the performance even when added to the already very strong model.
引用
收藏
页码:1657 / 1668
页数:12
相关论文
共 39 条
[1]  
[Anonymous], ABS161104741 CORR
[2]  
[Anonymous], 2016, ABS160704492 CORR
[3]  
[Anonymous], 2003, P 41 ANN M ASS COMP
[4]  
[Anonymous], 2014, P 8 WORKSH SYNT SEM
[5]  
[Anonymous], HUMAN LANGUAGE TECHN
[6]  
[Anonymous], 2015, ABS151106361 CORR
[7]  
[Anonymous], ABSARXIV170304617V2
[8]  
[Anonymous], 2015, P 4 JOINT C LEX COMP, DOI DOI 10.18653/V1/S15-1002
[9]  
[Anonymous], 1997, Neural Computation
[10]  
[Anonymous], P COLING 2014 25 INT