Multiple premises entailment recognition based on attention and gate mechanism

被引:3
作者
Wu, Pin [1 ]
Lei, Zhidan [1 ]
Zhou, Quan [1 ]
Zhu, Rukang [1 ]
Chang, Xuting [1 ]
Sun, Junwu [1 ]
Zhang, Wenjie [1 ]
Guo, Yike [1 ,2 ]
机构
[1] Shanghai Univ, Shanghai, Peoples R China
[2] Imperial Coll London, London, England
关键词
Natural language inference; Multiple premise entailment; Attention mechanism; Gate mechanism; Fine-tune;
D O I
10.1016/j.eswa.2020.113214
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-premise natural language inference provides important technical support for automatic question answering, machine reading comprehension and other application fields. Existing approaches for Multiple Premises Entailment (MPE) task are to convert MPE data into Single Premise Entailment (SPE) data format, then MPE is handled in the same way as SPE. This process ignores the unique characteristics of multi-premise, which will result in loss of semantics. This paper proposes a mechanism based on Attention and Gate Fusion Network (AGNet). AGNet adopts a "Local Matching-Integration" strategy to consider the characteristics of multi-premise. In this process, an attention mechanism combined with a matching gate mechanism can fully describe the relationship between the premise and hypothesis. A self-attention mechanism and a fusion gate mechanism can deeply exploit the relationship from the multi-premise. In order to avoid over-fitting problem, we propose a pre-training method for our model. In terms of computational complexity, AGNet has good parallelism, reduces the time complexity to O(1) in the process of matching. The experiments show that our model has achieved new state-of-the-art results on MPE test set. (C) 2020 Elsevier Ltd. All rights reserved.
引用
收藏
页数:8
相关论文
共 36 条
  • [1] Adams Rod., 2006, P OF THE 2 PASCAL CH, P128
  • [2] Akhmatova E., 2005, P PASC CHALL WORKSH
  • [3] [Anonymous], P INT C LEARN REPR I
  • [4] Ba Jimmy Lei, 2016, LAYER NORMALIZATION, V1050, P21, DOI 10.48550/arXiv.1607.06450
  • [5] Bayer S., 2005, P PATT AN STAT MOD C
  • [6] Bowman S.R., 2015, EMNLP
  • [7] Enhanced LSTM for Natural Language Inference
    Chen, Qian
    Zhu, Xiaodan
    Ling, Zhenhua
    Wei, Si
    Jiang, Hui
    Inkpen, Diana
    [J]. PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 1657 - 1668
  • [8] Dagan I., 2004, Probabilistic textual entailment: Generic applied modeling of language variability
  • [9] De Marneffe M.C., 2008, Proceedings of ACL-08: HLT, P1039
  • [10] Ghaeini R., 2018, P 2018 C N AM CHAPT