A Question-Focused Multi-Factor Attention Network for Question Answering

被引:0
作者
Kundu, Souvik [1 ]
Hwee Tou Ng [1 ]
机构
[1] Natl Univ Singapore, Dept Comp Sci, Singapore, Singapore
来源
THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE | 2018年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural network models recently proposed for question answering (QA) primarily focus on capturing the passage question relation. However, they have minimal capability to link relevant facts distributed across multiple sentences which is crucial in achieving deeper understanding, such as performing multi-sentence reasoning, co reference resolution, etc. They also do not explicitly focus on the question and answer type which often plays a critical role in QA. In this paper, we propose a novel end-to-end question-focused multi-factor attention network for answer extraction. Multi-factor attentive encoding using tensor-based transformation aggregates meaningful facts even when they are located in multiple sentences. To implicitly infer the answer type, we also propose a max-attentional question aggregation mechanism to encode a question vector based on the important words in a question. During prediction, we incorporate sequence-level encoding of the first wh-word and its immediately following word as an additional source of question type information. Our proposed model achieves significant improvements over the best prior state-of-the-art results on three large-scale challenging QA datasets, namely NewsQA, TriviaQA, and SearchQA.
引用
收藏
页码:5828 / 5835
页数:8
相关论文
共 31 条
[1]  
[Anonymous], 2014, P ACL
[2]  
[Anonymous], 2016, ARXIV161009996
[3]  
[Anonymous], 2016, ARXIV161204211
[4]  
[Anonymous], 2016, ARXIV161101436
[5]  
[Anonymous], 2016, ARXIV
[6]  
[Anonymous], 2017, ACL
[7]  
[Anonymous], 2017, P ACL
[8]  
[Anonymous], 2017, AAAI C ARTIFICIAL
[9]  
Chen Danqi, 2016, P ACL
[10]   Towards End-to-End Reinforcement Learning of Dialogue Agents for Information Access [J].
Dhingra, Bhuwan ;
Li, Lihong ;
Li, Xiujun ;
Gao, Jianfeng ;
Chen, Yun-Nung ;
Ahmed, Faisal ;
Deng, Li .
PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, :484-495