Multi-grained unsupervised evidence retrieval for question answering

被引:0
作者
You, Hao [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing 100093, Peoples R China
[2] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing 100049, Peoples R China
关键词
Evidence retrieval; Question answering; Machine reading comprehension; Pre-trained language model;
D O I
10.1007/s00521-023-08892-4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Evidence retrieval is a crucial step in question answering (QA) tasks, which can filter original context to provide supporting evidence for reading comprehension and reduce the time of answer inference. Most research interests in evidence retrieval have been paid to supervised methods, while unsupervised evidence retrieval received limited attention. However, most existing works for unsupervised evidence retrieval regard sentence-level and passage-level evidence retrieval as two independent processes. In order to fuse these two levels of information, we propose an efficient and unsupervised multi-grained evidence retrieval method for QA, which considers multiple interactions, including query-sentence, query-passage, and passage-passage. Specifically, we use sentence-level retrieval to obtain an evidence framework. Then we propose a score fusion mechanism to model the unilateral guiding relationship between sentence-level and passage-level retrieval. On the basis of score fusion, we propose a gated selection mechanism to retrieve evidence passages, which can supplement reasoning information for the evidence framework. Extensive experiments on open-domain and multi-hop QA datasets show that our method has fast retrieval speed while achieving better retrieval and QA performance than baselines.
引用
收藏
页码:21247 / 21257
页数:11
相关论文
共 35 条
  • [1] Banerjee P, 2020, PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), P151
  • [2] Berger A., 2000, SIGIR Forum, V34, P192
  • [3] Reading Wikipedia to Answer Open-Domain Questions
    Chen, Danqi
    Fisch, Adam
    Weston, Jason
    Bordes, Antoine
    [J]. PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 1870 - 1879
  • [4] Chen JF, 2021, Arxiv, DOI arXiv:1910.02610
  • [5] Cheng H, 2021, 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, P3080
  • [6] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [7] Dhingra B, 2017, Arxiv, DOI arXiv:1707.03904
  • [8] Hsu TY, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P5933
  • [9] Izacard G, 2022, Arxiv, DOI arXiv:2012.04584
  • [10] SpanBERT: Improving Pre-training by Representing and Predicting Spans
    Joshi, Mandar
    Chen, Danqi
    Liu, Yinhan
    Weld, Daniel S.
    Zettlemoyer, Luke
    Levy, Omer
    [J]. TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2020, 8 : 64 - 77