A Unified Generative Retriever for Knowledge-Intensive Language Tasks via Prompt Learning

被引:13
作者
Chen, Jiangui [1 ]
Zhang, Ruqing [1 ,2 ]
Guo, Jiafeng [1 ]
de Rijke, Maarten [2 ]
Liu, Yiqun [3 ]
Fan, Yixing [1 ]
Cheng, Xueqi [1 ]
机构
[1] Univ Chinese Acad Sci, CAS, ICT, CAS Key Lab Network Data Sci & Technol, Beijing, Peoples R China
[2] Univ Amsterdam, Amsterdam, Netherlands
[3] Tsinghua Univ, Dept Comp Sci & Technol, Beijing, Peoples R China
来源
PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023 | 2023年
基金
中国国家自然科学基金;
关键词
Knowledge-intensive language tasks; Generative retrieval; Unified retriever;
D O I
10.1145/3539618.3591631
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Knowledge-intensive language tasks (KILTs) benefit from retrieving high-quality relevant contexts from large external knowledge corpora. Learning task-specific retrievers that return relevant contexts at an appropriate level of semantic granularity, such as a document retriever, passage retriever, sentence retriever, and entity retriever, may help to achieve better performance on the end-to-end task. But a task-specific retriever usually has poor generalization ability to new domains and tasks, and it may be costly to deploy a variety of specialised retrievers in practice. We propose a unified generative retriever (UGR) that combines task-specific effectiveness with robust performance over different retrieval tasks in KILTs. To achieve this goal, we make two major contributions: (i) To unify different retrieval tasks into a single generative form, we introduce an n-gram-based identifier for relevant contexts at different levels of granularity in KILTs. And (ii) to address different retrieval tasks with a single model, we employ a prompt learning strategy and investigate three methods to design prompt tokens for each task. In this way, the proposed UGR model can not only share common knowledge across tasks for better generalization, but also perform different retrieval tasks effectively by distinguishing task-specific characteristics. We train UGR on a heterogeneous set of retrieval corpora with well-designed prompts in a supervised and multi-task fashion. Experimental results on the KILT benchmark demonstrate the effectiveness of UGR on in-domain datasets, out-of-domain datasets, and unseen tasks.
引用
收藏
页码:1448 / 1457
页数:10
相关论文
共 47 条
  • [1] Akbik A, 2019, NAACL HLT 2019: THE 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES: PROCEEDINGS OF THE DEMONSTRATIONS SESSION, P54
  • [2] Bevilacqua Michele, 2022, ARXIV220410628
  • [3] Burrows M, 1994, 124 DIG EQ CORP
  • [4] Reading Wikipedia to Answer Open-Domain Questions
    Chen, Danqi
    Fisch, Adam
    Weston, Jason
    Bordes, Antoine
    [J]. PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 1870 - 1879
  • [5] CorpusBrain: Pre-train a Generative Retrieval Model for Knowledge-Intensive Language Tasks
    Chen, Jiangui
    Zhang, Ruqing
    Guo, Jiafeng
    Liu, Yiqun
    Fan, Yixing
    Cheng, Xueqi
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 191 - 200
  • [6] GERE: Generative Evidence Retrieval for Fact Verification
    Chen, Jiangui
    Zhang, Ruqing
    Guo, Jiafeng
    Fan, Yixing
    Cheng, Xueqi
    [J]. PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 2184 - 2189
  • [7] Collobert R., 2008, P 25 INT C MACH LEAR, P160, DOI [10.1145/1390156.1390177, DOI 10.1145/1390156.1390177]
  • [8] De Cao Nicola, 2020, INT C LEARN REPR
  • [9] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [10] Dinan Emily, 2018, INT C LEARN REPR