Direct Fact Retrieval from Knowledge Graphs without Entity Linking

被引:0
作者
Baek, Jinheon [1 ,3 ]
Aji, Alham Fikri [2 ]
Lehmann, Jens [3 ]
Hwang, Sung Ju [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Daejeon, South Korea
[2] MBZUAI, Abu Dhabi, U Arab Emirates
[3] Amazon, Seattle, WA 98170 USA
来源
PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1 | 2023年
基金
新加坡国家研究基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
There has been a surge of interest in utilizing Knowledge Graphs (KGs) for various natural language processing/understanding tasks. The conventional mechanism to retrieve facts in KGs usually involves three steps: entity span detection, entity disambiguation, and relation classification. However, this approach requires additional labels for training each of the three subcomponents in addition to pairs of input texts and facts, and also may accumulate errors propagated from failures in previous steps. To tackle these limitations, we propose a simple knowledge retrieval framework, which directly retrieves facts from the KGs given the input text based on their representational similarities, which we refer to as Direct Fact Retrieval (DiFaR). Specifically, we first embed all facts in KGs onto a dense embedding space by using a language model trained by only pairs of input texts and facts, and then provide the nearest facts in response to the input text. Since the fact, consisting of only two entities and one relation, has little context to encode, we propose to further refine ranks of top-k retrieved facts with a reranker that contextualizes the input text and the fact jointly. We validate our DiFaR framework on multiple fact retrieval tasks, showing that it significantly outperforms relevant baselines that use the three-step approach.
引用
收藏
页码:10038 / 10055
页数:18
相关论文
共 60 条
[1]  
[Anonymous], 2014, P 2014 C EMP METH NA
[2]  
[Anonymous], 2015, Large-scale simple question answering with memory networks
[3]  
[Anonymous], 2013, P 2013 C EMPIRICAL M
[4]  
[Anonymous], 2021, P 2 WORKSH SCHOL DOC, DOI DOI 10.1007/S11422-021-10018-Y
[5]  
Ayoola T, 2022, 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, NAACL-HLT 2022, P209
[6]  
Bao Junwei, 2016, P COLING 2016 26 INT, P2503
[7]  
Bollacker K, 2008, Proceedings of SIGMOD, SIGMOD '08, P1247
[8]  
Brown TB, 2020, ADV NEUR IN, V33
[9]  
Chakraborty Nilesh, 2019, ARXIV190709361
[10]  
Chen Y, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P2913