Simple Large-scale Relation Extraction from Unstructured Text

被引:0
作者
Christodoulopoulos, Christos [1 ]
Mittal, Arpit [1 ]
机构
[1] Amazon Res Cambridge, Cambridge, England
来源
PROCEEDINGS OF THE ELEVENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2018) | 2018年
关键词
relation extraction; distant supervision; unstructured text;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Knowledge-based question answering relies on the availability of facts, the majority of which cannot be found in structured sources (e.g. Wikipedia info-boxes, Wikidata). One of the major components of extracting facts from unstructured text is Relation Extraction (RE). In this paper we propose a novel method for creating distant (weak) supervision labels for training a large-scale RE system. We also provide new evidence about the effectiveness of neural network approaches by decoupling the model architecture from the feature design of a state-of-the-art neural network system. Surprisingly, a much simpler classifier trained on similar features performs on par with the highly complex neural network system (at 75x reduction to the training time), suggesting that the features are a bigger contributor to the final performance.
引用
收藏
页码:215 / 222
页数:8
相关论文
共 30 条
[1]  
[Anonymous], 2005, P HUM LANG TECHN C C
[2]  
[Anonymous], 2002, MALLET: A machine learning for language toolkit
[3]  
[Anonymous], 2013, EFFICIENT ESTIMATION
[4]  
[Anonymous], 2011, P 2011 C EMPIRICAL M
[5]   DBpedia: A nucleus for a web of open data [J].
Auer, Soeren ;
Bizer, Christian ;
Kobilarov, Georgi ;
Lehmann, Jens ;
Cyganiak, Richard ;
Ives, Zachary .
SEMANTIC WEB, PROCEEDINGS, 2007, 4825 :722-+
[6]   SPACE/TIME TRADE/OFFS IN HASH CODING WITH ALLOWABLE ERRORS [J].
BLOOM, BH .
COMMUNICATIONS OF THE ACM, 1970, 13 (07) :422-&
[7]  
Bollacker K, 2008, P 2008 ACM SIGMOD IN, P1247, DOI DOI 10.1145/1376616.1376746
[8]  
Brown P. F., 1992, Computational Linguistics, V18, P467
[9]  
Chinchor N, 1993, Computational Linguistics, V19, P409
[10]  
Chollet F., 2015, about us