Probabilistic Belief Embedding for Large-Scale Knowledge Population

被引:6
作者
Fan, Miao [1 ]
Zhou, Qiang [1 ]
Abel, Andrew [2 ]
Zheng, Thomas Fang [1 ]
Grishman, Ralph [3 ]
机构
[1] Tsinghua Univ, Div Tech Innovat & Dev, Tsinghua Natl Lab Informat Sci & Technol, CSLT, Beijing 100084, Peoples R China
[2] Univ Stirling, Sch Nat Sci, Comp Sci & Math, Room 4B59,Cottrell Bldg, Stirling FK9 4LA, Scotland
[3] NYU, Courant Inst Math Sci, Dept Comp Sci, New York, NY 10003 USA
基金
美国国家科学基金会;
关键词
Knowledge population; Belief embedding; Entity inference; Relation prediction; Triplet classification;
D O I
10.1007/s12559-016-9425-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To populate knowledge repositories, such as WordNet, Freebase and NELL, two branches of research have grown separately for decades. On the one hand, corpus-based methods which leverage unstructured free texts have been explored for years; on the other hand, some recently emerged embedding-based approaches use structured knowledge graphs to learn distributed representations of entities and relations. But there are still few comprehensive and elegant models that can integrate those large-scale heterogeneous resources to satisfy multiple subtasks of knowledge population including entity inference, relation prediction and triplet classification. This paper contributes a novel embedding model which estimates the probability of each candidate belief < h,r,t,m > in a large-scale knowledge repository via simultaneously learning distributed representations for entities (h and t), relations (r) and the words in relation mentions (m). It facilitates knowledge population by means of simple vector operations to discover new beliefs. Given an imperfect belief, we can not only infer the missing entities and predict the unknown relations, but also identify the plausibility of the belief, just by leveraging the learned embeddings of remaining evidence. To demonstrate the scalability and the effectiveness of our model, experiments have been conducted on several large-scale repositories which contain millions of beliefs from WordNet, Freebase and NELL, and the results are compared with other cutting-edge approaches via comparing the performance assessed by the tasks of entity inference, relation prediction and triplet classification with their respective metrics. Extensive experimental results show that the proposed model outperforms the state of the arts with significant improvements. The essence of the improvements comes from the capability of our model that encodes not only structured knowledge graph information, but also unstructured relation mentions, into continuous vector spaces, so that we can bridge the gap of one-hot representations, and expect to discover certain relevance among entities, relations and even words in relation mentions.
引用
收藏
页码:1087 / 1102
页数:16
相关论文
共 37 条
[1]  
[Anonymous], 2012, Advances in neural information processing systems
[2]  
[Anonymous], 2013, P 2013 C EMP METH NA
[3]  
[Anonymous], 2007, AAAI
[4]  
Bollacker K., 2008, P 2008 ACM SIGMOD IN, P1247, DOI DOI 10.1145/1376616.1376746
[5]  
Bordes A, AAAI 2011
[6]  
Bordes A, 2013, P 26 INT C NEURAL IN, P2787
[7]   A semantic matching energy function for learning with multi-relational data Application to word-sense disambiguation [J].
Bordes, Antoine ;
Glorot, Xavier ;
Weston, Jason ;
Bengio, Yoshua .
MACHINE LEARNING, 2014, 94 (02) :233-259
[8]  
Carlson A., 2010, AAAI, V5, P3
[9]  
Fan M, 2015, ARXIV150308155
[10]  
Fan M, 2015, ARXIV150401683