Revisiting Text and Knowledge Graph Joint Embeddings: The Amount of Shared Information Matters!

被引:0
作者
Rosso, Paolo [1 ]
Yang, Dingqi [1 ]
Cudre-Mauroux, Philippe [1 ]
机构
[1] Univ Fribourg, eXascale Infolab, Fribourg, Switzerland
来源
2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA) | 2019年
基金
瑞士国家科学基金会;
关键词
Word embeddings; Knowledge Graph embeddings; regularization;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Jointly learning embeddings from text and a Knowledge Graph benefits both word and entity/relation embeddings by taking advantage of both large-scale unstructured content (text) and high-quality structured data (the Knowledge Graph). Current techniques leverage anchors to associate entities in the Knowledge Graph to corresponding words in the text corpus; these anchors are then used to generate additional learning samples during the embedding learning process. However, we show in this paper that such techniques yield suboptimal results, as they fail to control the amount of shared information between the two data sources during the joint learning process. Moreover, the additional learning samples often incur significant computational overhead. Aiming at releasing the power of such joint embeddings, we propose JOINER, a new joint text and Knowledge Graph embedding method using regularization. JOINER not only preserves co-occurrence between words in a text corpus and relations between entities in a Knowledge Graph, it also provides the flexibility to control the amount of information shared between the two data sources via regularization. Our method does not generate additional learning samples, which makes it computationally efficient. Our extensive empirical evaluation on real datasets shows the superiority of JOINER across different evaluation tasks, including analogical reasoning, link prediction, and relation extraction. Compared to state-of-the-art techniques generating additional learning samples from a set of anchors, our method yields better results (with up to 4.3% absolute improvement) and significantly less computational overhead (76% less learning time overhead).
引用
收藏
页码:2465 / 2473
页数:9
相关论文
共 43 条
[1]  
[Anonymous], 2005, P INT C VER LARG DAT
[2]  
[Anonymous], 2015, P 24 ACM INT C INF K
[3]  
[Anonymous], 2013, P 24 ACM C HYPERTEXT, DOI [DOI 10.1145/2481492.2481505, 10.1145/2481492.2481505]
[4]  
[Anonymous], 2016, CONLL
[5]  
Balazevic Ivana, 2018, ARXIV180807018
[6]   Aligning Knowledge Base and Document Embedding Models Using Regularized Multi-Task Learning [J].
Baumgartner, Matthias ;
Zhang, Wen ;
Paudel, Bibek ;
Dell'Aglio, Daniele ;
Chen, Huajun ;
Bernstein, Abraham .
SEMANTIC WEB - ISWC 2018, PT I, 2018, 11136 :21-37
[7]   A neural probabilistic language model [J].
Bengio, Y ;
Ducharme, R ;
Vincent, P ;
Jauvin, C .
JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (06) :1137-1155
[8]  
Bollacker K., 2008, P 2008 ACM SIGMOD IN, P1247
[9]  
Bordes A, 2013, ADV NEURAL INFORM PR, V26
[10]  
Chen L., 2016, IEEE T HUMAN MACHINE, V47, P380