Self-attention presents low-dimensional knowledge graph embeddings for link prediction

被引:50
作者
Baghershahi, Peyman [1 ]
Hosseini, Reshad [1 ]
Moradi, Hadi [1 ]
机构
[1] Univ Tehran, Coll Engn, Sch ECE, Tehran, Iran
关键词
Knowledge graph completion; Link prediction; Self-attention;
D O I
10.1016/j.knosys.2022.110124
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A few models have tried to tackle the link prediction problem, also known as knowledge graph comple-tion, by embedding knowledge graphs in comparably lower dimensions. However, the state-of-the-art results are attained at the cost of considerably increasing the dimensionality of embeddings which causes scalability issues in the case of huge knowledge bases. Transformers have been successfully used recently as powerful encoders for knowledge graphs, but available models still have scalability issues. To address this limitation, we introduce a Transformer-based model to gain expressive low-dimensional embeddings. We utilize a large number of self-attention heads as the key to applying query-dependent projections to capture mutual information between entities and relations. Empirical results on WN18RR and FB15k-237 as standard link prediction benchmarks demonstrate that our model has favorably comparable performance with the current state-of-the-art models. Notably, we yield our promising results with a significant reduction of 66.9% in the dimensionality of embeddings compared to the five best recent state-of-the-art competitors on average.(c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页数:8
相关论文
共 39 条
[1]  
[Anonymous], 2013, Procings of the 26th international conference on neural information processing systems, NIPS13
[2]  
Balazevic I, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P5185
[3]   Hypernetwork Knowledge Graph Embeddings [J].
Balazevic, Ivana ;
Allen, Carl ;
Hospedales, Timothy M. .
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: WORKSHOP AND SPECIAL SESSIONS, 2019, 11731 :553-565
[4]  
Balazevic Ivana, 2019, Advances in Neural Information Processing Systems, V32
[5]  
Bansal T, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P4387
[6]  
Bordes A., 2013, P 26 INT C NEURAL IN, P2787
[7]  
Chami I., 2020, P 58 ANN M ASS COMPU, P6901, DOI 10.18653/v1/2020.acl-main.617
[8]  
Chen SX, 2021, 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), P10395
[9]  
Das R., 2018, INT C LEARNING REPRE
[10]  
Dettmers T, 2018, AAAI CONF ARTIF INTE, P1811