Prompting disentangled embeddings for knowledge graph completion with pre-trained language model

被引:3
作者
Geng, Yuxia [1 ,2 ]
Chen, Jiaoyan [3 ]
Zeng, Yuhang [4 ]
Chen, Zhuo [5 ]
Zhang, Wen [6 ]
Pan, Jeff Z. [7 ]
Wang, Yuxiang [2 ]
Xu, Xiaoliang [2 ]
机构
[1] Powerchina Huadong Engn Corp Ltd, Hangzhou 311112, Peoples R China
[2] Hangzhou Dianzi Univ, Sch Comp Sci, Hangzhou 310018, Peoples R China
[3] UNIV MANCHESTER, Dept Comp Sci, MANCHESTER M13 9PL, England
[4] Hangzhou Dianzi Univ, HDU ITMO Joint Inst, Hangzhou 310018, Peoples R China
[5] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou 310028, Peoples R China
[6] Zhejiang Univ, Sch Software Technol, Ningbo 315048, Peoples R China
[7] Univ Edinburgh, Sch Informat, Edinburgh EH8 9AB, Scotland
基金
英国工程与自然科学研究理事会;
关键词
Knowledge graph completion; Pre-trained language model; Prompt tuning; Disentangled embedding;
D O I
10.1016/j.eswa.2024.126175
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Both graph structures and textual information play a critical role in Knowledge Graph Completion (KGC). With the success of Pre-trained Language Models (PLMs) such as BERT, they have been applied for text encoding for KGC. However, the current methods mostly prefer to fine-tune PLMs, leading to huge training costs and limited scalability to larger PLMs. In contrast, we propose to utilize prompts and perform KGC on a frozen PLM with only the prompts trained. Accordingly, we propose a new KGC method named PDKGC with two prompts - a hard task prompt which is to adapt the KGC task to the PLM pre-training task of token prediction, and a disentangled structure prompt which learns disentangled graph representation so as to enable the PLM to combine more relevant structure knowledge with the text information. With the two prompts, PDKGC builds a textual predictor and a structural predictor, respectively, and their combination leads to more comprehensive entity prediction. Solid evaluation on three widely used KGC datasets has shown that PDKGC often outperforms the baselines including the state-of-the-art, and its components are all effective. Our codes and data are available at https://github.com/genggengcss/PDKGC.
引用
收藏
页数:13
相关论文
共 58 条
[1]  
[Anonymous], 2015, ICLR POSTER
[2]  
Bollacker Kurt., P 2008 ACM SIGMOD IN
[3]  
Bordes A, 2013, ADV NEURAL INF PROCE, V26
[4]  
Brown TB, 2020, ADV NEUR IN, V33
[5]  
Chen C, 2023, FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), P11489
[6]  
Chen Chen, 2022, P 29 INT C COMP LING, P4005
[7]   Contextual semantic embeddings for ontology subsumption prediction [J].
Chen, Jiaoyan ;
He, Yuan ;
Geng, Yuxia ;
Jimenez-Ruiz, Ernesto ;
Dong, Hang ;
Horrocks, Ian .
WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2023, 26 (05) :2569-2591
[8]   Knowledge Graph Completion: A Review [J].
Chen, Zhe ;
Wang, Yuehan ;
Zhao, Bin ;
Cheng, Jing ;
Zhao, Xin ;
Duan, Zongtao .
IEEE ACCESS, 2020, 8 :192435-192456
[9]  
Chepurova A, 2023, Arxiv, DOI [arXiv:2311.01326, 10.18653/v1/2023.findingsemnlp.352, DOI 10.18653/V1/2023.FINDINGSEMNLP.352]
[10]   MEM-KGC: Masked Entity Model for Knowledge Graph Completion With Pre-Trained Language Model [J].
Choi, Bonggeun ;
Jang, Daesik ;
Ko, Youngjoong .
IEEE ACCESS, 2021, 9 :132025-132032