SSMFRP: Semantic Similarity Model for Relation Prediction in KBQA Based on Pre-trained Models

被引:1
作者
Wang, Ziming [1 ]
Xu, Xirong [1 ]
Li, Xinzi [1 ]
Song, Xiaoying [1 ]
Wei, Xiaopeng [1 ]
Huang, Degen [1 ]
机构
[1] Dalian Univ Technol, Sch Comp Sci & Technol, Dalian, Peoples R China
来源
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2022, PT II | 2022年 / 13530卷
关键词
Knowledge base question answering; Relation prediction; Semantic similarity; Pre-trained model;
D O I
10.1007/978-3-031-15931-2_25
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pre-trained Relation Extraction (RE) models are widely employed in relation prediction in Knowledge Base Question Answering (KBQA). However, pre-trained models are usually optimized and evaluated on datasets (e.g. GLUE) which contain various Natural Language Processing (NLP) tasks except a RE task. As a result, it is difficult to select a best pre-trained model for relation prediction unless we evaluate all available pre-trained models on a relation prediction dataset. As the Semantic Similarity (SS) task in GLUE is similar to a RE task, a Semantic Similarity Model for Relation Prediction (SSMFRP) is proposed in this paper to convert a RE task in relation prediction to a SS task. In our model, a relation candidate in a RE model is converted into the corresponding question which contains a relation candidate. Then a modified SS model is employed to find the best-matched relation. Experimental results show that the effectiveness of our proposed model in relation prediction is related to the effectiveness of the original pre-trained model in SS and GLUE. Our model achieves an average accuracy of 91.2% with various pre-trained models and outperforms original models by an average margin of 1.8% with the similar training cost. In addition, further experiments show that our model is robust to abnormal input and outperforms original models by an average margin of 1.0% on datasets of abnormal input.
引用
收藏
页码:294 / 306
页数:13
相关论文
共 20 条
  • [1] Bollacker K., 2008, SIGMOD C, P1247, DOI [10.1145/1376616.1376746, DOI 10.1145/1376616.1376746]
  • [2] Bordes A, 2015, Arxiv, DOI arXiv:1506.02075
  • [3] Clark K., 2020, P INT C LEARNING REP
  • [4] Dai ZH, 2016, PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, P800
  • [5] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [6] Fikri FB, 2021, 1ST WORKSHOP ON NATURAL LANGUAGE GENERATION, EVALUATION, AND METRICS (GEM 2021), P24
  • [7] Hu ZK, 2021, 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, P4662
  • [8] Huang B, 2021, P 15 INT WORKSH SEM
  • [9] Huang K, 2021, REPL4NLP 2021: PROCEEDINGS OF THE 6TH WORKSHOP ON REPRESENTATION LEARNING FOR NLP, P307
  • [10] Kingma D P., 2014, P INT C LEARN REPR