Enhancing SPARQL Query Generation for Knowledge Base Question Answering Systems by Learning to Correct Triplets

被引:0
作者
Qi, Jiexing [1 ]
Su, Chang [1 ]
Guo, Zhixin [1 ]
Wu, Lyuwen [1 ]
Shen, Zanwei [1 ]
Fu, Luoyi [1 ]
Wang, Xinbing [1 ]
Zhou, Chenghu [1 ,2 ]
机构
[1] Shanghai Jiao Tong Univ, Sch Elect Informat & Elect Engn, Shanghai 200240, Peoples R China
[2] Chinese Acad Sci, Inst Geog Sci & Nat Resources Res, Beijing 100101, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 04期
关键词
Knowledge Base Question Answering; Text-to-SPARQL; semantic parsing; further pretraining; Triplet Structure;
D O I
10.3390/app14041521
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Generating SPARQL queries from natural language questions is challenging in Knowledge Base Question Answering (KBQA) systems. The current state-of-the-art models heavily rely on fine-tuning pretrained models such as T5. However, these methods still encounter critical issues such as triple-flip errors (e.g., (subject, relation, object) is predicted as (object, relation, subject)). To address this limitation, we introduce TSET (Triplet Structure Enhanced T5), a model with a novel pretraining stage positioned between the initial T5 pretraining and the fine-tuning for the Text-to-SPARQL task. In this intermediary stage, we introduce a new objective called Triplet Structure Correction (TSC) to train the model on a SPARQL corpus derived from Wikidata. This objective aims to deepen the model's understanding of the order of triplets. After this specialized pretraining, the model undergoes fine-tuning for SPARQL query generation, augmenting its query-generation capabilities. We also propose a method named "semantic transformation" to fortify the model's grasp of SPARQL syntax and semantics without compromising the pre-trained weights of T5. Experimental results demonstrate that our proposed TSET outperforms existing methods on three well-established KBQA datasets: LC-QuAD 2.0, QALD-9 plus, and QALD-10, establishing a new state-of-the-art performance (95.0% F1 and 93.1% QM on LC-QuAD 2.0, 75.85% F1 and 61.76% QM on QALD-9 plus, 51.37% F1 and 40.05% QM on QALD-10).
引用
收藏
页数:19
相关论文
共 50 条
  • [31] Dynamic Updating of the Knowledge Base for a Large-Scale Question Answering System
    Liu, Xiao-Yang
    Zhang, Yimeng
    Liao, Yukang
    Jiang, Ling
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2020, 19 (03)
  • [32] Translational relation embeddings for multi-hop knowledge base question answering
    Li, Ziyan
    Wang, Haofen
    Zhang, Wenqiang
    JOURNAL OF WEB SEMANTICS, 2022, 74
  • [33] An Integrated Method of Semantic Parsing and Information Retrieval for Knowledge Base Question Answering
    Zhen, Shiqi
    Yi, Xianwei
    Lin, Zhishu
    Xiao, Weiqi
    Su, Haibo
    Liu, Yijing
    CCKS 2021 - EVALUATION TRACK, 2022, 1553 : 44 - 51
  • [34] Knowledge Base Question Answering With a Matching-Aggregation Model and Question-Specific Contextual Relations
    Lan, Yunshi
    Wang, Shuohang
    Jiang, Jing
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2019, 27 (10) : 1629 - 1638
  • [35] Conversational Question Answering Over Knowledge Base using Chat-Bot Framework
    Sharath, Japa Sai
    Banafsheh, Rekabdar
    2021 IEEE 15TH INTERNATIONAL CONFERENCE ON SEMANTIC COMPUTING (ICSC 2021), 2021, : 84 - 85
  • [36] Multi-hop Knowledge Base Question Answering with an Iterative Sequence Matching Model
    Lan, Yunshi
    Wang, Shuohang
    Jiang, Jing
    2019 19TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2019), 2019, : 359 - 368
  • [37] A dynamic graph expansion network for multi-hop knowledge base question answering
    Wu, Wenqing
    Zhu, Zhenfang
    Qi, Jiangtao
    Wang, Wenling
    Zhang, Guangyuan
    Liu, Peiyu
    NEUROCOMPUTING, 2023, 515 : 37 - 47
  • [38] DAM: Transformer-based relation detection for Question Answering over Knowledge Base
    Chen, Yongrui
    Li, Huiying
    KNOWLEDGE-BASED SYSTEMS, 2020, 201
  • [39] Relation Extraction in Knowledge Base Question Answering: From General-Domain to the Catering Industry
    Chen, Hung-Chen
    Chen, Zi-Yuan
    Huang, Sin-Yi
    Ku, Lun-Wei
    Chiu, Yu-Shian
    Yang, Wei-Jen
    HCI IN BUSINESS, GOVERNMENT, AND ORGANIZATIONS, 2018, 10923 : 26 - 41
  • [40] A BERT-based Approach with Relation-aware Attention for Knowledge Base Question Answering
    Luo, Da
    Su, Jindian
    Yu, Shanshan
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,