Enhancing SPARQL Query Generation for Knowledge Base Question Answering Systems by Learning to Correct Triplets

被引:0
作者
Qi, Jiexing [1 ]
Su, Chang [1 ]
Guo, Zhixin [1 ]
Wu, Lyuwen [1 ]
Shen, Zanwei [1 ]
Fu, Luoyi [1 ]
Wang, Xinbing [1 ]
Zhou, Chenghu [1 ,2 ]
机构
[1] Shanghai Jiao Tong Univ, Sch Elect Informat & Elect Engn, Shanghai 200240, Peoples R China
[2] Chinese Acad Sci, Inst Geog Sci & Nat Resources Res, Beijing 100101, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 04期
关键词
Knowledge Base Question Answering; Text-to-SPARQL; semantic parsing; further pretraining; Triplet Structure;
D O I
10.3390/app14041521
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Generating SPARQL queries from natural language questions is challenging in Knowledge Base Question Answering (KBQA) systems. The current state-of-the-art models heavily rely on fine-tuning pretrained models such as T5. However, these methods still encounter critical issues such as triple-flip errors (e.g., (subject, relation, object) is predicted as (object, relation, subject)). To address this limitation, we introduce TSET (Triplet Structure Enhanced T5), a model with a novel pretraining stage positioned between the initial T5 pretraining and the fine-tuning for the Text-to-SPARQL task. In this intermediary stage, we introduce a new objective called Triplet Structure Correction (TSC) to train the model on a SPARQL corpus derived from Wikidata. This objective aims to deepen the model's understanding of the order of triplets. After this specialized pretraining, the model undergoes fine-tuning for SPARQL query generation, augmenting its query-generation capabilities. We also propose a method named "semantic transformation" to fortify the model's grasp of SPARQL syntax and semantics without compromising the pre-trained weights of T5. Experimental results demonstrate that our proposed TSET outperforms existing methods on three well-established KBQA datasets: LC-QuAD 2.0, QALD-9 plus, and QALD-10, establishing a new state-of-the-art performance (95.0% F1 and 93.1% QM on LC-QuAD 2.0, 75.85% F1 and 61.76% QM on QALD-9 plus, 51.37% F1 and 40.05% QM on QALD-10).
引用
收藏
页数:19
相关论文
共 50 条
  • [21] Question Answering over Knowledge Base with Variational Auto-Encoder
    Sharath, Japa Sai
    Sarah, Green
    2022 IEEE EIGHTH INTERNATIONAL CONFERENCE ON MULTIMEDIA BIG DATA (BIGMM 2022), 2022, : 29 - 36
  • [22] Robust cross-lingual knowledge base question answering via knowledge distillation
    Wang, Shaofei
    Dang, Depeng
    DATA TECHNOLOGIES AND APPLICATIONS, 2021, 55 (05) : 661 - 681
  • [23] Knowing What I Don't Know: A Generation Assisted Rejection Framework in Knowledge Base Question Answering
    Huang, Junyang
    Lu, Xuantao
    Liang, Jiaqing
    Bao, Qiaoben
    Huang, Chen
    Xiao, Yanghua
    Liu, Bang
    Chen, Yunwen
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, DASFAA 2022, PT III, 2022, : 238 - 246
  • [24] Relation Extraction as Text Matching: A Scheme for Multi-hop Knowledge Base Question Answering
    Li, Ziyan
    Ni, Kan
    Wang, Haofen
    Zhang, Wenqiang
    CCKS 2022 - EVALUATION TRACK, 2022, 1711 : 191 - 201
  • [25] Complex Knowledge Base Question Answering via Structure and Content Dual-Driven Method
    Zhang, Jinhao
    Zhang, Lizong
    Zhang, Jinchuan
    Xin, Yichen
    Zheng, Xu
    WEB AND BIG DATA, APWEB-WAIM 2024, PT II, 2024, 14962 : 243 - 255
  • [26] A contrastive framework for enhancing Knowledge Graph Question Answering: Alleviating exposure bias
    Du, Huifang
    Zhang, Xixie
    Wang, Meng
    Chen, Yunwen
    Ji, Daqi
    Ma, Jun
    Wang, Haofen
    KNOWLEDGE-BASED SYSTEMS, 2023, 280
  • [27] BT-CKBQA: An efficient approach for Chinese knowledge base question answering
    Yang, Erhe
    Hao, Fei
    Shang, Jiaxing
    Chen, Xiaoliang
    Park, Doo-Soon
    DATA & KNOWLEDGE ENGINEERING, 2023, 147
  • [28] A Dynamic Graph Reasoning Model with an Auxiliary Task for Knowledge Base Question Answering
    Wu, Zhichao
    Tian, Xuan
    ELECTRONICS, 2024, 13 (24):
  • [29] SPaReL: A Semantic Parsing Relation Linking Method for Knowledge Base Question Answering
    Dai, Xinbang
    Li, Huiying
    Wang, Tenggou
    Li, Lei
    Feng, Yuxi
    Li, Xin
    PROCEEDINGS OF THE 11TH INTERNATIONAL JOINT CONFERENCE ON KNOWLEDGE GRAPHS, IJCKG 2022, 2022, : 73 - 81
  • [30] Improving Core Path Reasoning for the Weakly Supervised Knowledge Base Question Answering
    Hu, Nan
    Bi, Sheng
    Qi, Guilin
    Wang, Meng
    Hua, Yuncheng
    Shen, Shirong
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, DASFAA 2022, PT I, 2022, : 162 - 170