An Improved Nested Named-Entity Recognition Model for Subject Recognition Task under Knowledge Base Question Answering

被引:2
作者
Wang, Ziming [1 ]
Xu, Xirong [1 ]
Li, Xinzi [1 ]
Li, Haochen [1 ]
Wei, Xiaopeng [1 ]
Huang, Degen [1 ]
机构
[1] Dalian Univ Technol, Sch Comp Sci & Technol, Dalian 116024, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2023年 / 13卷 / 20期
关键词
natural language processing; knowledge base question answering; subject recognition; named-entity recognition; deep learning; robustness;
D O I
10.3390/app132011249
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
In the subject recognition (SR) task under Knowledge Base Question Answering (KBQA), a common method is by training and employing a general flat Named-Entity Recognition (NER) model. However, it is not effective and robust enough in the case that the recognized entity could not be strictly matched to any subjects in the Knowledge Base (KB). Compared to flat NER models, nested NER models show more flexibility and robustness in general NER tasks, whereas it is difficult to employ a nested NER model directly in an SR task. In this paper, we take advantage of features of a nested NER model and propose an Improved Nested NER Model (INNM) for the SR task under KBQA. In our model, each question token is labeled as either an entity token, a start token, or an end token by a modified nested NER model based on semantics. Then, entity candidates would be generated based on such labels, and an approximate matching strategy is employed to score all subjects in the KB based on string similarity to find the best-matched subject. Experimental results show that our model is effective and robust to both single-relation questions and complex questions, which outperforms the baseline flat NER model by a margin of 3.3% accuracy on the SimpleQuestions dataset and a margin of 11.0% accuracy on the WebQuestionsSP dataset.
引用
收藏
页数:13
相关论文
共 43 条
  • [1] Automatic question generation based on sentence structure analysis using machine learning approach
    Blstak, Miroslav
    Rozinajova, Viera
    [J]. NATURAL LANGUAGE ENGINEERING, 2022, 28 (04) : 487 - 517
  • [2] Bollacker Kurt., 2008, P 2008 ACM SIGMOD IN, P1247, DOI 10.1145/1376616.1376746
  • [3] Bordes A, 2015, Arxiv, DOI arXiv:1506.02075
  • [4] Chuang YS, 2020, Arxiv, DOI arXiv:1910.11559
  • [5] ELECTRA: PRE-TRAINING TEXT ENCODERS AS DISCRIMINATORS RATHER THAN GENERATORS
    Clark, Kevin
    Luong, Minh-Thang
    Le, Quoc V.
    Manning, Christopher D.
    [J]. INFORMATION SYSTEMS RESEARCH, 2020,
  • [6] Dai ZH, 2016, PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, P800
  • [7] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [8] Gangwar A., 2021, Proceedings of the 15th International Workshop on Semantic Evaluation, P1232
  • [9] Gu Y., 2022, arXiv
  • [10] Nguyen H, 2022, PROCEEDINGS OF THE 5TH WORKSHOP ON E-COMMERCE AND NLP (ECNLP 5), P171