Improving unified named entity recognition by incorporating mention relevance

被引:4
作者
Ji, Lijun [1 ]
Yan, Danfeng [1 ]
Cheng, Zhuoran [1 ]
Song, Yan [2 ]
机构
[1] Beijing Univ Posts & Telecommun, State Key Lab Networking & Switching Technol, Beijing 100876, Peoples R China
[2] Shanghai Int Studies Univ, Sch Business & Management, Shanghai 200092, Peoples R China
关键词
Named entity recognition; Sequence-to-sequence model; Machine reading comprehension; Mention relevance attention;
D O I
10.1007/s00521-023-08820-6
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Named entity recognition (NER) is a fundamental task for natural language processing, which aims to detect mentions of real-world entities from text and classifying them into predefined types. Recently, research on overlapped and discontinuous named entity recognition has received increasing attention. However, we note that few studies have considered both overlapped and discontinuous entities. In this paper, we proposed a novel sequence-to-sequence model that is capable of recognizing both overlapped and discontinuous entities based on machine reading comprehension. The model utilizes machine reading comprehension formulation to encode significant inferior information about the entity category. Then input sequence passes through a question-answering model to predict the mention relevance of the given source sentences to the query. Finally, we incorporate the mention relevance into the BART-based generation model. We conducted experiments on three type of NER datasets to show the generality of our model. The experimental results demonstrate that our model beats almost all the current top-performing baselines achieves a vast amount of performance boost over current SOTA models on overlapped and discontinuous NER datasets.
引用
收藏
页码:22223 / 22234
页数:12
相关论文
共 39 条
[1]  
Alex B., 2007, TRANSLATIONAL CLIN L, P65
[2]  
[Anonymous], 2015, 3 INT C LEARNING REP
[3]  
Chen X., 2022, P 29 INT C COMP LING, P2374, DOI DOI 10.48550/ARXIV.2109.00720
[4]  
Chiu Jason PC, 2016, Transactions of the Association for Computational Linguistics, V4, P357, DOI DOI 10.1162/TACLA00104
[5]  
Cuiy LY, 2021, FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, P1835
[6]  
Dai Xiang, 2020, P 58 ANN M ASS COMPU, DOI [DOI 10.18653/V1/2020.ACL-MAIN.520, 10.18653/V1/2020.ACL-MAIN.520]
[7]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[8]  
Fei H, 2021, AAAI CONF ARTIF INTE, V35, P12785
[9]  
Gillick Dan, 2016, P 2016 C N AM CHAPT, P1296, DOI DOI 10.18653/V1/N16-1155
[10]  
Gu J, INT C LEARNING REPRE