CMiNER: Named entity recognition on imperfectly annotated data via confidence and meta weight adaptation

被引:0
作者
Wei, Shangfei
Lai, Zhaohong
Shi, Xiaodong [1 ]
机构
[1] Xiamen Univ, Sch Informat, Xiamen 36100, Fujian, Peoples R China
关键词
NER; Meta-learning; Confident-learning; Noisy-tolerance;
D O I
10.1016/j.eswa.2025.126987
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Named Entity Recognition(NER) is a very critical task in natural language processing. The task often suffers from noisy labels that affect the robustness of the model. In particular, when there are a large number of both mislabeled and missing labeled labels in the dataset, the performance of the model is usually severely degraded. Through empirical studies of multiple NER models and datasets, we find that the degree of degradation increases with the noise rate, which poses a challenge to robustly training models on imperfectly annotated datasets. To alleviate the problem, this study designs a confidence and meta-weight based adaptive NER training framework (CMiNER). We design entity-level confidence estimators that not only initially assess the quality of a given noisy dataset, but also greatly assist in the training phase. Specifically, we design a meta-weight adaptive strategy that incorporates confidence partitioning for robust training of noisy datasets. We partition the data into support and query sets using a confidence estimator and employ a meta-weight adaptation strategy for training, ultimately enhancing the model performance. Experiments on real NER datasets and synthetic noisy datasets show that the CMiNER framework is extremely robust. The performance after introducing our framework on the NER datasets outperforms other noise reduction methods. The performance is improved by an average of 10% on low-noise synthetic datasets and 3.1% on high-noise synthetic datasets. In addition, there is a 7.6% improvement over the real NER dataset.
引用
收藏
页数:13
相关论文
共 76 条
  • [1] Alt C, 2020, 58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), P1558
  • [2] Amiri H., 2018, P 2018 C N AM CHAPT, V1, P2006, DOI DOI 10.18653/V1/N18-1182
  • [3] Barnes J, 2019, BLACKBOXNLP WORKSHOP ON ANALYZING AND INTERPRETING NEURAL NETWORKS FOR NLP AT ACL 2019, P12
  • [4] Boyd A., 2008, Research on Language and Computation, V6, P113
  • [5] Randomly Wired Graph Neural Network for Chinese NER
    Chen, Jie
    Xi, Xuefeng
    Sheng, Victor S.
    Cui, Zhiming
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2023, 227
  • [6] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [7] Dickinson M, 2003, EACL 2003: 10TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, PROCEEDINGS OF THE CONFERENCE, P107
  • [8] Ding ZJ, 2024, PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, P6252
  • [9] Elkan Charles, 2008, P 14 ACM SIGKDD INT, P213, DOI [10.1145/1401890.1401920, DOI 10.1145/1401890.1401920]
  • [10] Finn C, 2017, PR MACH LEARN RES, V70