Chinese Named Entity Recognition Based on BERT and Lightweight Feature Extraction Model

被引:8
作者
Yang, Ruisen [1 ]
Gan, Yong [2 ]
Zhang, Chenfang [1 ]
机构
[1] Zhengzhou Univ Light Ind, Sch Comp Commun & Engn, Zhengzhou 450000, Peoples R China
[2] Zhengzhou Inst Engn & Technol, Sch Comp Commun & Engn, Zhengzhou 450000, Peoples R China
关键词
named entity recognition; deep learning; neural network; BERT;
D O I
10.3390/info13110515
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the early named entity recognition models, most text processing focused only on the representation of individual words and character vectors, and paid little attention to the semantic relationships between the preceding and following text in an utterance, which led to the inability to handle the problem of multiple meanings of a word during recognition. To address this problem, most models introduce the attention mechanism of Transformer model to solve the problem of multiple meanings of a word in text. However, the traditional Transformer model leads to a high computational overhead due to its fully connected structure. Therefore, this paper proposes a new model, the BERT-Star-Transformer-CNN-BiLSTM-CRF model, to solve the problem of the computational efficiency of the traditional Transformer. First, the input text is dynamically generated into a character vector using the BERT model pre-trained in large-scale preconditioning to solve the problem of multiple meanings of words, and then the lightweight Star-Transformer model is used as the feature extraction module to perform local feature extraction on the word vector sequence, while the CNN-BiLSTM joint model is used to perform global feature extraction on the context in the text. The obtained feature sequences are fused. Finally, the fused feature vector sequences are input to CRF for prediction of the final results. After the experiments, it is shown that the model has a significant improvement in precision, recall and Fl value compared with the traditional model, and the computational efficiency is improved by nearly 40%.
引用
收藏
页数:21
相关论文
共 39 条
[1]  
Agarwal M., 2018, P 2 INT WORKSHOP SEN
[2]  
[Anonymous], 2016, T ASSOC COMPUT LING, DOI DOI 10.1162/TACLA00104
[3]  
[Anonymous], 2008, P 25 INT C MACH LEAR, DOI DOI 10.1145/1390156.1390177
[4]  
Cao PF, 2018, 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), P182
[5]  
Chen H, 2019, AAAI CONF ARTIF INTE, P6236
[6]  
Chinchor N. A., 1995, Empirical Methods in Discourse Interpretation and Generation. Papers from the 1995 AAAI Symposium (TR SS-95-06), P21
[7]  
Collobert R, 2011, J MACH LEARN RES, V12, P2493
[8]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[9]   Character-Based LSTM-CRF with Radical-Level Features for Chinese Named Entity Recognition [J].
Dong, Chuanhai ;
Zhang, Jiajun ;
Zong, Chengqing ;
Hattori, Masanori ;
Di, Hui .
NATURAL LANGUAGE UNDERSTANDING AND INTELLIGENT APPLICATIONS (NLPCC 2016), 2016, 10102 :239-250
[10]  
Dyer C, 2015, Arxiv, DOI arXiv:1505.08075