Chinese named entity recognition based on adaptive transformer

被引:0
作者
Yan Yang [1 ]
Yin, Guozhe [1 ]
机构
[1] Heilongjiang Univ, Harbin, Peoples R China
来源
2022 5TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND NATURAL LANGUAGE PROCESSING, MLNLP 2022 | 2022年
关键词
Vocabulary enhancement; Chinese named entity recognition; caracter encoding; Transformer;
D O I
10.1145/3578741.3578804
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, lexical enhancement methods have been the focus of research in NER tasks. In Chinese NER research,the SoftLexion model has achieved the best results on several Chinese datasets using lexical augmentation methods. In this paper,SoftLexion is investigated, and it is found that the model does not encode the input text characters but directly uses the corresponding embedding vectors found from the pre-trained character embedding matrix. Secondly, Bi-LSTM is used in the sequence modeling layer,but this method is inferior to Transformer in terms of contextual feature extraction capability, long-range feature capture capability, and parallel computing capability, so the research work in this paper adopts the latest Adaptive Transformer to encode the input text of the model and then fuse it with lexical information. The character representation of the fused lexical information is then sequence modeled by the adaptive Transformer and finally decoded by the tag decoding layer. In this paper, experiments are conducted on three Chinese datasets and the study shows that the model performs better with the addition of a character encoding layer and sequence modeling by Adaptive Transformer.
引用
收藏
页码:327 / 331
页数:5
相关论文
共 16 条
[1]  
Devlin J, 2019, Arxiv, DOI [arXiv:1810.04805, 10.48550/arXiv.1810.04805]
[2]  
Ding RX, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P1462
[3]  
Gui T, 2019, PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P4982
[4]  
Huang ZH, 2015, Arxiv, DOI [arXiv:1508.01991, 10.48550/arXiv.1508.01991, DOI 10.48550/ARXIV.1508.01991]
[5]  
King DB, 2015, ACS SYM SER, V1214, P1, DOI 10.1021/bk-2015-1214.ch001
[6]  
Lafferty J.D., 2001, ICML, P282
[7]  
Li X., 2020, P 58 ANN M ASS COMPU, P6836, DOI DOI 10.18653/V1/2020.ACL-MAIN.611
[8]  
Liu W, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P2379
[9]  
Liu ZX, 2010, LECT NOTES ARTIF INT, V6216, P634
[10]   VITERBI ALGORITHM AS AN AID IN TEXT RECOGNITION [J].
NEUHOFF, DL .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1975, 21 (02) :222-226