Deep Neural Network with Embedding Fusion for Chinese Named Entity Recognition

被引:2
作者
Long, Kaifang [1 ]
Zhao, Han [1 ]
Shao, Zengzhen [1 ,2 ]
Cao, Yang [1 ]
Geng, Yanfang [1 ]
Sun, Yintai [1 ]
Xu, Weizhi [3 ,4 ]
Yu, Hui [5 ]
机构
[1] Shandong Normal Univ, Sch Informat Sci & Engn, 1 Univ Rd, Jinan, Peoples R China
[2] Shandong Womens Univ, Inst Data Sci & Comp, 2399 Univ Rd, Jinan, Peoples R China
[3] Shandong Normal Univ, Sch Informat Sci & Engn, 2399 Univ Rd, Jinan, Peoples R China
[4] State Key Lab High End Server & Storage Technol, Jinan, Peoples R China
[5] Shandong Normal Univ, Sch Business, 1 Univ Rd, Jinan, Peoples R China
基金
中国国家自然科学基金;
关键词
Chinese Named Entity Recognition; self-attention mechanism; natural language processing; deep neural network; MODEL;
D O I
10.1145/3570328
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Chinese Named Entity Recognition (NER) is an essential task in natural language processing, and its performance directly impacts the downstream tasks. The main challenges in Chinese NER are the high dependence of named entities on context and the lack of word boundary information. Therefore, how to integrate relevant knowledge into the corresponding entity has become the primary task for Chinese NER. Both the lattice LSTM model and the WC-LSTM model did not make excellent use of contextual information. Additionally, the lattice LSTM model had a complex structure and did not exploit the word information well. To address the preceding problems, we propose a Chinese NERmethod based on the deep neural network with multiple ways of embedding fusion. First, we use a convolutional neural network to combine the contextual information of the input sequence and apply a self-attention mechanism to integrate lexicon knowledge, compensating for the lack of word boundaries. The word feature, context feature, bigram feature, and bigram context feature are obtained for each character. Second, four different features are used to fuse information at the embedding layer. As a result, four different word embeddings are obtained through cascading. Last, the fused feature information is input to the encoding and decoding layer. Experiments on three datasets show that our model can effectively improve the performance of Chinese NER.
引用
收藏
页数:16
相关论文
共 51 条
  • [1] Cao PF, 2018, 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), P182
  • [2] Chen X, 2021, Arxiv, DOI arXiv:2109.00720
  • [3] Chen Xinchi, 2015, C EMP METH NAT LANG, P1197
  • [4] Chen YB, 2015, PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1, P167
  • [5] Chiu J.P.C., 2016, T ASS COMPUTATIONAL, V4, P357, DOI [DOI 10.1162/TACLA00104, 10.1162/tacl_a_00104, DOI 10.1162/TACL_A_00104]
  • [6] Cuiy LY, 2021, FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, P1835
  • [7] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [8] Core techniques of question answering systems over knowledge bases: a survey
    Diefenbach, Dennis
    Lopez, Vanessa
    Singh, Kamal
    Maret, Pierre
    [J]. KNOWLEDGE AND INFORMATION SYSTEMS, 2018, 55 (03) : 529 - 569
  • [9] Ding RX, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P1462
  • [10] Character-Based LSTM-CRF with Radical-Level Features for Chinese Named Entity Recognition
    Dong, Chuanhai
    Zhang, Jiajun
    Zong, Chengqing
    Hattori, Masanori
    Di, Hui
    [J]. NATURAL LANGUAGE UNDERSTANDING AND INTELLIGENT APPLICATIONS (NLPCC 2016), 2016, 10102 : 239 - 250