Kiwifruit Planting Entity Recognition Based on Character and Word Information Fusion

被引:0
作者
Li S. [1 ]
Zhang M. [1 ]
Liu B. [1 ]
机构
[1] College of Information Engineering, Northwest A&F University, Shaanxi, Yangling
来源
Nongye Jixie Xuebao/Transactions of the Chinese Society for Agricultural Machinery | 2022年 / 53卷 / 12期
关键词
kiwifruit planting; named entity recognition; pre-trained language model; self-attention mechanisim; semantic enhancement; word fusion;
D O I
10.6041/j.issn.1000-1298.2022.12.032
中图分类号
学科分类号
摘要
Aiming at the problem of high complexity of real words and low recognition accuracy in the named entity recognition task of kiwifmit planting field, a entity recognition method of kiwifruit planting integrating character and word information was proposed. Based on BiGRU - CRF model, word level and character level information were fused. At the word level, by introducing word set information and using multiple self-attention mechanisms ( MHA) to adjust the weights of different words in the word set. At the same time, attention mechanism was used to ignore the unreliable word sets and focus on the important word sets to improve the entity recognition effect. At the character level, the unsupervised bidirectional encoder representations form transformers (BERT) pre-training model was introduced to enhance the semantic representation of words. Experiments were conducted on a homemade corpus in the kiwifruit cultivation domain containing 12 477 annotated samples and seven categories of entities, and the results showed that the Fl value of the model was improved by 1.58 percentage points compared with the SoftLexicon model. In addition, the experimental comparison of the model ResumeNER with Lattice - LSTM, WC - LSTM and other models in the open data set ResumeNER was carried out, and the best recognition effect was achieved. The Fl value reached 96. 17%, indicating that the method proposed had certain generalization ability. © 2022 Chinese Society of Agricultural Machinery. All rights reserved.
引用
收藏
页码:323 / 331
页数:8
相关论文
共 24 条
[21]  
MA R, PENG M, ZHANG Q, Et al., Simplify the usage of Lexicon in Chinese NER [C], Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, (2020)
[22]  
VASWANI A, SHAZEER N, PARMAR N, Et al., Attention is all you need [C], Proceedings of the 31st International Conference on Neural Information Processing Systems, pp. 5998-6008, (2017)
[23]  
HUANG Z, WEI X, KAI Y., Bidirectional LSTM - CRF models for sequence tagging
[24]  
WANG Ziniu, JIANG Meng, GAO Jianling, Et al., Chinese named entity recognition method based on BERT [J], Computer Science, 46, pp. 138-142, (2019)