A multi-head adjacent attention-based pyramid layered model for nested named entity recognition

被引:4
作者
Cui, Shengmin [1 ]
Joe, Inwhee [1 ]
机构
[1] Hanyang Univ, Dept Comp Sci, 222 Wangsimni Ro, Seoul 04763, South Korea
关键词
Nested named entity recognition; Named entity recognition; Attention; Pyramid; Natural language processing; EXTRACTION;
D O I
10.1007/s00521-022-07747-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Named entity recognition (NER) is one of the widely studied natural language processing tasks in recent years. Conventional solutions treat the NER as a sequence labeling problem, but these approaches cannot handle nested NER. This is due to the fact that nested NER refers to the case where one entity contains another entity and it is not feasible to tag each token with a single tag. The pyramid model stacks L flat NER layers for prediction, which subtly enumerates all spans with length less than or equal to L. However, the original model introduces a block consisting of a convolutional layer and a bidirectional long short-term memory (Bi-LSTM) layer as the decoder, which does not consider the dependency between adjacent inputs and the Bi-LSTM cannot perform parallel computation on sequential inputs. For the purpose of improving performance and reducing the forward computation, we propose a Multi-Head Adjacent Attention-based Pyramid Layered model. In addition, when constructing a pyramid structure for span representation, the information of the intermediate words has more proportion than words on the two sides. To address this imbalance in the span representation, we fuse the output of the attention layer with the features of head and tail words when doing classification. We conducted experiments on nested NER datasets such as GENIA, SciERC, and ADE to validate the effectiveness of our proposed model.
引用
收藏
页码:2561 / 2574
页数:14
相关论文
共 50 条
  • [1] A multi-head adjacent attention-based pyramid layered model for nested named entity recognition
    Shengmin Cui
    Inwhee Joe
    Neural Computing and Applications, 2023, 35 : 2561 - 2574
  • [2] Fast Neural Chinese Named Entity Recognition with Multi-head Self-attention
    Qi, Tao
    Wu, Chuhan
    Wu, Fangzhao
    Ge, Suyu
    Liu, Junxin
    Huang, Yongfeng
    Xie, Xing
    KNOWLEDGE GRAPH AND SEMANTIC COMPUTING: KNOWLEDGE COMPUTING AND LANGUAGE UNDERSTANDING, 2019, 1134 : 98 - 110
  • [3] Adversarial Transfer Learning for Named Entity Recognition Based on Multi-Head Attention Mechanism and Feature Fusion
    Zhao, Dandan
    Zhang, Pan
    Meng, Jiana
    Wu, Yue
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT I, 2022, 13551 : 272 - 284
  • [4] Attention-based interactive multi-level feature fusion for named entity recognition
    Xu, Yiwu
    Chen, Yun
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [5] A Multi-Agent Communication Based Model for Nested Named Entity Recognition
    Li, Canguang
    Wang, Guohua
    Cao, Jin
    Cai, Yi
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2021, 29 (29) : 2123 - 2136
  • [6] An Attention-Based BiLSTM-CRF Model for Chinese Clinic Named Entity Recognition
    Wu, Guohua
    Tang, Guangen
    Wang, Zhongru
    Zhang, Zhen
    Wang, Zhen
    IEEE ACCESS, 2019, 7 (113942-113949) : 113942 - 113949
  • [7] Nested Named Entity Recognition via an Independent-Layered Pretrained Model
    Jia, Liruizhi
    Liu, Shengquan
    Wei, Fuyuan
    Kong, Bo
    Wang, Guangyao
    IEEE ACCESS, 2021, 9 : 109693 - 109703
  • [8] An attention-based multi-task model for named entity recognition and intent analysis of Chinese online medical questions
    Wu, Chaochen
    Luo, Guan
    Guo, Chao
    Ren, Yin
    Zheng, Anni
    Yang, Cheng
    JOURNAL OF BIOMEDICAL INFORMATICS, 2020, 108 (108)
  • [9] An Attention-Based Approach for Mongolian News Named Entity Recognition
    Tan, Mingyan
    Bao, Feilong
    Gao, Guanglai
    Wang, Weihua
    CHINESE COMPUTATIONAL LINGUISTICS, CCL 2019, 2019, 11856 : 424 - 435
  • [10] An Attention-Based Approach for Chemical Compound and Drug Named Entity Recognition
    Yang P.
    Yang Z.
    Luo L.
    Lin H.
    Wang J.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2018, 55 (07): : 1548 - 1556