A multi-head adjacent attention-based pyramid layered model for nested named entity recognition

被引:4
|
作者
Cui, Shengmin [1 ]
Joe, Inwhee [1 ]
机构
[1] Hanyang Univ, Dept Comp Sci, 222 Wangsimni Ro, Seoul 04763, South Korea
来源
NEURAL COMPUTING & APPLICATIONS | 2023年 / 35卷 / 03期
关键词
Nested named entity recognition; Named entity recognition; Attention; Pyramid; Natural language processing; EXTRACTION;
D O I
10.1007/s00521-022-07747-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Named entity recognition (NER) is one of the widely studied natural language processing tasks in recent years. Conventional solutions treat the NER as a sequence labeling problem, but these approaches cannot handle nested NER. This is due to the fact that nested NER refers to the case where one entity contains another entity and it is not feasible to tag each token with a single tag. The pyramid model stacks L flat NER layers for prediction, which subtly enumerates all spans with length less than or equal to L. However, the original model introduces a block consisting of a convolutional layer and a bidirectional long short-term memory (Bi-LSTM) layer as the decoder, which does not consider the dependency between adjacent inputs and the Bi-LSTM cannot perform parallel computation on sequential inputs. For the purpose of improving performance and reducing the forward computation, we propose a Multi-Head Adjacent Attention-based Pyramid Layered model. In addition, when constructing a pyramid structure for span representation, the information of the intermediate words has more proportion than words on the two sides. To address this imbalance in the span representation, we fuse the output of the attention layer with the features of head and tail words when doing classification. We conducted experiments on nested NER datasets such as GENIA, SciERC, and ADE to validate the effectiveness of our proposed model.
引用
收藏
页码:2561 / 2574
页数:14
相关论文
共 50 条
  • [21] Attention-Based Bi-LSTM for Chinese Named Entity Recognition
    Zhang, Kai
    Ren, Weiping
    Zhang, Yangsen
    CHINESE LEXICAL SEMANTICS, CLSW 2018, 2018, 11173 : 643 - 652
  • [22] Nested Named Entity Recognition via an Independent-Layered Pretrained Model
    Jia, Liruizhi
    Liu, Shengquan
    Wei, Fuyuan
    Kong, Bo
    Wang, Guangyao
    IEEE ACCESS, 2021, 9 : 109693 - 109703
  • [23] A Nested Named Entity Recognition Model Based on Multi-agent Communication Mechanism
    Li, Canguang
    Wang, Guohua
    Cao, Jin
    Cai, Yi
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 15823 - 15824
  • [24] An attention-based multi-task model for named entity recognition and intent analysis of Chinese online medical questions
    Wu, Chaochen
    Luo, Guan
    Guo, Chao
    Ren, Yin
    Zheng, Anni
    Yang, Cheng
    JOURNAL OF BIOMEDICAL INFORMATICS, 2020, 108 (108)
  • [25] Attention-Based End-to-End Named Entity Recognition from Speech
    Porjazovski, Dejan
    Leinonen, Juho
    Kurimo, Mikko
    TEXT, SPEECH, AND DIALOGUE, TSD 2021, 2021, 12848 : 469 - 480
  • [26] Multi-head attention-based model for reconstructing continuous missing time series data
    Wu, Huafeng
    Zhang, Yuxuan
    Liang, Linian
    Mei, Xiaojun
    Han, Dezhi
    Han, Bing
    Weng, Tien-Hsiung
    Li, Kuan-Ching
    JOURNAL OF SUPERCOMPUTING, 2023, 79 (18): : 20684 - 20711
  • [27] Multi-head Attention-Based Masked Sequence Model for Mapping Functional Brain Networks
    He, Mengshen
    Hou, Xiangyu
    Wang, Zhenwei
    Kang, Zili
    Zhang, Xin
    Qiang, Ning
    Ge, Bao
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT I, 2022, 13431 : 295 - 304
  • [28] Multi-head attention-based model for reconstructing continuous missing time series data
    Huafeng Wu
    Yuxuan Zhang
    Linian Liang
    Xiaojun Mei
    Dezhi Han
    Bing Han
    Tien-Hsiung Weng
    Kuan-Ching Li
    The Journal of Supercomputing, 2023, 79 : 20684 - 20711
  • [29] Recursive label attention network for nested named entity recognition
    Kim, Hongjin
    Kim, Harksoo
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 249
  • [30] Multi-head attention-based masked sequence model for mapping functional brain networks
    He, Mengshen
    Hou, Xiangyu
    Ge, Enjie
    Wang, Zhenwei
    Kang, Zili
    Qiang, Ning
    Zhang, Xin
    Ge, Bao
    FRONTIERS IN NEUROSCIENCE, 2023, 17