Local or global? A novel transformer for Chinese named entity recognition based on multi-view and sliding attention

被引:4
|
作者
Wang, Yuke [1 ]
Lu, Ling [1 ]
Yang, Wu [1 ]
Chen, Yinong [2 ]
机构
[1] Chongqing Univ Technol, Sch Comp Sci & Engn, 69 Hongguang Ave, Chongqing, Peoples R China
[2] Arizona State Univ, Sch Comp & Augmented Intelligence, Tempe, AZ USA
关键词
Chinese NER; Human semantic understanding; Multimodal information fusion; Sliding attention; CONTEXT; CONSCIOUSNESS; INFORMATION;
D O I
10.1007/s13042-023-02023-0
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transformer is widely used in natural language processing (NLP) tasks due to the parallel and modeling of long texts. However, its performance in Chinese named entity recognition (NER) is not effective. While distance, direction, and information on global and local perspectives of sequence are all important for NER tasks, the traditional transformer structure only focus on distance and partial global information by fully connected self-attention mechanism. In this paper, we propose a multi-view and sliding attention (MVSA) model to enhance transformer's ability to model Chinese character-word features in NER task. MVSA combines directional information to extract character-word features from multiple views, proposes a weighted ternary fusion method for feature fusion and uses slider attention mechanisms to enhance the local representation ability of the model. Experiments on five Chinese NER datasets show that MVSA achieves superior performance than CNN-based, LSTM-based and traditional transformer-based models.
引用
收藏
页码:2199 / 2208
页数:10
相关论文
共 50 条
  • [41] A self-attention based neural architecture for Chinese medical named entity recognition
    Wan, Qian
    Liu, Jie
    Wei, Luona
    Ji, Bin
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2020, 17 (04) : 3498 - 3511
  • [42] MATE: Multi-view Attention for Table Transformer Efficiency
    Eisenschlos, Julian Martin
    Gor, Maharshi
    Mueller, Thomas
    Cohen, William W.
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 7606 - 7619
  • [43] Think More Ambiguity Less: A Novel Dual Interactive Model with Local and Global Semantics for Chinese Named Entity Recognition
    Jia, Yue
    Fang, Wei
    Lu, Heng-Yang
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (06)
  • [44] Chinese Named Entity Recognition Model Based on Multi-Task Learning
    Fang, Qin
    Li, Yane
    Feng, Hailin
    Ruan, Yaoping
    APPLIED SCIENCES-BASEL, 2023, 13 (08):
  • [45] Chinese Named Entity Recognition Based on Multi-Level Representation Learning
    Li, Weijun
    Ding, Jianping
    Liu, Shixia
    Liu, Xueyang
    Su, Yilei
    Wang, Ziyi
    APPLIED SCIENCES-BASEL, 2024, 14 (19):
  • [46] AERNs: Attention-Based Entity Region Networks for Multi-Grained Named Entity Recognition
    Dai, Jianghai
    Feng, Chong
    Bai, Xuefeng
    Dai, Jinming
    Zhang, Huanhuan
    2019 IEEE 31ST INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2019), 2019, : 408 - 415
  • [47] Incorporating word⁃set attention into Chinese named entity recognition Method
    Zhong S.-S.
    Chen X.
    Zhao M.-H.
    Zhang Y.-J.
    Jilin Daxue Xuebao (Gongxueban)/Journal of Jilin University (Engineering and Technology Edition), 2022, 52 (05): : 1098 - 1105
  • [48] Chinese Named Entity Recognition Method Combining ALBERT and a Local Adversarial Training and Adding Attention Mechanism
    Zhang Runmei
    Li Lulu
    Yin Lei
    Liu Jingjing
    Xu Weiyi
    Cao Weiwei
    Chen Zhong
    INTERNATIONAL JOURNAL ON SEMANTIC WEB AND INFORMATION SYSTEMS, 2022, 18 (01)
  • [49] Medical Named Entity Recognition Based on Multi-Feature and Co-Attention
    Xinning, L.I.U.
    Computer Engineering and Applications, 2024, 60 (06) : 188 - 198
  • [50] Attention-based Multi-level Feature Fusion for Named Entity Recognition
    Yang, Zhiwei
    Chen, Hechang
    Zhang, Jiawei
    Ma, Jing
    Chang, Yi
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 3594 - 3600