Local or global? A novel transformer for Chinese named entity recognition based on multi-view and sliding attention

被引:4
|
作者
Wang, Yuke [1 ]
Lu, Ling [1 ]
Yang, Wu [1 ]
Chen, Yinong [2 ]
机构
[1] Chongqing Univ Technol, Sch Comp Sci & Engn, 69 Hongguang Ave, Chongqing, Peoples R China
[2] Arizona State Univ, Sch Comp & Augmented Intelligence, Tempe, AZ USA
关键词
Chinese NER; Human semantic understanding; Multimodal information fusion; Sliding attention; CONTEXT; CONSCIOUSNESS; INFORMATION;
D O I
10.1007/s13042-023-02023-0
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transformer is widely used in natural language processing (NLP) tasks due to the parallel and modeling of long texts. However, its performance in Chinese named entity recognition (NER) is not effective. While distance, direction, and information on global and local perspectives of sequence are all important for NER tasks, the traditional transformer structure only focus on distance and partial global information by fully connected self-attention mechanism. In this paper, we propose a multi-view and sliding attention (MVSA) model to enhance transformer's ability to model Chinese character-word features in NER task. MVSA combines directional information to extract character-word features from multiple views, proposes a weighted ternary fusion method for feature fusion and uses slider attention mechanisms to enhance the local representation ability of the model. Experiments on five Chinese NER datasets show that MVSA achieves superior performance than CNN-based, LSTM-based and traditional transformer-based models.
引用
收藏
页码:2199 / 2208
页数:10
相关论文
共 50 条
  • [1] MSFM: Multi-view Semantic Feature Fusion Model for Chinese Named Entity Recognition
    Liu, Jingxin
    Cheng, Jieren
    Peng, Xin
    Zhao, Zeli
    Tang, Xiangyan
    Sheng, Victor S.
    KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS, 2022, 16 (06): : 1833 - 1848
  • [2] Chinese named entity recognition based on adaptive transformer
    Yan Yang
    Yin, Guozhe
    2022 5TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND NATURAL LANGUAGE PROCESSING, MLNLP 2022, 2022, : 327 - 331
  • [3] Chinese named entity recognition based on Transformer encoder
    Guo X.-R.
    Luo P.
    Wang W.-L.
    Jilin Daxue Xuebao (Gongxueban)/Journal of Jilin University (Engineering and Technology Edition), 2021, 51 (03): : 989 - 995
  • [4] Joint multi-view character embedding model for named entity recognition of Chinese car reviews
    Jiaming Ding
    Wenping Xu
    Anning Wang
    Shuangyao Zhao
    Qiang Zhang
    Neural Computing and Applications, 2023, 35 : 14947 - 14962
  • [5] Joint multi-view character embedding model for named entity recognition of Chinese car reviews
    Ding, Jiaming
    Xu, Wenping
    Wang, Anning
    Zhao, Shuangyao
    Zhang, Qiang
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (20): : 14947 - 14962
  • [6] MULTI-VIEW SELF-ATTENTION BASED TRANSFORMER FOR SPEAKER RECOGNITION
    Wang, Rui
    Ao, Junyi
    Zhou, Long
    Liu, Shujie
    Wei, Zhihua
    Ko, Tom
    Li, Qing
    Zhang, Yu
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 6732 - 6736
  • [7] Chinese Medical Named Entity Recognition Based on Fusion of Global Features and Multi-Local Features
    Sun, Huarong
    Wang, Jianfeng
    Li, Bo
    Cao, Xiyuan
    Zang, Junbin
    Xue, Chenyang
    Zhang, Zhidong
    IEEE ACCESS, 2023, 11 : 137506 - 137520
  • [8] Multi-Feature Fusion Transformer for Chinese Named Entity Recognition
    Han, Xiaokai
    Yue, Qi
    Chu, Jing
    Han, Zhan
    Shi, Yifan
    Wang, Chengfeng
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 4227 - 4232
  • [9] Multi-Task Multi-Attention Transformer for Generative Named Entity Recognition
    Mo, Ying
    Liu, Jiahao
    Tang, Hongyin
    Wang, Qifan
    Xu, Zenglin
    Wang, Jingang
    Quan, Xiaojun
    Wu, Wei
    Li, Zhoujun
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 4171 - 4183
  • [10] Named Entity Recognition of Chinese Text Based on Attention Mechanism
    Shen, Tong-Ping
    Dumlao, Menchita
    Meng, Qing-Quan
    Zhan, Zhong-Hua
    Journal of Network Intelligence, 2023, 8 (02): : 505 - 518