PL-Transformer: a POS-aware and layer ensemble transformer for text classification

被引:3
|
作者
Shi, Yu [1 ]
Zhang, Xi [1 ]
Yu, Ning [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Key Lab Trustworthy Distributed Comp & Serv BUPT, Minist Educ, Beijing, Peoples R China
关键词
Text classification; Transformer; Part-of-speech; Layer ensemble;
D O I
10.1007/s00521-022-07872-4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The transformer-based models have become the de-facto standard for natural language processing (NLP) tasks. However, most of these models are only designed to capture the implicit semantics among tokens without considering the extra off-the-shelf knowledge (e.g., parts-of-speech) to facilitate the NLP tasks. Additionally, despite using multiple attention-based encoders, they only utilize the embeddings from the last layer, ignoring that from other layers. To address these issues, in this paper, we propose a novel POS-aware and layer ensemble transformer neural network (named as PL-Transformer). PL-Transformer utilizes the parts-of-speech information explicitly and leverages the outputs from different encoder layers with correlation coefficient attention (C-Encoder) jointly. Moreover, we use correlation coefficient attention to bound dot product in C-Encoder, which improves the overall model performance. Extensive experiments on four datasets demonstrate that PL-Transformer can improve the text classification performance. For example, the accuracy on the MPQA dataset is improved by 3.95% over the vanilla transformer.
引用
收藏
页码:1971 / 1982
页数:12
相关论文
共 44 条
  • [1] PL-Transformer: a POS-aware and layer ensemble transformer for text classification
    Yu Shi
    Xi Zhang
    Ning Yu
    Neural Computing and Applications, 2023, 35 : 1971 - 1982
  • [2] Transformer and Graph Convolutional Network for Text Classification
    Liu, Boting
    Guan, Weili
    Yang, Changjin
    Fang, Zhijie
    Lu, Zhiheng
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE SYSTEMS, 2023, 16 (01)
  • [3] Transformer and Graph Convolutional Network for Text Classification
    Boting Liu
    Weili Guan
    Changjin Yang
    Zhijie Fang
    Zhiheng Lu
    International Journal of Computational Intelligence Systems, 16
  • [4] Interactive POS-aware network for aspect-level sentiment classification
    Shuang, Kai
    Gu, Mengyu
    Li, Rui
    Loo, Jonathan
    Su, Sen
    NEUROCOMPUTING, 2021, 420 : 181 - 196
  • [5] A Comprehensive Verification of Transformer in Text Classification
    Yang, Xiuyuan
    Yang, Liang
    Bi, Ran
    Lin, Hongfei
    CHINESE COMPUTATIONAL LINGUISTICS, CCL 2019, 2019, 11856 : 207 - 218
  • [6] Text classification using improved bidirectional transformer
    Tezgider, Murat
    Yildiz, Beytullah
    Aydin, Galip
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2022, 34 (09)
  • [7] Learning to combine classifiers outputs with the transformer for text classification
    Bugueno, Margarita
    Mendoza, Marcelo
    INTELLIGENT DATA ANALYSIS, 2020, 24 (S1) : S15 - S41
  • [8] ARRHYTHMIA CLASSIFICATION WITH HEARTBEAT-AWARE TRANSFORMER
    Wang, Bin
    Liu, Chang
    Hu, Chuanyan
    Liu, Xudong
    Cao, Jun
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 1025 - 1029
  • [9] Graph Receptive Transformer Encoder for Text Classification
    Aras, Arda Can
    Alikasifoglu, Tuna
    Koc, Aykut
    IEEE TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING OVER NETWORKS, 2024, 10 : 347 - 359
  • [10] Graph affine Transformer with a symmetric adaptation strategy for text classification
    Ma, Minyi
    Gong, Hongfang
    Ding, Yingjing
    JOURNAL OF SUPERCOMPUTING, 2025, 81 (03)