PL-Transformer: a POS-aware and layer ensemble transformer for text classification

被引:3
|
作者
Shi, Yu [1 ]
Zhang, Xi [1 ]
Yu, Ning [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Key Lab Trustworthy Distributed Comp & Serv BUPT, Minist Educ, Beijing, Peoples R China
关键词
Text classification; Transformer; Part-of-speech; Layer ensemble;
D O I
10.1007/s00521-022-07872-4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The transformer-based models have become the de-facto standard for natural language processing (NLP) tasks. However, most of these models are only designed to capture the implicit semantics among tokens without considering the extra off-the-shelf knowledge (e.g., parts-of-speech) to facilitate the NLP tasks. Additionally, despite using multiple attention-based encoders, they only utilize the embeddings from the last layer, ignoring that from other layers. To address these issues, in this paper, we propose a novel POS-aware and layer ensemble transformer neural network (named as PL-Transformer). PL-Transformer utilizes the parts-of-speech information explicitly and leverages the outputs from different encoder layers with correlation coefficient attention (C-Encoder) jointly. Moreover, we use correlation coefficient attention to bound dot product in C-Encoder, which improves the overall model performance. Extensive experiments on four datasets demonstrate that PL-Transformer can improve the text classification performance. For example, the accuracy on the MPQA dataset is improved by 3.95% over the vanilla transformer.
引用
收藏
页码:1971 / 1982
页数:12
相关论文
共 44 条
  • [21] Heterogeneous Graph Transformer for Meta-structure Learning with Application in Text Classification
    Wang, Shuhai
    Liu, Xin
    Pan, Xiao
    Xu, Hanjie
    Liu, Mingrui
    ACM TRANSACTIONS ON THE WEB, 2023, 17 (03)
  • [22] Comparative Analysis of Traditional Machine Learning and Transformer-based Deep Learning Models for Text Classification
    Aydin, Nazif
    Erdem, Osman Ayhan
    Tekerek, Adem
    JOURNAL OF POLYTECHNIC-POLITEKNIK DERGISI, 2025, 28 (02):
  • [23] Transformer-based active learning for multi-class text annotation and classification
    Afzal, Muhammad
    Hussain, Jamil
    Abbas, Asim
    Hussain, Maqbool
    Attique, Muhammad
    Lee, Sungyoung
    DIGITAL HEALTH, 2024, 10
  • [24] Enhancing Spam Message Classification and Detection Using Transformer-Based Embedding and Ensemble Learning
    Ghourabi, Abdallah
    Alohaly, Manar
    SENSORS, 2023, 23 (08)
  • [25] UVaT: Uncertainty Incorporated View-Aware Transformer for Robust Multi-View Classification
    Li, Yapeng
    Luo, Yong
    Du, Bo
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 5129 - 5143
  • [26] A transformer-based generative adversarial learning to detect sarcasm from Bengali text with correct classification of confusing text
    Lora, Sanzana Karim
    Jahan, Ishrat
    Hussain, Rahad
    Shahriyar, Rifat
    Islam, A. B. M. Alim Al
    HELIYON, 2023, 9 (12)
  • [27] APTrans: Transformer-Based Multilayer Semantic and Locational Feature Integration for Efficient Text Classification
    Ji, Gaoyang
    Chen, Zengzhao
    Liu, Hai
    Liu, Tingting
    Wang, Bing
    APPLIED SCIENCES-BASEL, 2024, 14 (11):
  • [28] Image and Text Aspect Level Multimodal Sentiment Classification Model Using Transformer and Multilayer Attention Interaction
    Yin, Xiuye
    Chen, Liyong
    INTERNATIONAL JOURNAL OF DATA WAREHOUSING AND MINING, 2023, 19 (01) : 22 - 22
  • [29] Ensemble and Transformer Encoder-based Models for the Cervical Cancer Classification Using Pap-smear Images
    Alzahrani, Maysoon
    Khan, Usman Ali
    Al-Garni, Sultan
    JOURNAL OF ELECTRICAL SYSTEMS, 2024, 20 (02) : 1637 - 1646
  • [30] Laplacian Mesh Transformer: Dual Attention and Topology Aware Network for 3D Mesh Classification and Segmentation
    Li, Xiao-Juan
    Yang, Jie
    Zhang, Fang-Lue
    COMPUTER VISION, ECCV 2022, PT XXIX, 2022, 13689 : 541 - 560