A Study on Performance Enhancement by Integrating Neural Topic Attention with Transformer-Based Language Model

被引:1
|
作者
Um, Taehum [1 ]
Kim, Namhyoung [1 ]
机构
[1] Gachon Univ, Dept Appl Stat, 1342 Seongnam Daero, Seongnam 13120, South Korea
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 17期
基金
新加坡国家研究基金会;
关键词
natural language processing; neural topic model; ELECTRA; ALBERT; multi-classification;
D O I
10.3390/app14177898
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
As an extension of the transformer architecture, the BERT model has introduced a new paradigm for natural language processing, achieving impressive results in various downstream tasks. However, high-performance BERT-based models-such as ELECTRA, ALBERT, and RoBERTa-suffer from limitations such as poor continuous learning capability and insufficient understanding of domain-specific documents. To address these issues, we propose the use of an attention mechanism to combine BERT-based models with neural topic models. Unlike traditional stochastic topic modeling, neural topic modeling employs artificial neural networks to learn topic representations. Furthermore, neural topic models can be integrated with other neural models and trained to identify latent variables in documents, thereby enabling BERT-based models to sufficiently comprehend the contexts of specific fields. We conducted experiments on three datasets-Movie Review Dataset (MRD), 20Newsgroups, and YELP-to evaluate our model's performance. Compared to the vanilla model, the proposed model achieved an accuracy improvement of 1-2% for the ALBERT model in multiclassification tasks across all three datasets, while the ELECTRA model showed an accuracy improvement of less than 1%.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] AN EMPIRICAL STUDY OF TRANSFORMER-BASED NEURAL LANGUAGE MODEL ADAPTATION
    Li, Ke
    Liu, Zhe
    He, Tianxing
    Huang, Hongzhao
    Peng, Fuchun
    Povey, Daniel
    Khudanpur, Sanjeev
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 7934 - 7938
  • [2] Roles and Utilization of Attention Heads in Transformer-based Neural Language Models
    Jo, Jae-young
    Myaeng, Sung-hyon
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 3404 - 3417
  • [3] Integrating Social Media Insights for Innovation Performance Enhancement: A Transformer-Based Analysis
    Wang, Ang
    Niu, Yue
    JOURNAL OF THE KNOWLEDGE ECONOMY, 2024,
  • [4] Empirical Study of Tweets Topic Classification Using Transformer-Based Language Models
    Mandal, Ranju
    Chen, Jinyan
    Becken, Susanne
    Stantic, Bela
    INTELLIGENT INFORMATION AND DATABASE SYSTEMS, ACIIDS 2021, 2021, 12672 : 340 - 350
  • [5] Is Transformer-Based Attention Agnostic of the Pretraining Language and Task?
    Martin, R. H. J.
    Visser, R.
    Dunaiski, M.
    SOUTH AFRICAN COMPUTER SCIENCE AND INFORMATION SYSTEMS RESEARCH TRENDS, SAICSIT 2024, 2024, 2159 : 95 - 123
  • [6] Tweets Topic Classification and Sentiment Analysis Based on Transformer-Based Language Models
    Mandal, Ranju
    Chen, Jinyan
    Becken, Susanne
    Stantic, Bela
    VIETNAM JOURNAL OF COMPUTER SCIENCE, 2023, 10 (02) : 117 - 134
  • [7] Transformer-based heart language model with electrocardiogram annotations
    Tudjarski, Stojancho
    Gusev, Marjan
    Kanoulas, Evangelos
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [8] LVBERT: Transformer-Based Model for Latvian Language Understanding
    Znotins, Arturs
    Barzdins, Guntis
    HUMAN LANGUAGE TECHNOLOGIES - THE BALTIC PERSPECTIVE (HLT 2020), 2020, 328 : 111 - 115
  • [9] ParsBERT: Transformer-based Model for Persian Language Understanding
    Mehrdad Farahani
    Mohammad Gharachorloo
    Marzieh Farahani
    Mohammad Manthouri
    Neural Processing Letters, 2021, 53 : 3831 - 3847
  • [10] ParsBERT: Transformer-based Model for Persian Language Understanding
    Farahani, Mehrdad
    Gharachorloo, Mohammad
    Farahani, Marzieh
    Manthouri, Mohammad
    NEURAL PROCESSING LETTERS, 2021, 53 (06) : 3831 - 3847