Using of Transformers Models for Text Classification to Mobile Educational Applications

被引:2
作者
Garrido, Anabel Pilicita [1 ]
Arias, Enrique Barra [1 ]
机构
[1] Univ Politecn Madrid, Madrid, Spain
关键词
Bit error rate; Transformers; Internet; Training; Text categorization; Recurrent neural networks; IEEE transactions; Natural Language Processing; Multiclass Text Classification; Bidirectional Encoder Representations from Transformers;
D O I
10.1109/TLA.2023.10172138
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In Q2 2022, educational apps were the second most popular category on the Google Play Store, accounting for 10.47% of the apps available worldwide. This work explores the application of five BERT-based pre-trained models with the Transformers architecture to classify mobile educational applications. These five models are according to the knowledge field: bert-base-cased, bert-base-uncased, roberta-base, albert-base-v2 and distilbert-base-uncased. This study uses a dataset with educational apps of Google Play, this dataset was enriched with description and category because it lacked this information. In all models, a tokenizer and fine-tuning works were applied for training in the classification task. After training the data, the testing phase was performed in which the models had to go through four training epochs to obtain better results: roberta-base with 81% accuracy, bert-base-uncased with 79% accuracy, bert-base-cased obtained 80% accuracy, albert-base-v2 obtained 78% accuracy and distilbert-base-uncased obtained 76% accuracy.
引用
收藏
页码:730 / 736
页数:7
相关论文
共 64 条
  • [1] Alam T., 2020, BANGLA TEXT CLASSIFI
  • [2] [Anonymous], 2022, EST EST 2022 TIEND G
  • [3] A Comparison of Pre-Trained Language Models for Multi-Class Text Classification in the Financial Domain
    Arslan, Yusuf
    Allix, Kevin
    Veiber, Lisa
    Lothritz, Cedric
    Bissyande, Tegawende F.
    Klein, Jacques
    Goujon, Anne
    [J]. WEB CONFERENCE 2021: COMPANION OF THE WORLD WIDE WEB CONFERENCE (WWW 2021), 2021, : 260 - 268
  • [4] Ataei TS, 2020, FIGURATIVE LANGUAGE PROCESSING, P67
  • [5] Transforming Fake News: Robust Generalisable News Classification Using Transformers
    Blackledge, Ciara
    Atapour-Abarghouei, Amir
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, : 3960 - 3968
  • [6] Blog Bellas Artes, CUALES SON BELLAS AR
  • [7] COMECSO, VENT DISC CIENC SOC
  • [8] computerworld, NAC MARIA PRIM MOD M
  • [9] GPT-3: What's it good for?
    Dale, Robert
    [J]. NATURAL LANGUAGE ENGINEERING, 2021, 27 (01) : 113 - 118
  • [10] Dale R, 2010, CH CRC MACH LEARN PA, P3