Prediction of Author's Profile Basing on Fine-Tuning BERT Model

被引:0
|
作者
Bsir B. [1 ,2 ]
Khoufi N. [3 ]
Zrigui M. [1 ,2 ]
机构
[1] ISITCom, University of Sousse, Hammam Sousse
[2] Laboratory in Algebra, Numbers Theory and Intelligent Systems, University of Monastir, Monastir
[3] ANLP Research Group, FSEGS, Sfax
来源
Informatica (Slovenia) | 2024年 / 48卷 / 01期
关键词
Author profiling (AP); BERT; deep learning; fine tuning; NLP; PAN 2018 Corpus dataset; Self-attention Transformers; Transformer-model;
D O I
10.31449/inf.v48i1.4839
中图分类号
学科分类号
摘要
The task of author profiling consists in specifying the infer-demographic features of the social networks' users by studying their published content or the interactions between them. In the literature, many research works were conducted to enhance the accuracy of the techniques used in this process. In fact, the existing methods can be divided into two types: simple linear models and complex deep neural network models. Among them, the transformer-based model exhibited the highest efficiency in NLP analysis in several languages (English, German, French, Turk, Arabic, etc.). Despite their good performance, these approaches do not cover author profiling analysis and, thus, should be further enhanced. So, we propose in this paper a new deep learning strategy by training a customized transformer-model to learn the optimal features of our dataset. In this direction, we fine-tune the model by using the transfer learning approach to improve the results with random initialization. We have achieved about 79% of accuracy by modifying model to apply the retraining process using PAN 2018 authorship dataset. © 2024 Slovene Society Informatika. All rights reserved.
引用
收藏
页码:69 / 78
页数:9
相关论文
共 50 条
  • [1] Efficient Fine-Tuning of BERT Models on the Edge
    Vucetic, Danilo
    Tayaranian, Mohammadreza
    Ziaeefard, Maryam
    Clark, James J.
    Meyer, Brett H.
    Gross, Warren J.
    2022 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS 22), 2022, : 1838 - 1842
  • [2] Fine-Tuning BERT Model for Materials Named Entity Recognition
    Zhao, Xintong
    Greenberg, Jane
    An, Yuan
    Hu, Xiaohua Tony
    2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, : 3717 - 3720
  • [3] Fine-Tuning BERT Models for Multiclass Amharic News Document Categorization
    Endalie, Demeke
    COMPLEXITY, 2025, 2025 (01)
  • [4] Transfer fine-tuning of BERT with phrasal paraphrases
    Arase, Yuki
    Tsujii, Junichi
    COMPUTER SPEECH AND LANGUAGE, 2021, 66
  • [5] SPEECH RECOGNITION BY SIMPLY FINE-TUNING BERT
    Huang, Wen-Chin
    Wu, Chia-Hua
    Luo, Shang-Bao
    Chen, Kuan-Yu
    Wang, Hsin-Min
    Toda, Tomoki
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7343 - 7347
  • [6] Investigation of BERT Model on Biomedical Relation Extraction Based on Revised Fine-tuning Mechanism
    Su, Peng
    Vijay-Shanker, K.
    2020 IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE, 2020, : 2522 - 2529
  • [7] Fine-Tuning BERT on Twitter and Reddit Data in Luganda and English
    Kimera, Richard
    Rim, Daniela N.
    Choi, Heeyoul
    PROCEEDINGS OF 2023 7TH INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING AND INFORMATION RETRIEVAL, NLPIR 2023, 2023, : 63 - 70
  • [8] Boosting generalization of fine-tuning BERT for fake news detection
    Qin, Simeng
    Zhang, Mingli
    INFORMATION PROCESSING & MANAGEMENT, 2024, 61 (04)
  • [9] Improving BERT Fine-Tuning via Self-Ensemble and Self-Distillation
    Yi-Ge Xu
    Xi-Peng Qiu
    Li-Gao Zhou
    Xuan-Jing Huang
    Journal of Computer Science and Technology, 2023, 38 : 853 - 866
  • [10] Improving BERT Fine-Tuning via Self-Ensemble and Self-Distillation
    Xu, Yi-Ge
    Qiu, Xi-Peng
    Zhou, Li-Gao
    Huang, Xuan-Jing
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2023, 38 (04) : 853 - 866