Prediction of Author's Profile Basing on Fine-Tuning BERT Model

被引:0
|
作者
Bsir B. [1 ,2 ]
Khoufi N. [3 ]
Zrigui M. [1 ,2 ]
机构
[1] ISITCom, University of Sousse, Hammam Sousse
[2] Laboratory in Algebra, Numbers Theory and Intelligent Systems, University of Monastir, Monastir
[3] ANLP Research Group, FSEGS, Sfax
来源
Informatica (Slovenia) | 2024年 / 48卷 / 01期
关键词
Author profiling (AP); BERT; deep learning; fine tuning; NLP; PAN 2018 Corpus dataset; Self-attention Transformers; Transformer-model;
D O I
10.31449/inf.v48i1.4839
中图分类号
学科分类号
摘要
The task of author profiling consists in specifying the infer-demographic features of the social networks' users by studying their published content or the interactions between them. In the literature, many research works were conducted to enhance the accuracy of the techniques used in this process. In fact, the existing methods can be divided into two types: simple linear models and complex deep neural network models. Among them, the transformer-based model exhibited the highest efficiency in NLP analysis in several languages (English, German, French, Turk, Arabic, etc.). Despite their good performance, these approaches do not cover author profiling analysis and, thus, should be further enhanced. So, we propose in this paper a new deep learning strategy by training a customized transformer-model to learn the optimal features of our dataset. In this direction, we fine-tune the model by using the transfer learning approach to improve the results with random initialization. We have achieved about 79% of accuracy by modifying model to apply the retraining process using PAN 2018 authorship dataset. © 2024 Slovene Society Informatika. All rights reserved.
引用
收藏
页码:69 / 78
页数:9
相关论文
共 50 条
  • [31] AstroMAE: Redshift Prediction Using a Masked Autoencoder with a Novel Fine-Tuning Architecture
    Fathkouhi, Amirreza Dolatpour
    Fox, Geoffrey Charles
    2024 IEEE 20TH INTERNATIONAL CONFERENCE ON E-SCIENCE, E-SCIENCE 2024, 2024,
  • [32] Prompt-Oriented Fine-Tuning Dual Bert for Aspect-Based Sentiment Analysis
    Yin, Wen
    Xu, Yi
    Liu, Cencen
    Zheng, Dezhang
    Wang, Qi
    Liu, Chuanjie
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PART X, 2023, 14263 : 505 - 517
  • [33] Fine-Tuning BERT-Based Pre-Trained Models for Arabic Dependency Parsing
    Al-Ghamdi, Sharefah
    Al-Khalifa, Hend
    Al-Salman, Abdulmalik
    APPLIED SCIENCES-BASEL, 2023, 13 (07):
  • [34] Comprehensive Review of Large Language Model Fine-Tuning
    Zhang, Qintong
    Wang, Yuchao
    Wang, Hexi
    Wang, Junxin
    Chen, Hai
    Computer Engineering and Applications, 2024, 60 (17) : 17 - 33
  • [35] SelfCCL: Curriculum Contrastive Learning by Transferring Self-Taught Knowledge for Fine-Tuning BERT
    Dehghan, Somaiyeh
    Amasyali, Mehmet Fatih
    APPLIED SCIENCES-BASEL, 2023, 13 (03):
  • [36] Range-Invariant Approximation of Non-Linear Operations for Efficient BERT Fine-Tuning
    Kim, Janghyeon
    Lee, Janghwan
    Choi, Jungwook
    Han, JeongHo
    Lee, Sangheon
    2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC, 2023,
  • [37] Knowledge Graph Fusion for Language Model Fine-Tuning
    Bhana, Nimesh
    van Zyl, Terence L.
    2022 9TH INTERNATIONAL CONFERENCE ON SOFT COMPUTING & MACHINE INTELLIGENCE, ISCMI, 2022, : 167 - 172
  • [38] Fine-tuning and the infrared bull's-eye
    Roberts, John T.
    PHILOSOPHICAL STUDIES, 2012, 160 (02) : 287 - 303
  • [39] Fine-tuning and the infrared bull’s-eye
    John T. Roberts
    Philosophical Studies, 2012, 160 : 287 - 303
  • [40] Fine-Tuning Channel-Pruned Deep Model via Knowledge Distillation
    Zhang, Chong
    Wang, Hong-Zhi
    Liu, Hong-Wei
    Chen, Yi-Lin
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2024, 39 (06) : 1238 - 1247