Prediction of Author's Profile Basing on Fine-Tuning BERT Model

被引:0
作者
Bsir B. [1 ,2 ]
Khoufi N. [3 ]
Zrigui M. [1 ,2 ]
机构
[1] ISITCom, University of Sousse, Hammam Sousse
[2] Laboratory in Algebra, Numbers Theory and Intelligent Systems, University of Monastir, Monastir
[3] ANLP Research Group, FSEGS, Sfax
来源
Informatica (Slovenia) | 2024年 / 48卷 / 01期
关键词
Author profiling (AP); BERT; deep learning; fine tuning; NLP; PAN 2018 Corpus dataset; Self-attention Transformers; Transformer-model;
D O I
10.31449/inf.v48i1.4839
中图分类号
学科分类号
摘要
The task of author profiling consists in specifying the infer-demographic features of the social networks' users by studying their published content or the interactions between them. In the literature, many research works were conducted to enhance the accuracy of the techniques used in this process. In fact, the existing methods can be divided into two types: simple linear models and complex deep neural network models. Among them, the transformer-based model exhibited the highest efficiency in NLP analysis in several languages (English, German, French, Turk, Arabic, etc.). Despite their good performance, these approaches do not cover author profiling analysis and, thus, should be further enhanced. So, we propose in this paper a new deep learning strategy by training a customized transformer-model to learn the optimal features of our dataset. In this direction, we fine-tune the model by using the transfer learning approach to improve the results with random initialization. We have achieved about 79% of accuracy by modifying model to apply the retraining process using PAN 2018 authorship dataset. © 2024 Slovene Society Informatika. All rights reserved.
引用
收藏
页码:69 / 78
页数:9
相关论文
共 45 条
[1]  
Conneau A., Khandelwal K., Goyal N., Chaudhary V., Wenzek G., Guzman F., Grave E., Ott M., Zettlemoyer L., Stoyanov V., Unsupervised cross-lingual representation learning at scale, (2020)
[2]  
Ai M., BERT for Russian news clustering, Proceedings of the International Conference “Dialogue 2021, (2021)
[3]  
Radford A., Narasimhan K., Salimans T., Sutskever I., Improving language understanding by generative pre-training, (2018)
[4]  
Vaswani A., Shazeer N., Parmar N., Uszkoreit J., Jones L., Gomez A. N., Kaiser L. u., Polosukhin I., Attention is all you need, Advances in Neural Information Processing Systems, 30, pp. 5998-6008, (2017)
[5]  
Radford Alec, Wu Jeffrey, Child Rewon, Luan David, Amodei Dario, Sutskever Ilya, Language models are unsupervised multitask learners, (2019)
[6]  
Alvarez-Carmona M. A., Lopez-Monroy A. P., Montes- y Gomez M., Villase nor-Pineda L., Meza I., Evaluating topic-based representations for author profiling in social media, (2016)
[7]  
Antoun Wissam, Baly Fady, Hajj Hazem, Arabert: Transformer-based model for arabic language understanding, (2020)
[8]  
Bernard G., Resources to compute TF-IDF weightings on press articles and tweets, (2022)
[9]  
Bassem B., Zrigui M., Gender identification: a comparative study of deep learning architectures, Intelligent Systems Design and Applications: 18th International Conference on Intelligent Systems Design and Applications (ISDA 2018) held in Vellore, 2, pp. 792-800, (2020)
[10]  
Butt S., Ashraf N., Sidorov G., Gelbukh A. F., Sexism Identification using BERT and Data Augmentation-EXIST2021, IberLEF@ SEPLN, pp. 381-389, (2021)