ParsBERT: Transformer-based Model for Persian Language Understanding

被引:74
作者
Farahani, Mehrdad [1 ]
Gharachorloo, Mohammad [2 ]
Farahani, Marzieh [3 ]
Manthouri, Mohammad [4 ]
机构
[1] Islamic Azad Univ, Dept Comp Engn, North Tehran Branch, Tehran, Iran
[2] Queensland Univ Technol, Sch Elect Engn & Robot, Brisbane, Qld, Australia
[3] Umea Univ, Dept Comp Sci, Umea, Sweden
[4] Shahed Univ, Dept Elect & Elect Engn, Tehran, Iran
关键词
Persian; Transformers; BERT; Language Models; NLP; NLU;
D O I
10.1007/s11063-021-10528-4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The surge of pre-trained language models has begun a new era in the field of Natural Language Processing (NLP) by allowing us to build powerful language models. Among these models, Transformer-based models such as BERT have become increasingly popular due to their state-of-the-art performance. However, these models are usually focused on English, leaving other languages to multilingual models with limited resources. This paper proposes a monolingual BERT for the Persian language (ParsBERT), which shows its state-of-the-art performance compared to other architectures and multilingual models. Also, since the amount of data available for NLP tasks in Persian is very restricted, a massive dataset for different NLP tasks as well as pre-training the model is composed. ParsBERT obtains higher scores in all datasets, including existing ones and gathered ones, and improves the state-of-the-art performance by outperforming both multilingual BERT and other prior works in Sentiment Analysis, Text Classification, and Named Entity Recognition tasks.
引用
收藏
页码:3831 / 3847
页数:17
相关论文
共 50 条
[41]   Quantifying the Bias of Transformer-Based Language Models for African American English in Masked Language Modeling [J].
Salutari, Flavia ;
Ramos, Jerome ;
Rahmani, Hossein A. ;
Linguaglossa, Leonardo ;
Lipani, Aldo .
ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2023, PT I, 2023, 13935 :532-543
[42]   Influence of Language Proficiency on the Readability of Review Text and Transformer-based Models for Determining Language Proficiency [J].
Sazzed, Salim .
COMPANION PROCEEDINGS OF THE WEB CONFERENCE 2022, WWW 2022 COMPANION, 2022, :881-886
[43]   A Transformer-based Model for Older Adult Behavior Change Detection [J].
Akbari, Fateme ;
Sartipi, Kamran .
2022 IEEE 10TH INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS (ICHI 2022), 2022, :27-35
[44]   MPformer: A Transformer-Based Model for Earthen Ruins Climate Prediction [J].
Xu, Guodong ;
Wang, Hai ;
Ji, Shuo ;
Ma, Yuhui ;
Feng, Yi .
TSINGHUA SCIENCE AND TECHNOLOGY, 2024, 29 (06) :1829-1838
[45]   A Hybrid Transformer-Based Model for Optimizing Fake News Detection [J].
Al-Quayed, Fatima ;
Javed, Danish ;
Jhanjhi, N. Z. ;
Humayun, Mamoona ;
Alnusairi, Thanaa S. .
IEEE ACCESS, 2024, 12 :160822-160834
[46]   Transformer-Based Hybrid Forecasting Model for Multivariate Renewable Energy [J].
Padilha, Guilherme Afonso Galindo ;
Ko, JeongRyun ;
Jung, Jason J. ;
Gomes de Mattos Neto, Paulo Salgado .
APPLIED SCIENCES-BASEL, 2022, 12 (21)
[47]   BiCalBERT: An Efficient Transformer-based Model for Chinese Question Answering [J].
Han, Yanbo ;
Zhan, Buchao ;
Zhang, Bin ;
Zhao, Chao ;
Yan, Shankai .
2024 8TH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS, METAHEURISTICS & SWARM INTELLIGENCE, ISMSI 2024, 2024, :100-104
[48]   Transformer-Based Microbubble Localization [J].
Gharamaleki, Sepideh K. ;
Helfield, Brandon ;
Rivaz, Hassan .
2022 IEEE INTERNATIONAL ULTRASONICS SYMPOSIUM (IEEE IUS), 2022,
[49]   Incorporating Relative Position Information in Transformer-Based Sign Language Recognition and Translation [J].
Aloysius, Neena ;
Geetha, M. ;
Nedungadi, Prema .
IEEE ACCESS, 2021, 9 :145929-145942
[50]   A Systematic Review of Transformer-Based Pre-Trained Language Models through Self-Supervised Learning [J].
Kotei, Evans ;
Thirunavukarasu, Ramkumar .
INFORMATION, 2023, 14 (03)