ParsBERT: Transformer-based Model for Persian Language Understanding

被引:63
|
作者
Farahani, Mehrdad [1 ]
Gharachorloo, Mohammad [2 ]
Farahani, Marzieh [3 ]
Manthouri, Mohammad [4 ]
机构
[1] Islamic Azad Univ, Dept Comp Engn, North Tehran Branch, Tehran, Iran
[2] Queensland Univ Technol, Sch Elect Engn & Robot, Brisbane, Qld, Australia
[3] Umea Univ, Dept Comp Sci, Umea, Sweden
[4] Shahed Univ, Dept Elect & Elect Engn, Tehran, Iran
关键词
Persian; Transformers; BERT; Language Models; NLP; NLU;
D O I
10.1007/s11063-021-10528-4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The surge of pre-trained language models has begun a new era in the field of Natural Language Processing (NLP) by allowing us to build powerful language models. Among these models, Transformer-based models such as BERT have become increasingly popular due to their state-of-the-art performance. However, these models are usually focused on English, leaving other languages to multilingual models with limited resources. This paper proposes a monolingual BERT for the Persian language (ParsBERT), which shows its state-of-the-art performance compared to other architectures and multilingual models. Also, since the amount of data available for NLP tasks in Persian is very restricted, a massive dataset for different NLP tasks as well as pre-training the model is composed. ParsBERT obtains higher scores in all datasets, including existing ones and gathered ones, and improves the state-of-the-art performance by outperforming both multilingual BERT and other prior works in Sentiment Analysis, Text Classification, and Named Entity Recognition tasks.
引用
收藏
页码:3831 / 3847
页数:17
相关论文
共 50 条
  • [1] ParsBERT: Transformer-based Model for Persian Language Understanding
    Mehrdad Farahani
    Mohammad Gharachorloo
    Marzieh Farahani
    Mohammad Manthouri
    Neural Processing Letters, 2021, 53 : 3831 - 3847
  • [2] LVBERT: Transformer-Based Model for Latvian Language Understanding
    Znotins, Arturs
    Barzdins, Guntis
    HUMAN LANGUAGE TECHNOLOGIES - THE BALTIC PERSPECTIVE (HLT 2020), 2020, 328 : 111 - 115
  • [3] Transformer-based Natural Language Understanding and Generation
    Zhang, Feng
    An, Gaoyun
    Ruan, Qiuqi
    2022 16TH IEEE INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING (ICSP2022), VOL 1, 2022, : 281 - 284
  • [4] Bangla-BERT: Transformer-Based Efficient Model for Transfer Learning and Language Understanding
    Kowsher, M.
    Sami, Abdullah A. S.
    Prottasha, Nusrat Jahan
    Arefin, Mohammad Shamsul
    Dhar, Pranab Kumar
    Koshiba, Takeshi
    IEEE ACCESS, 2022, 10 : 91855 - 91870
  • [5] Vision transformer-based visual language understanding of the construction process
    Yang, Bin
    Zhang, Binghan
    Han, Yilong
    Liu, Boda
    Hu, Jiniming
    Jin, Yiming
    ALEXANDRIA ENGINEERING JOURNAL, 2024, 99 : 242 - 256
  • [6] A transformer-based deep learning model for Persian moral sentiment analysis
    Karami, Behnam
    Bakouie, Fatemeh
    Gharibzadeh, Shahriar
    JOURNAL OF INFORMATION SCIENCE, 2023,
  • [7] Transformer-based heart language model with electrocardiogram annotations
    Tudjarski, Stojancho
    Gusev, Marjan
    Kanoulas, Evangelos
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [8] AN EMPIRICAL STUDY OF TRANSFORMER-BASED NEURAL LANGUAGE MODEL ADAPTATION
    Li, Ke
    Liu, Zhe
    He, Tianxing
    Huang, Hongzhao
    Peng, Fuchun
    Povey, Daniel
    Khudanpur, Sanjeev
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 7934 - 7938
  • [9] TransPolymer: a Transformer-based language model for polymer property predictions
    Xu, Changwen
    Wang, Yuyang
    Farimani, Amir Barati
    NPJ COMPUTATIONAL MATERIALS, 2023, 9 (01)
  • [10] Transformer-Based Single-Cell Language Model: A Survey
    Lan, Wei
    He, Guohang
    Liu, Mingyang
    Chen, Qingfeng
    Cao, Junyue
    Peng, Wei
    BIG DATA MINING AND ANALYTICS, 2024, 7 (04): : 1169 - 1186