MarIA and BETO are sexist: evaluating gender bias in large language models for Spanish

被引:3
作者
Garrido-Munoz, Ismael [1 ]
Martinez-Santiago, Fernando [1 ]
Montejo-Raez, Arturo [1 ]
机构
[1] Univ Jaen, CEATIC, Campus Las Lagunillas, Jaen 23071, Spain
关键词
Deep learning; Gender bias; Bias evaluation; Language model; BERT; RoBERTa;
D O I
10.1007/s10579-023-09670-3
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
The study of bias in language models is a growing area of work, however, both research and resources are focused on English. In this paper, we make a first approach focusing on gender bias in some freely available Spanish language models trained using popular deep neural networks, like BERT or RoBERTa. Some of these models are known for achieving state-of-the-art results on downstream tasks. These promising results have promoted such models' integration in many real-world applications and production environments, which could be detrimental to people affected for those systems. This work proposes an evaluation framework to identify gender bias in masked language models, with explainability in mind to ease the interpretation of the evaluation results. We have evaluated 20 different models for Spanish, including some of the most popular pretrained ones in the research community. Our findings state that varying levels of gender bias are present across these models.This approach compares the adjectives proposed by the model for a set of templates. We classify the given adjectives into understandable categories and compute two new metrics from model predictions, one based on the internal state (probability) and the other one on the external state (rank). Those metrics are used to reveal biased models according to the given categories and quantify the degree of bias of the models under study.
引用
收藏
页码:1387 / 1417
页数:31
相关论文
共 52 条
  • [1] Abdaoui Amine, 2020, GREGOIRE SIGEL
  • [2] Persistent Anti-Muslim Bias in Large Language Models
    Abid, Abubakar
    Farooqi, Maheen
    Zou, James
    [J]. AIES '21: PROCEEDINGS OF THE 2021 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, 2021, : 298 - 306
  • [3] Al Kuwatly H., 2020, P 4 WORKSH ONL AB HA, P184, DOI DOI 10.18653/V1/2020.ALW-1.21
  • [4] Quantifying Gender Bias in Different Corpora
    Babaeianjelodar, Marzieh
    Lorenz, Stephen
    Gordon, Josh
    Matthews, Jeanna
    Freitag, Evan
    [J]. WWW'20: COMPANION PROCEEDINGS OF THE WEB CONFERENCE 2020, 2020, : 752 - 759
  • [5] Bartl M., 2020, P 2 WORKSHOP GENDER
  • [6] On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?
    Bender, Emily M.
    Gebru, Timnit
    McMillan-Major, Angelina
    Shmitchell, Shmargaret
    [J]. PROCEEDINGS OF THE 2021 ACM CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, FACCT 2021, 2021, : 610 - 623
  • [7] Bertin project, 2021, BERTIN PROJECT BERTI
  • [8] Investigating Gender Bias in BERT
    Bhardwaj, Rishabh
    Majumder, Navonil
    Poria, Soujanya
    [J]. COGNITIVE COMPUTATION, 2021, 13 (04) : 1008 - 1018
  • [9] Bianchi F, 2021, 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), P10065
  • [10] Algorithmic Factors Influencing Bias in Machine Learning
    Blanzeisky, William
    Cunningham, Padraig
    [J]. MACHINE LEARNING AND PRINCIPLES AND PRACTICE OF KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2021, PT I, 2021, 1524 : 559 - 574