Gender bias and stereotypes in Large Language Models

被引:68
作者
Kotek, Hadas [1 ,2 ]
Dockum, Rikker [3 ]
Sun, David Q. [1 ]
机构
[1] Apple, Cupertino, CA 95014 USA
[2] MIT, Cupertino, CA 95014 USA
[3] Swarthmore Coll, Swarthmore, PA USA
来源
PROCEEDINGS OF THE ACM COLLECTIVE INTELLIGENCE CONFERENCE, CI 2023 | 2023年
关键词
gender; ethics; large language models; explanations; bias; stereotypes; occupations; INFORMATION; RESOLUTION; COHERENCE; ENGLISH;
D O I
10.1145/3582269.3615599
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large Language Models (LLMs) have made substantial progress in the past several months, shattering state-of-the-art benchmarks in many domains. This paper investigates LLMs' behavior with respect to gender stereotypes, a known issue for prior models. We use a simple paradigm to test the presence of gender bias, building on but differing from WinoBias, a commonly used gender bias dataset, which is likely to be included in the training data of current LLMs. We test four recently published LLMs and demonstrate that they express biased assumptions about men and women's occupations. Our contributions in this paper are as follows: (a) LLMs are 3-6 times more likely to choose an occupation that stereotypically aligns with a person's gender; (b) these choices align with people's perceptions better than with the ground truth as reflected in official job statistics; (c) LLMs in fact amplify the bias beyond what is reflected in perceptions or the ground truth; (d) LLMs ignore crucial ambiguities in sentence structure 95% of the time in our study items, but when explicitly prompted, they recognize the ambiguity; (e) LLMs provide explanations for their choices that are factually inaccurate and likely obscure the true reason behind their predictions. That is, they provide rationalizations of their biased behavior. This highlights a key property of these models: LLMs are trained on imbalanced datasets; as such, even with the recent successes of reinforcement learning with human feedback, they tend to reflect those imbalances back at us. As with other types of societal biases, we suggest that LLMs must be carefully tested to ensure that they treat minoritized individuals and communities equitably.
引用
收藏
页码:12 / 24
页数:13
相关论文
共 105 条
  • [1] Persistent Anti-Muslim Bias in Large Language Models
    Abid, Abubakar
    Farooqi, Maheen
    Zou, James
    [J]. AIES '21: PROCEEDINGS OF THE 2021 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, 2021, : 298 - 306
  • [2] Wikipedia, sociology, and the promise and pitfalls of Big Data
    Adams, Julia
    Brueckner, Hannah
    [J]. BIG DATA & SOCIETY, 2015, 2 (02):
  • [3] Armstrong Ashley B, 2023, Who's Afraid of ChatGPT? An Examination of ChatGPT's Implications for LegalWriting, DOI DOI 10.2139/SSRN.4336929
  • [4] The rapid use of gender information: evidence of the time course of pronoun resolution from eyetracking
    Arnold, JE
    Eisenband, JG
    Brown-Schmidt, S
    Trueswell, JC
    [J]. COGNITION, 2000, 76 (01) : B13 - B26
  • [5] Arthur A.E., 2008, INTERGROUP ATTITUDES, P66
  • [6] Azar Z., 2016, Pragmatic relativity: Gender and context affect the use of personal pronouns in discourse differentially across languages, P1295
  • [7] Hashimoto TB, 2018, Arxiv, DOI arXiv:1806.08010
  • [8] Bang Y, 2023, Arxiv, DOI arXiv:2302.04023
  • [9] Barocas S., 2020, P 58 ANN M ASS COMP, P5454, DOI DOI 10.18653/V1/2020.ACL-MAIN.485
  • [10] Big Data's Disparate Impact
    Barocas, Solon
    Selbst, Andrew D.
    [J]. CALIFORNIA LAW REVIEW, 2016, 104 (03) : 671 - 732