Persistent Anti-Muslim Bias in Large Language Models

被引:140
作者
Abid, Abubakar [1 ]
Farooqi, Maheen [2 ]
Zou, James [1 ]
机构
[1] Stanford Univ, Stanford, CA 94305 USA
[2] McMaster Univ, Hamilton, ON, Canada
来源
AIES '21: PROCEEDINGS OF THE 2021 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY | 2021年
关键词
machine learning; language models; bias; stereotypes; ethics;
D O I
10.1145/3461702.3462624
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It has been observed that large-scale language models capture undesirable societal biases, e.g. relating to race and gender; yet religious bias has been relatively unexplored. We demonstrate that GPT-3, a state-of-the-art contextual language model, captures persistent Muslim-violence bias. We probe GPT-3 in various ways, including prompt completion, analogical reasoning, and story generation, to understand this anti-Muslim bias, demonstrating that it appears consistently and creatively in different uses of the model and that it is severe even compared to biases about other religious groups. For instance, "Muslim" is analogized to "terrorist" in 23% of test cases, while "Jewish" is mapped to its most common stereotype, "money," in 5% of test cases. We quantify the positive distraction needed to overcome this bias with adversarial text prompts, and find that use of the most positive 6 adjectives reduces violent completions for "Muslims" from 66% to 20%, but which is still higher than for other religious groups.
引用
收藏
页码:298 / 306
页数:9
相关论文
共 50 条
  • [41] Vision of the future: large language models in ophthalmology
    Tailor, Prashant D.
    D'Souza, Haley S.
    Li, Hanzhou
    Starr, Matthew R.
    CURRENT OPINION IN OPHTHALMOLOGY, 2024, 35 (05) : 391 - 402
  • [42] LARGE LANGUAGE MODELS (LLMS) AND CHATGPT FOR BIOMEDICINE
    Arighi, Cecilia
    Brenner, Steven
    Lu, Zhiyong
    BIOCOMPUTING 2024, PSB 2024, 2024, : 641 - 644
  • [43] Ethical considerations for large language models in ophthalmology
    Kalaw, Fritz Gerald P.
    Baxter, Sally L.
    CURRENT OPINION IN OPHTHALMOLOGY, 2024, 35 (06) : 438 - 446
  • [44] OffensEval 2023: Offensive language identification in the age of Large Language Models
    Zampieri, Marcos
    Rosenthal, Sara
    Nakov, Preslav
    Dmonte, Alphaeus
    Ranasinghe, Tharindu
    NATURAL LANGUAGE ENGINEERING, 2023, 29 (06) : 1416 - 1435
  • [45] Evaluating Nuanced Bias in Large Language Model Free Response Answers
    Healey, Jennifer
    Byrum, Laurie
    Akhtar, Md Nadeem
    Sinha, Moumita
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS, PT II, NLDB 2024, 2024, 14763 : 378 - 391
  • [46] Knowledge-Enhanced Language Models Are Not Bias-Proof: Situated Knowledge and Epistemic Injustice in AI
    Kraft, Angelie
    Soulier, Eloise
    PROCEEDINGS OF THE 2024 ACM CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, ACM FACCT 2024, 2024, : 1433 - 1445
  • [47] Disparities in seizure outcomes revealed by large language models
    Xie, Kevin
    Ojemann, William K. S.
    Gallagher, Ryan S.
    Shinohara, Russell T.
    Lucas, Alfredo
    Hill, Chloe E.
    Hamilton, Roy H.
    Johnson, Kevin B.
    Roth, Dan
    Litt, Brian
    Ellis, Colin A.
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (06) : 1348 - 1355
  • [48] A paradigm shift?-On the ethics of medical large language models
    Grote, Thomas
    Berens, Philipp
    BIOETHICS, 2024, 38 (05) : 383 - 390
  • [49] Artificial intelligence: Augmenting telehealth with large language models
    Snoswell, Centaine L.
    Snoswell, Aaron J.
    Kelly, Jaimon T.
    Caffery, Liam J.
    Smith, Anthony C.
    JOURNAL OF TELEMEDICINE AND TELECARE, 2025, 31 (01) : 150 - 154
  • [50] Exploring Large Language Models in a Limited Resource Scenario
    Panchbhai, Anand
    Pankanti, Smarana
    2021 11TH INTERNATIONAL CONFERENCE ON CLOUD COMPUTING, DATA SCIENCE & ENGINEERING (CONFLUENCE 2021), 2021, : 147 - 152