Dissecting bias of ChatGPT in college major recommendations

被引:2
作者
Zheng, Alex [1 ]
机构
[1] Carnegie Mellon Univ, 3801 Evesham Dr, Plano, TX 75025 USA
基金
美国安德鲁·梅隆基金会;
关键词
Large language models (LLM); ChatGPT; Bias; Prompt engineering; College major recommendation;
D O I
10.1007/s10799-024-00430-5
中图分类号
G25 [图书馆学、图书馆事业]; G35 [情报学、情报工作];
学科分类号
1205 ; 120501 ;
摘要
Large language models (LLMs) such as ChatGPT play a crucial role in guiding critical decisions nowadays, such as in choosing a college major. Therefore, it is essential to assess the limitations of these models' recommendations and understand any potential biases that may mislead human decisions. In this study, I investigate bias in terms of GPT-3.5 Turbo's college major recommendations for students with various profiles, looking at demographic disparities in factors such as race, gender, and socioeconomic status, as well as educational disparities such as score percentiles. To conduct this analysis, I sourced public data for California seniors who have taken standardized tests like the California Standard Test (CAST) in 2023. By constructing prompts for the ChatGPT API, allowing the model to recommend majors based on high school student profiles, I evaluate bias using various metrics, including the Jaccard Coefficient, Wasserstein Metric, and STEM Disparity Score. The results of this study reveal a significant disparity in the set of recommended college majors, irrespective of the bias metric applied. Notably, the most pronounced disparities are observed for students who fall into minority categories, such as LGBTQ + , Hispanic, or the socioeconomically disadvantaged. Within these groups, ChatGPT demonstrates a lower likelihood of recommending STEM majors compared to a baseline scenario where these criteria are unspecified. For example, when employing the STEM Disparity Score metric, an LGBTQ + student scoring at the 50th percentile faces a 50% reduced chance of receiving a STEM major recommendation in comparison to a male student, with all other factors held constant. Additionally, an average Asian student is three times more likely to receive a STEM major recommendation than an African-American student. Meanwhile, students facing socioeconomic disadvantages have a 30% lower chance of being recommended a STEM major compared to their more privileged counterparts. These findings highlight the pressing need to acknowledge and rectify biases within language models, especially when they play a critical role in shaping personalized decisions. Addressing these disparities is essential to foster a more equitable educational and career environment for all students.
引用
收藏
页数:12
相关论文
共 16 条
  • [1] Addressing algorithmic bias and the perpetuation of health inequities: An AI bias aware framework
    Agarwal, R.
    Bjarnadottir, M.
    Rhue, L.
    Dugas, M.
    Crowley, K.
    Clark, J.
    Gao, G.
    [J]. HEALTH POLICY AND TECHNOLOGY, 2023, 12 (01)
  • [2] Alwahaidi K, 2023, CBC Radio
  • [3] An efficient recommendation generation using relevant Jaccard similarity
    Bag, Sujoy
    Kumar, Sri Krishna
    Tiwari, Manoj Kumar
    [J]. INFORMATION SCIENCES, 2019, 483 : 53 - 64
  • [4] Algorithmic Bias in Education
    Baker, Ryan S.
    Hawn, Aaron
    [J]. INTERNATIONAL JOURNAL OF ARTIFICIAL INTELLIGENCE IN EDUCATION, 2022, 32 (04) : 1052 - 1092
  • [5] Barocas S, 2023, Fairness and Machine Learning: Limitations and Opportunities
  • [6] Chen J., 2023, ACM Trans Inf Syst, V4, P1
  • [7] CollegeData, 2023, Five things college applicants should know about using ChatGPT
  • [8] Liang PP, 2021, PR MACH LEARN RES, V139
  • [9] Luo N, 2023, INFORMS WORKSH DAT S
  • [10] Dissecting racial bias in an algorithm used to manage the health of populations
    Obermeyer, Ziad
    Powers, Brian
    Vogeli, Christine
    Mullainathan, Sendhil
    [J]. SCIENCE, 2019, 366 (6464) : 447 - +