Measuring biases in AI-generated co-authorship networks

被引:0
作者
Kalhor, Ghazal [1 ]
Ali, Shiza [2 ]
Mashhadi, Afra [2 ]
机构
[1] Univ Tehran, Coll Engn, Sch Elect & Comp Engn, Tehran, Iran
[2] Univ Washington Bothell, Comp & Software Syst, Bothell, WA 98011 USA
关键词
Large language models; Co-authorship networks; Computer science; Gender; Ethnicity; Biases in network representation; COLLABORATION; DBLP;
D O I
10.1140/epjds/s13688-025-00555-9
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Large Language Models (LLMs) have significantly advanced prompt-based information retrieval, yet their potential to reproduce or amplify social biases remains insufficiently understood. In this study, we investigate this issue through the concrete task of reconstructing real-world co-authorship networks of computer science (CS) researchers using two widely used LLMs-GPT-3.5 Turbo and Mixtral 8x7B. This task offers a structured and quantifiable way to evaluate whether LLM-generated scholarly relationships reflect demographic disparities, as co-authorship is a key proxy for collaboration and recognition in academia. We compare the LLM-generated networks to baseline networks derived from DBLP and Google Scholar, employing both statistical and network science approaches to assess biases related to gender and ethnicity. Our findings show that both LLMs tend to produce more accurate co-authorship links for individuals with Asian or White names, particularly among researchers with lower visibility or limited academic impact. While we find no significant gender disparities in accuracy, the models systematically favor generating co-authorship links that overrepresent Asian and White individuals. Additionally, the structural properties of the LLM-generated networks differ from those of the baseline networks. These results highlight the importance of examining how LLMs represent social and scientific relationships, particularly in contexts where they are increasingly used for knowledge discovery and scholarly search.
引用
收藏
页数:33
相关论文
共 76 条
[1]  
Agarwal S, 2025, Arxiv, DOI [arXiv:2402.01788, DOI 10.48550/ARXIV.2402.01788]
[2]  
Aggrawal N, 2016, PROCEEDINGS ON 2016 2ND INTERNATIONAL CONFERENCE ON NEXT GENERATION COMPUTING TECHNOLOGIES (NGCT), P494, DOI 10.1109/NGCT.2016.7877466
[3]   On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? [J].
Bender, Emily M. ;
Gebru, Timnit ;
McMillan-Major, Angelina ;
Shmitchell, Shmargaret .
PROCEEDINGS OF THE 2021 ACM CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, FACCT 2021, 2021, :610-623
[4]  
Bird S., 2020, FAIRLEARN TOOLKIT AS
[5]  
Blodgett SL, 2020, Arxiv, DOI [arXiv:2005.14050, 10.48550/arXiv.2005.14050, DOI 10.48550/ARXIV.2005.14050]
[6]   Gender and collaboration patterns in a temporal scientific authorship network [J].
Bravo-Hermsdorff, Gecia ;
Felso, Valkyrie ;
Ray, Emily ;
Gunderson, Lee M. ;
Helander, Mary E. ;
Maria, Joana ;
Niv, Yael .
APPLIED NETWORK SCIENCE, 2019, 4 (01)
[7]   Academic careers in Computer Science: continuance and transience of lifetime co-authorships [J].
Cabanac, Guillaume ;
Hubert, Gilles ;
Milard, Beatrice .
SCIENTOMETRICS, 2015, 102 (01) :135-150
[8]   On the Independence of Association Bias and Empirical Fairness in Language Models [J].
Cabello, Laura ;
Jorgensen, Anna Katrine ;
Sogaard, Anders .
PROCEEDINGS OF THE 6TH ACM CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, FACCT 2023, 2023, :370-378
[9]   Building Classifiers with Independency Constraints [J].
Calders, Toon ;
Kamiran, Faisal ;
Pechenizkiy, Mykola .
2009 IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW 2009), 2009, :13-18
[10]   Gender Bias in Word Embeddings: A Comprehensive Analysis of Frequency, Syntax, and Semantics [J].
Caliskan, Aylin ;
Ajay, Pimparkar Parth ;
Charlesworth, Tessa ;
Wolfe, Robert ;
Banaji, Mahzarin R. .
PROCEEDINGS OF THE 2022 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, AIES 2022, 2022, :156-170