A novel approach for assessing fairness in deployed machine learning algorithms

被引:1
作者
Uddin, Shahadat [1 ]
Lu, Haohui [1 ]
Rahman, Ashfaqur [2 ]
Gao, Junbin [3 ]
机构
[1] Univ Sydney, Fac Engn, Sch Project Management, Camperdown, NSW 2037, Australia
[2] CSIRO Data61, Hobart, Tas, Australia
[3] Univ Sydney, Business Sch, Discipline Business Analyt, Camperdown, NSW 2006, Australia
来源
SCIENTIFIC REPORTS | 2024年 / 14卷 / 01期
关键词
Fair machine learning; Fairness; Machine learning; OMITTED VARIABLE BIAS;
D O I
10.1038/s41598-024-68651-w
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Fairness in machine learning (ML) emerges as a critical concern as AI systems increasingly influence diverse aspects of society, from healthcare decisions to legal judgments. Many studies show evidence of unfair ML outcomes. However, the current body of literature lacks a statistically validated approach that can evaluate the fairness of a deployed ML algorithm against a dataset. A novel evaluation approach is introduced in this research based on k-fold cross-validation and statistical t-tests to assess the fairness of ML algorithms. This approach was exercised across five benchmark datasets using six classical ML algorithms. Considering four fair ML definitions guided by the current literature, our analysis showed that the same dataset generates a fair outcome for one ML algorithm but an unfair result for another. Such an observation reveals complex, context-dependent fairness issues in ML, complicated further by the varied operational mechanisms of the underlying ML models. Our proposed approach enables researchers to check whether deploying any ML algorithms against a protected attribute within datasets is fair. We also discuss the broader implications of the proposed approach, highlighting a notable variability in its fairness outcomes. Our discussion underscores the need for adaptable fairness definitions and the exploration of methods to enhance the fairness of ensemble approaches, aiming to advance fair ML practices and ensure equitable AI deployment across societal sectors.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] A reinforcement federated learning based strategy for urinary disease dataset processing
    Ahmed, Saleem
    Groenli, Tor-Morten
    Lakhan, Abdullah
    Chen, Yi
    Liang, Guoxi
    [J]. COMPUTERS IN BIOLOGY AND MEDICINE, 2023, 163
  • [2] Angwin J., 2022, Ethics of data and analytics
  • [3] Reconciling modern machine-learning practice and the classical bias-variance trade-off
    Belkin, Mikhail
    Hsu, Daniel
    Ma, Siyuan
    Mandal, Soumik
    [J]. PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2019, 116 (32) : 15849 - 15854
  • [4] Fairness in Criminal Justice Risk Assessments: The State of the Art
    Berk, Richard
    Heidari, Hoda
    Jabbari, Shahin
    Kearns, Michael
    Roth, Aaron
    [J]. SOCIOLOGICAL METHODS & RESEARCH, 2021, 50 (01) : 3 - 44
  • [5] Bogen M., 2018, Analysis & Policy Observatory, P1
  • [6] Random forests
    Breiman, L
    [J]. MACHINE LEARNING, 2001, 45 (01) : 5 - 32
  • [7] Cross-validation methods
    Browne, MW
    [J]. JOURNAL OF MATHEMATICAL PSYCHOLOGY, 2000, 44 (01) : 108 - 132
  • [8] Fairness in Machine Learning: A Survey
    Caton, Simon
    Haas, Christian
    [J]. ACM COMPUTING SURVEYS, 2024, 56 (07) : 1 - 38
  • [9] Cellini SR, 2008, REV HIGH EDUC, V31, P329
  • [10] The phantom menace: Omitted variable bias in econometric research
    Clarke, KA
    [J]. CONFLICT MANAGEMENT AND PEACE SCIENCE, 2005, 22 (04) : 341 - 352