An Empirical Study on the Perceived Fairness of Realistic, Imperfect Machine Learning Models

被引:67
作者
Harrison, Galen [1 ]
Hanson, Julia [1 ]
Jacinto, Christine [1 ]
Ramirez, Julio [1 ]
Ur, Blase [1 ]
机构
[1] Univ Chicago, Chicago, IL 60637 USA
来源
FAT* '20: PROCEEDINGS OF THE 2020 CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY | 2020年
基金
美国国家科学基金会;
关键词
Fairness; Accountability; Machine Learning; Survey; Data Science;
D O I
10.1145/3351095.3372831
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
There are many competing definitions of what statistical properties make a machine learning model fair. Unfortunately, research has shown that some key properties are mutually exclusive. Realistic models are thus necessarily imperfect, choosing one side of a tradeoff or the other. To gauge perceptions of the fairness of such realistic, imperfect models, we conducted a between-subjects experiment with 502 Mechanical Turk workers. Each participant compared two models for deciding whether to grant bail to criminal defendants. The first model equalized one potentially desirable model property, with the other property varying across racial groups. The second model did the opposite. We tested pairwise trade-offs between the following four properties: accuracy; false positive rate; outcomes; and the consideration of race. We also varied which racial group the model disadvantaged. We observed a preference among participants for equalizing the false positive rate between groups over equalizing accuracy. Nonetheless, no preferences were overwhelming, and both sides of each trade-off we tested were strongly preferred by a non-trivial fraction of participants. We observed nuanced distinctions between participants considering a model "unbiased" and considering it "fair." Furthermore, even when a model within a trade-off pair was seen as fair and unbiased by a majority of participants, we did not observe consensus that a machine learning model was preferable to a human judge. Our results highlight challenges for building machine learning models that are perceived as fair and broadly acceptable in realistic situations.
引用
收藏
页码:392 / 402
页数:11
相关论文
共 29 条
[1]   Auditing black-box models for indirect influence [J].
Adler, Philip ;
Falk, Casey ;
Friedler, Sorelle A. ;
Nix, Tionney ;
Rybeck, Gabriel ;
Scheidegger, Carlos ;
Smith, Brandon ;
Venkatasubramanian, Suresh .
KNOWLEDGE AND INFORMATION SYSTEMS, 2018, 54 (01) :95-122
[2]  
Angwin Julia, 2016, MACHINE BIAS
[3]  
[Anonymous], 2016, COMPAS RISK SCALES: DEMONSTRATING ACCURACY EQUITY AND PREDICTIVE PARITY
[4]  
[Anonymous], 2018, P AAAI
[5]  
[Anonymous], 2018, P CHI, DOI DOI 10.1145/3173574.3174230
[6]   The Moral Machine experiment [J].
Awad, Edmond ;
Dsouza, Sohan ;
Kim, Richard ;
Schulz, Jonathan ;
Henrich, Joseph ;
Shariff, Azim ;
Bonnefon, Jean-Francois ;
Rahwan, Iyad .
NATURE, 2018, 563 (7729) :59-+
[7]   'It's Reducing a Human Being to a Percentage'; Perceptions of Justice in Algorithmic Decisions [J].
Binns, Reuben ;
Van Kleek, Max ;
Veale, Michael ;
Lyngs, Ulrik ;
Zhao, Jun ;
Shadbolt, Nigel .
PROCEEDINGS OF THE 2018 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI 2018), 2018,
[8]   Fair Prediction with Disparate Impact: A Study of Bias in Recidivism Prediction Instruments [J].
Chouldechova, Alexandra .
BIG DATA, 2017, 5 (02) :153-163
[9]   Algorithmic Transparency via Quantitative Input Influence: Theory and Experiments with Learning Systems [J].
Datta, Anupam ;
Sen, Shayak ;
Zick, Yair .
2016 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2016, :598-617
[10]  
Feldman M., 2014, Certifying and Removing Disparate Impact