FlipTest: Fairness Testing via Optimal Transport

被引:41
作者
Black, Emily [1 ]
Yeom, Samuel [1 ]
Fredrikson, Matt [1 ]
机构
[1] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
来源
FAT* '20: PROCEEDINGS OF THE 2020 CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY | 2020年
基金
美国国家科学基金会;
关键词
fairness; machine learning; optimal transport; disparate impact;
D O I
10.1145/3351095.3372845
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present FlipTest, a black-box technique for uncovering discrimination in classifiers. FlipTest is motivated by the intuitive question: had an individual been of a different protected status, would the model have treated them differently? Rather than relying on causal information to answer this question, FlipTest leverages optimal transport to match individuals in different protected groups, creating similar pairs of in-distribution samples. We show how to use these instances to detect discrimination by constructing a flipset: the set of individuals whose classifier output changes post-translation, which corresponds to the set of people who may be harmed because of their group membership. To shed light on why the model treats a given subgroup differently, FlipTest produces a transparency report: a ranking of features that are most associated with the model's behavior on the flipset. Evaluating the approach on three case studies, we show that this provides a computationally inexpensive way to identify subgroups that may be harmed by model discrimination, including in cases where the model satisfies group fairness criteria.
引用
收藏
页码:111 / 121
页数:11
相关论文
共 47 条
  • [21] Equivant, 2019, PRACT GUID COMPAS CO
  • [22] Certifying and Removing Disparate Impact
    Feldman, Michael
    Friedler, Sorelle A.
    Moeller, John
    Scheidegger, Carlos
    Venkatasubramanian, Suresh
    [J]. KDD'15: PROCEEDINGS OF THE 21ST ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2015, : 259 - 268
  • [23] Fairness Testing: Testing Software for Discrimination
    Galhotra, Sainyam
    Brun, Yuriy
    Meliou, Alexandra
    [J]. ESEC/FSE 2017: PROCEEDINGS OF THE 2017 11TH JOINT MEETING ON FOUNDATIONS OF SOFTWARE ENGINEERING, 2017, : 498 - 510
  • [24] Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672
  • [25] Hardt M, 2016, ADV NEUR IN, V29
  • [26] Hebert-Johnson Ursula, 2018, P 35 INT C MACH LEAR, P1944
  • [27] Joseph M., 2016, Advances in Neural Information Processing Systems
  • [28] Kearns M, 2018, PR MACH LEARN RES, V80
  • [29] An Empirical Study of Rich Subgroup Fairness for Machine Learning
    Kearns, Michael
    Neel, Seth
    Roth, Aaron
    Wu, Zhiwei Steven
    [J]. FAT*'19: PROCEEDINGS OF THE 2019 CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, 2019, : 100 - 109
  • [30] Kleinberg J., 2017, Leibniz International Proceedings in Informatics, V67, p43:1