Improving fairness generalization through a sample-robust optimization method

被引:6
作者
Ferry, Julien [1 ]
Aivodji, Ulrich [2 ]
Gambs, Sebastien [3 ]
Huguet, Marie-Jose [1 ]
Siala, Mohamed [1 ]
机构
[1] Univ Toulouse, INSA, CNRS, LAAS CNRS, Toulouse, France
[2] Ecole Technol Super, Montreal, PQ, Canada
[3] Univ Quebec Montreal, Montreal, PQ, Canada
关键词
Supervised learning; Fairness; Generalization; Distributionally robust optimization;
D O I
10.1007/s10994-022-06191-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Unwanted bias is a major concern in machine learning, raising in particular significant ethical issues when machine learning models are deployed within high-stakes decision systems. A common solution to mitigate it is to integrate and optimize a statistical fairness metric along with accuracy during the training phase. However, one of the main remaining challenges is that current approaches usually generalize poorly in terms of fairness on unseen data. We address this issue by proposing a new robustness framework for statistical fairness in machine learning. The proposed approach is inspired by the domain of distributionally robust optimization and works in ensuring fairness over a variety of samplings of the training set. Our approach can be used to quantify the robustness of fairness but also to improve it when training a model. We empirically evaluate the proposed method and show that it effectively improves fairness generalization. In addition, we propose a simple yet powerful heuristic application of our framework that can be integrated into a wide range of existing fair classification techniques to enhance fairness generalization. Our extensive empirical study using two existing fair classification methods demonstrates the efficiency and scalability of the proposed heuristic approach.
引用
收藏
页码:2131 / 2192
页数:62
相关论文
共 52 条
  • [41] Saeys Y, 2008, LECT NOTES ARTIF INT, V5212, P313, DOI 10.1007/978-3-540-87481-2_21
  • [42] Sagawa Shiori, 2020, ICLR
  • [43] Fairness Warnings and Fair-MAML: Learning Fairly with Minimal Data
    Slack, Dylan
    Friedler, Sorelle A.
    Givental, Emile
    [J]. FAT* '20: PROCEEDINGS OF THE 2020 CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, 2020, : 200 - 209
  • [44] Taskesen B., 2020, ARXIV200709530
  • [45] Tommasi T, 2017, ADV COMPUT VIS PATT, P37, DOI 10.1007/978-3-319-58347-1_2
  • [46] Verma S, 2018, 2018 IEEE/ACM INTERNATIONAL WORKSHOP ON SOFTWARE FAIRNESS (FAIRWARE 2018), P1, DOI [10.1145/3194770.3194776, 10.23919/FAIRWARE.2018.8452913]
  • [47] Understanding and Improving Fairness-Accuracy Trade-offs in Multi-Task Learning
    Wang, Yuyan
    Wang, Xuezhi
    Beutel, Alex
    Prost, Flavien
    Chen, Jilin
    Chi, Ed H.
    [J]. KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 1748 - 1757
  • [48] The comparisons of data mining techniques for the predictive accuracy of probability of default of credit card clients
    Yeh, I-Cheng
    Lien, Che-hui
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2009, 36 (02) : 2473 - 2480
  • [49] Yurochkin Mikhail, 2020, P INT C LEARN REPR I
  • [50] Fairness Beyond Disparate Treatment & Disparate Impact: Learning Classification without Disparate Mistreatment
    Zafar, Muhammad Bilal
    Valera, Isabel
    Rodriguez, Manuel Gomez
    Gummadi, Krishna P.
    [J]. PROCEEDINGS OF THE 26TH INTERNATIONAL CONFERENCE ON WORLD WIDE WEB (WWW'17), 2017, : 1171 - 1180