Wasserstein Robust Classification with Fairness Constraints

被引:3
作者
Wang, Yijie [1 ]
Nguyen, Viet Anh [2 ]
Hanasusanto, Grani A. [3 ]
机构
[1] Tongji Univ, Sch Econ & Management, Shanghai 200092, Peoples R China
[2] Chinese Univ Hong Kong, Dept Syst Engn & Engn Management, Hong Kong, Peoples R China
[3] Univ Illinois, Dept Ind & Enterprise Syst Engn, Urbana, IL 61801 USA
基金
美国国家科学基金会;
关键词
programming; stochastic methods; MACHINE; RISK;
D O I
10.1287/msom.2022.0230
中图分类号
C93 [管理学];
学科分类号
12 ; 1201 ; 1202 ; 120202 ;
摘要
Problem definition: : Data analytics models and machine learning algorithms are increasingly deployed to support consequential decision -making processes, from deciding which applicants will receive job offers and loans to university enrollments and medical interventions. However, recent studies show these models may unintentionally amplify human bias and yield significant unfavorable decisions to specific groups. Methodology/ results: : We propose a distributionally robust classification model with a fairness constraint that encourages the classifier to be fair in the equality of opportunity criterion. We use a type-infinity infinity Wasserstein ambiguity set centered at the empirical distribution to represent distributional uncertainty and derive a conservative reformulation for the worst -case equal opportunity unfairness measure. We show that the model is equivalent to a mixed binary conic optimization problem, which standard off -the -shelf solvers can solve. We propose a convex, hinge -loss -based model for large problem instances whose reformulation does not incur binary variables to improve scalability. Moreover, we also consider the distributionally robust learning problem with a generic ground transportation cost to hedge against the label and sensitive attribute uncertainties. We numerically examine the performance of our proposed models on five real -world data sets related to individual analysis. Compared with the state-of-the-art methods, our proposed approaches significantly improve fairness with negligible loss of predictive accuracy in the testing data set. Managerial implications: : Our paper raises awareness that bias may arise when predictive models are used in service and operations. It generally comes from human bias, for example, imbalanced data collection or low sample sizes, and is further amplified by algorithms. Incorporating fairness constraints and the distributionally robust optimization (DRO) scheme is a powerful way to alleviate algorithmic biases.
引用
收藏
页码:1567 / 1585
页数:20
相关论文
共 66 条
[1]  
Agarwal Alekh, 2018, P MACHINE LEARNING R, V80
[2]  
Aliprantis CD., 2006, Infinite dimensional analysis: a Hitchhiker's Guide
[3]  
Angwin Julia., 2016, Machine bias: Risk assessments in criminal sentencing
[4]  
[Anonymous], 2013, news release
[5]   Detecting Customer Trends for Optimal Promotion Targeting [J].
Baardman, Lennart ;
Boroujeni, Setareh Borjian ;
Cohen-Hillel, Tamar ;
Panchamgam, Kiran ;
Perakis, Georgia .
M&SOM-MANUFACTURING & SERVICE OPERATIONS MANAGEMENT, 2023, 25 (02) :448-467
[6]  
Bach F., 2021, LEARNING THEORY 1 PR
[7]   Fairness in Criminal Justice Risk Assessments: The State of the Art [J].
Berk, Richard ;
Heidari, Hoda ;
Jabbari, Shahin ;
Kearns, Michael ;
Roth, Aaron .
SOCIOLOGICAL METHODS & RESEARCH, 2021, 50 (01) :3-44
[8]  
Bertsimas D, 2018, DATA DRIVEN APPROACH
[9]   Technical Note-Two-Stage Sample Robust Optimization [J].
Bertsimas, Dimitris ;
Shtern, Shimrit ;
Sturt, Bradley .
OPERATIONS RESEARCH, 2022, 70 (01) :624-640
[10]  
Bird S., 2020, FAIRLEARN: A TOOLKIT FOR ASSESSING AND IMPROVING FAIRNESS IN AI