Towards Fair and Robust Classification

被引:4
作者
Sun, Haipei [1 ]
Wu, Kun [2 ]
Wang, Ting [3 ]
Wang, Wendy Hui [2 ]
机构
[1] Facebook Inc, Seattle, WA 98109 USA
[2] Stevens Inst Technol, Hoboken, NJ 07030 USA
[3] Penn State Univ, University Pk, PA 16802 USA
来源
2022 IEEE 7TH EUROPEAN SYMPOSIUM ON SECURITY AND PRIVACY (EUROS&P 2022) | 2022年
基金
美国国家科学基金会;
关键词
Algorithmic fairness; adversarial robustness; classification; trustworthy machine learning; ATTACKS;
D O I
10.1109/EuroSP53844.2022.00030
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Robustness and fairness are two equally important issues for machine learning systems. Despite the active research on robustness and fairness of ML recently, these efforts focus on either fairness or robustness, but not both. To bridge this gap, in this paper, we design Fair and Robust Classification (FRoC) models that equip the classification models with both fairness and robustness. Meeting both fairness and robustness constraints is not trivial due to the tension between them. The trade-off between fairness, robustness, and model accuracy also introduces additional challenge. To address these challenges, we design two FRoC methods, namely FROC-PRE that modifies the input data before model training, and FROC-IN that modifies the model with an adversarial objective function to address both fairness and robustness during training. FROC-IN is suitable to the settings where the users (e.g., ML service providers) only have the access to the model but not the original data, while FROC-PRE works for the settings where the users (e.g., data owners) have the access to both data and a surrogate model that may have similar architecture as the target model. Our extensive experiments on real-world datasets demonstrate that both FROC-IN and FROC-PRE can achieve both fairness and robustness with insignificant accuracy loss of the target model.
引用
收藏
页码:356 / 376
页数:21
相关论文
共 63 条
[1]   Threat of Adversarial Attacks on Deep Learning in Computer Vision: A Survey [J].
Akhtar, Naveed ;
Mian, Ajmal .
IEEE ACCESS, 2018, 6 :14410-14430
[2]  
Alzantot M, 2018, Arxiv, DOI arXiv:1804.07998
[3]  
Angwin J., 2022, Ethics of data and analytics, P254
[4]  
Ateniese Giuseppe, 2015, International Journal of Security and Networks, V10, P137
[5]  
Bechavod Y., 2017, ARXIV
[6]  
Biggio B., 2012, arXiv
[7]   Security Evaluation of Pattern Classifiers under Attack [J].
Biggio, Battista ;
Fumera, Giorgio ;
Roli, Fabio .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2014, 26 (04) :984-996
[8]   Three naive Bayes approaches for discrimination-free classification [J].
Calders, Toon ;
Verwer, Sicco .
DATA MINING AND KNOWLEDGE DISCOVERY, 2010, 21 (02) :277-292
[9]   Building Classifiers with Independency Constraints [J].
Calders, Toon ;
Kamiran, Faisal ;
Pechenizkiy, Mykola .
2009 IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW 2009), 2009, :13-18
[10]  
Carlini N, 2019, Arxiv, DOI arXiv:1902.06705