Boundary-guided Black-box Fairness Testing

被引:0
作者
Yin, Ziliiang [1 ]
Zhao, Wentian [1 ]
Song, Tian [1 ]
机构
[1] Beijing Inst Technol, Sch Cyberspace Sci & Technol, Beijing, Peoples R China
来源
2024 IEEE 48TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE, COMPSAC 2024 | 2024年
关键词
Fairness Testing; Boundary-Guided Method; Individual Discriminatory Samples;
D O I
10.1109/COMPSAC61105.2024.00163
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Although deep learning models have achieved outstanding performance in many applications, there are still concerns about their fairness. A series of fairness testing methods, which evaluate the fairness of deep learning models by generating discriminatory samples, have been proposed. However, these methods either neglect the naturalness of discriminatory samples or roughly select natural discriminatory samples, leading to a decrease in efficiency. In this paper, we introduce a boundary-guided black-box fairness testing method to effectively generate individual discriminatory samples with high efficiency and enhanced naturalness. Our boundary-guided method involves a global exploration phase, which explores multiple paths from the initial samples to the surrogate decision boundary of the target model, imitated from the semantic latent space of a generative adversarial network (GAN). Then, a local perturbation phase explores the nearby space around a given sample for identifying potential discriminatory samples. Extensive experiments on various datasets demonstrate that our approach outperforms state-of-the-art methods in terms of efficiency and effectiveness while maintaining high naturalness.
引用
收藏
页码:1230 / 1239
页数:10
相关论文
共 41 条
  • [1] Black Box Fairness Testing of Machine Learning Models
    Aggarwal, Aniya
    Lohia, Pranay
    Nagar, Seema
    Dey, Kuntal
    Saha, Diptikalyan
    [J]. ESEC/FSE'2019: PROCEEDINGS OF THE 2019 27TH ACM JOINT MEETING ON EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING, 2019, : 625 - 635
  • [2] Big Data's Disparate Impact
    Barocas, Solon
    Selbst, Andrew D.
    [J]. CALIFORNIA LAW REVIEW, 2016, 104 (03) : 671 - 732
  • [3] Becker B., 1996, UCI Machine Learning Repository
  • [4] Fairness in Criminal Justice Risk Assessments: The State of the Art
    Berk, Richard
    Heidari, Hoda
    Jabbari, Shahin
    Kearns, Michael
    Roth, Aaron
    [J]. SOCIOLOGICAL METHODS & RESEARCH, 2021, 50 (01) : 3 - 44
  • [5] Black A., 2020, Dictionary of dimensions of data quality (3dq), dictionary of 60 standardized definitions
  • [6] Algorithmic fairness in credit scoring
    Bono, Teresa
    Croxson, Karen
    Giles, Adam
    [J]. OXFORD REVIEW OF ECONOMIC POLICY, 2021, 37 (03) : 585 - 617
  • [7] Brendel W., 2018, 6 INT C LEARN REPR
  • [8] Brendel Wieland, 2018, INT C LEARNING REPRE
  • [9] Brenninkmeijer B., 2019, Doctoral dissertation
  • [10] HopSkipJumpAttack: A Query-Efficient Decision-Based Attack
    Chen, Jianbo
    Jordan, Michael, I
    Wainwright, Martin J.
    [J]. 2020 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP 2020), 2020, : 1277 - 1294