Empirical analysis of fairness-aware data segmentation

被引:0
|
作者
Okura, Seiji [1 ]
Mohri, Takao [1 ]
机构
[1] Fujitsu Ltd, Res Ctr AI Eth, Kawasaki, Kanagawa, Japan
来源
2022 IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS, ICDMW | 2022年
关键词
fairness; machine learning; data segmentation; empirical analysis; BIAS;
D O I
10.1109/ICDMW58026.2022.00029
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Fairness in machine learning is a research area that is recently established, for mitigating bias of unfair models that treat unprivileged people unfavorably based on protected attributes. We want to take an approach for mitigating such bias based on the idea of data segmentation, that is, dividing data into segments where people should be treated similarly. Such an approach should be useful in the sense that the mitigation process itself is explainable for cases that similar people should be treated similarly. Although research on such cases exists, the question of effectiveness of data segmentation itself, however, remains to be answered. In this paper, we answer this question by empirically analyzing the experimental results of data segmentation by using two datasets, i.e., the UCI Adult dataset and the Kaggle Give me some credit (gmsc) dataset. We empirically show that (1) fairness can be controllable during training models by the way of dividing data into segments; more specifically, by selecting the attributes and setting the number of segments for adjusting statistics such as statistical parity of the segments and mutual information between the attributes, etc. (2) the effects of data segmentation is dependent on classifiers, and (3) there exist weak trade-offs between fairness and accuracy with regard to data segmentation.
引用
收藏
页码:155 / 162
页数:8
相关论文
共 50 条
  • [1] Fairness-aware Data Integration
    Mazilu, Lacramioara
    Paton, Norman W.
    Konstantinou, Nikolaos
    Fernandes, Alvaro A. A.
    ACM JOURNAL OF DATA AND INFORMATION QUALITY, 2022, 14 (04):
  • [2] Considerations on Fairness-aware Data Mining
    Kamishima, Toshihiro
    Akaho, Shotaro
    Asoh, Hideki
    Sakuma, Jun
    12TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW 2012), 2012, : 378 - 385
  • [3] Fairness-Aware Programming
    Albarghouthi, Aws
    Vinitsky, Samuel
    FAT*'19: PROCEEDINGS OF THE 2019 CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, 2019, : 211 - 219
  • [4] Fairness-Aware PageRank
    Tsioutsiouliklis, Sotiris
    Pitoura, Evaggelia
    Tsaparas, Panayiotis
    Kleftakis, Ilias
    Mamoulis, Nikos
    PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE 2021 (WWW 2021), 2021, : 3815 - 3826
  • [5] Fairness-Aware Graph Sampling for Network Analysis
    Masrour, Farzan
    Santos, Francisco
    Tan, Pang-Ning
    Esfahanian, Abdol-Hossein
    2022 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2022, : 1107 - 1112
  • [6] Tailoring Data Source Distributions for Fairness-aware Data Integration
    Nargesian, Fatemeh
    Asudeh, Abolfazl
    Jagadish, H., V
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2021, 14 (11): : 2519 - 2532
  • [7] Fairness-Aware PAC Learning from Corrupted Data
    Konstantinov, Nikola
    Lampert, Christoph H.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [8] Fairness-Aware Range Queries for Selecting Unbiased Data
    Shetiya, Suraj
    Swift, Ian P.
    Asudeh, Abolfazl
    Das, Gautam
    2022 IEEE 38TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2022), 2022, : 1423 - 1436
  • [9] Fairness-Aware PAC Learning from Corrupted Data
    Konstantinov, Nikola
    Lampert, Christoph H.
    Journal of Machine Learning Research, 2022, 23 : 1 - 60
  • [10] On the Impossibility of Fairness-Aware Learning from Corrupted Data
    Konstantinov, Nikola
    Lampert, Christoph H.
    ALGORITHMIC FAIRNESS THROUGH THE LENS OF CAUSALITY AND ROBUSTNESS WORKSHOP, VOL 171, 2021, 171 : 59 - 72