Fairness-aware feature selection: A causal path approach

被引:0
作者
Zhang, Wenqiong [1 ]
Li, Yun [1 ]
Liu, Yue [1 ]
机构
[1] Nanjing Univ Posts & Telecommun, Nanjing 210023, Peoples R China
基金
中国国家自然科学基金;
关键词
Machine learning; Fairness; Feature selection; Causal path; DISCRIMINATION; CLASSIFICATION; INFORMATION;
D O I
10.1016/j.knosys.2025.113708
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The real-world datasets used for training machine learning models may contain sensitive features such as gender, race, and age, raising concerns about unfair treatment of specific groups or individuals. These concerns can frequently be alleviated by feature selection techniques. However, most existing research primarily focuses on removing statistically correlated features with sensitive attributes, which frequently fails to address the underlying causal relationships that contribute to unfair treatment. In this paper, we propose a novel fairness-aware causal feature selection method (FCFS). By constructing a causal diagram and leveraging Markov blanket theory, we identify the discrimination-prone feature set that undermine the fairness of model decisions, then propose direct and indirect causal discriminatory effect measures for each feature within the set. Features with high discrimination effect are iteratively assigned priority for removal. Experimental results on public datasets, including Adult, Boston, and German, demonstrate that the proposed method, FCFS, achieves an ideal tradeoff between accuracy and fairness (measured by EOs, DP, EO, and BAD) compared to existing fairness-aware feature selection methods.
引用
收藏
页数:15
相关论文
共 54 条
[1]  
Angwin J., 2022, Ethics of Data and Analytics, P254
[2]   Automating Procedurally Fair Feature Selection in Machine Learning [J].
Belitz, Clara ;
Jiang, Lan ;
Bosch, Nigel .
AIES '21: PROCEEDINGS OF THE 2021 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, 2021, :379-389
[3]   Associations of Race, Insurance, and Zip Code-Level Income with Nonadherence Diagnoses in Primary and Specialty Diabetes Care [J].
Beltran, Sourik ;
Arenas, Daniel J. ;
Lopez-Hinojosa, Itzel J. ;
Tung, Elizabeth L. ;
Cronholm, Peter F. .
JOURNAL OF THE AMERICAN BOARD OF FAMILY MEDICINE, 2021, 34 (05) :891-897
[4]   Categorical Feature Encoding Techniques for Improved Classifier Performance when Dealing with Imbalanced Data of Fraudulent Transactions [J].
Breskuviene, Dalia ;
Dzemyda, Gintautas .
INTERNATIONAL JOURNAL OF COMPUTERS COMMUNICATIONS & CONTROL, 2023, 18 (03)
[5]   Three naive Bayes approaches for discrimination-free classification [J].
Calders, Toon ;
Verwer, Sicco .
DATA MINING AND KNOWLEDGE DISCOVERY, 2010, 21 (02) :277-292
[6]   Fair Transition Loss: From label noise robustness to bias mitigation [J].
Canalli, Ygor ;
Braida, Filipe ;
Alvim, Leandro ;
Zimbrao, Geraldo .
KNOWLEDGE-BASED SYSTEMS, 2024, 294
[7]   Stable learning establishes some common ground between causal inference and machine learning [J].
Cui, Peng ;
Athey, Susan .
NATURE MACHINE INTELLIGENCE, 2022, 4 (02) :110-115
[8]  
Du MN, 2021, ADV NEUR IN, V34
[9]  
Dutta S, 2020, AAAI CONF ARTIF INTE, V34, P3825
[10]  
Ehyaei AR, 2024, AAAI CONF ARTIF INTE, P11847