Disagreement amongst counterfactual explanations: how transparency can be misleading

被引:3
作者
Brughmans, Dieter [1 ]
Melis, Lissa [2 ,3 ]
Martens, David [1 ]
机构
[1] Univ Antwerp, Engn Management Dept, Prinsstr 13, B-2000 Antwerp, Belgium
[2] Penn State Univ, Civil & Environm Engn Dept, 212 Sackett Bldg, University Pk, PA 16802 USA
[3] Maastricht Univ, Sch Business & Econ, Tongersestr 53, NL-6211 LM Maastricht, Netherlands
关键词
XAI; Counterfactual explanations; Machine learning; Disagreement problem;
D O I
10.1007/s11750-024-00670-2
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
Counterfactual explanations are increasingly used as an Explainable Artificial Intelligence (XAI) technique to provide stakeholders of complex machine learning algorithms with explanations for data-driven decisions. The popularity of counterfactual explanations resulted in a boom in the algorithms generating them. However, not every algorithm creates uniform explanations for the same instance. Even though in some contexts multiple possible explanations are beneficial, there are circumstances where diversity amongst counterfactual explanations results in a potential disagreement problem among stakeholders. Ethical issues arise when for example, malicious agents use this diversity to fairwash an unfair machine learning model by hiding sensitive features. As legislators worldwide tend to start including the right to explanations for data-driven, high-stakes decisions in their policies, these ethical issues should be understood and addressed. Our literature review on the disagreement problem in XAI reveals that this problem has never been empirically assessed for counterfactual explanations. Therefore, in this work, we conduct a large-scale empirical analysis, on 40 data sets, using 12 explanation-generating methods, for two black-box models, yielding over 192,000 explanations. Our study finds alarmingly high disagreement levels between the methods tested. A malicious user is able to both exclude and include desired features when multiple counterfactual explanations are available. This disagreement seems to be driven mainly by the data set characteristics and the type of counterfactual algorithm. XAI centers on the transparency of algorithmic decision-making, but our analysis advocates for transparency about this self-proclaimed transparency.
引用
收藏
页码:429 / 462
页数:34
相关论文
共 46 条
  • [1] Aïvodji U, 2019, PR MACH LEARN RES, V97
  • [2] Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
    Barredo Arrieta, Alejandro
    Diaz-Rodriguez, Natalia
    Del Ser, Javier
    Bennetot, Adrien
    Tabik, Siham
    Barbado, Alberto
    Garcia, Salvador
    Gil-Lopez, Sergio
    Molina, Daniel
    Benjamins, Richard
    Chatila, Raja
    Herrera, Francisco
    [J]. INFORMATION FUSION, 2020, 58 : 82 - 115
  • [3] Bordt Sebastian, 2022, FAccT '22: 2022 ACM Conference on Fairness, Accountability, and Transparency, P891, DOI 10.1145/3531146.3533153
  • [4] NICE: an algorithm for nearest instance counterfactual explanations
    Brughmans, Dieter
    Leyman, Pieter
    Martens, David
    [J]. DATA MINING AND KNOWLEDGE DISCOVERY, 2024, 38 (05) : 2665 - 2703
  • [5] Generating collective counterfactual explanations in score-based classification via mathematical optimization
    Carrizosa, Emilio
    Ramirez-Ayerbe, Jasone
    Morales, Dolores Romero
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2024, 238
  • [6] Counterfactual explanations as interventions in latent space
    Crupi, Riccardo
    Castelnovo, Alessandro
    Regoli, Daniele
    San Miguel Gonzalez, Beatriz
    [J]. DATA MINING AND KNOWLEDGE DISCOVERY, 2024, 38 (05) : 2733 - 2769
  • [7] Multi-Objective Counterfactual Explanations
    Dandl, Susanne
    Molnar, Christoph
    Binder, Martin
    Bischl, Bernd
    [J]. PARALLEL PROBLEM SOLVING FROM NATURE - PPSN XVI, PT I, 2020, 12269 : 448 - 469
  • [8] Doshi-Velez Finale, 2017, arXiv
  • [9] Dua D., 2017, UCI MACHINE LEARNING
  • [10] Explainable AI (XAI): Core Ideas, Techniques, and Solutions
    Dwivedi, Rudresh
    Dave, Devam
    Naik, Het
    Singhal, Smiti
    Omer, Rana
    Patel, Pankesh
    Qian, Bin
    Wen, Zhenyu
    Shah, Tejal
    Morgan, Graham
    Ranjan, Rajiv
    [J]. ACM COMPUTING SURVEYS, 2023, 55 (09)