Counterfactual explanations for misclassified images: How human and machine explanations differ

被引:6
作者
Delaney, Eoin [1 ,2 ,3 ]
Pakrashi, Arjun [1 ,3 ]
Greene, Derek [1 ,2 ,3 ]
Keane, Mark T. [1 ,3 ]
机构
[1] Univ Coll Dublin, Sch Comp Sci, Dublin, Ireland
[2] Insight Ctr Data Analyt, Dublin, Ireland
[3] VistaMilk SFI Res Ctr, Dublin, Ireland
基金
爱尔兰科学基金会;
关键词
XAI; Counterfactual explanation; User testing; BLACK-BOX;
D O I
10.1016/j.artint.2023.103995
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Counterfactual explanations have emerged as a popular solution for the eXplainable AI (XAI) problem of elucidating the predictions of black-box deep-learning systems because people easily understand them, they apply across different problem domains and seem to be legally compliant. Although over 100 counterfactual methods exist in the XAI literature, each claiming to generate plausible explanations akin to those preferred by people, few of these methods have actually been tested on users (similar to 7%). Even fewer studies adopt a user-centered perspective; for instance, asking people for their counterfactual explanations to determine their perspective on a "good explanation". This gap in the literature is addressed here using a novel methodology that (i) gathers human-generated counterfactual explanations for misclassified images, in two user studies and, then, (ii) compares these human-generated explanations to computationally-generated explanations for the same misclassifications. Results indicate that humans do not "minimally edit" images when generating counterfactual explanations. Instead, they make larger, "meaningful" edits that better approximate prototypes in the counterfactual class. An analysis based on "explanation goals" is proposed to account for this divergence between human and machine explanations. The implications of these proposals for future work are discussed. (c) 2023 The Author(s). Published by Elsevier B.V. This is an open access article under the CC BY license (http://creativecommons .org /licenses /by /4 .0/).
引用
收藏
页数:25
相关论文
共 50 条
  • [1] Counterfactual Explanations and Algorithmic Recourses for Machine Learning: A Review
    Verma, Sahil
    Boonsanong, Varich
    Hoang, Minh
    Hines, Keegan
    Dickerson, John
    Shah, Chirag
    ACM COMPUTING SURVEYS, 2024, 56 (12)
  • [2] The Intriguing Relation Between Counterfactual Explanations and Adversarial Examples
    Freiesleben, Timo
    MINDS AND MACHINES, 2022, 32 (01) : 77 - 109
  • [3] DECE: Decision Explorer with Counterfactual Explanations for Machine Learning Models
    Cheng, Furui
    Ming, Yao
    Qu, Huamin
    IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2021, 27 (02) : 1438 - 1447
  • [4] Evaluating Robustness of Counterfactual Explanations
    Artelt, Andre
    Vaquet, Valerie
    Velioglu, Riza
    Hinder, Fabian
    Brinkrolf, Johannes
    Schilling, Malte
    Hammer, Barbara
    2021 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2021), 2021,
  • [5] Generating Robust Counterfactual Explanations
    Guyomard, Victor
    Fessant, Francoise
    Guyet, Thomas
    Bouadi, Tassadit
    Termier, Alexandre
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT III, 2023, 14171 : 394 - 409
  • [6] Disagreement amongst counterfactual explanations: how transparency can be misleading
    Brughmans, Dieter
    Melis, Lissa
    Martens, David
    TOP, 2024, 32 (03) : 429 - 462
  • [7] On the robustness of sparse counterfactual explanations to adverse perturbations
    Virgolin, Marco
    Fracaros, Saverio
    ARTIFICIAL INTELLIGENCE, 2023, 316
  • [8] Counterfactual Explanations in Explainable AI: A Tutorial
    Wang, Cong
    Li, Xiao-Hui
    Han, Haocheng
    Wang, Shendi
    Wang, Luning
    Cao, Caleb Chen
    Chen, Lei
    KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 4080 - 4081
  • [9] The Intriguing Relation Between Counterfactual Explanations and Adversarial Examples
    Timo Freiesleben
    Minds and Machines, 2022, 32 : 77 - 109
  • [10] An empirical study on how humans appreciate automated counterfactual explanations which embrace imprecise information
    Stepin, Ilia
    Alonso-Moral, Jose M.
    Catala, Alejandro
    Pereira-Farina, Martin
    INFORMATION SCIENCES, 2022, 618 : 379 - 399