Counterfactual explanations for misclassified images: How human and machine explanations differ

被引:6
作者
Delaney, Eoin [1 ,2 ,3 ]
Pakrashi, Arjun [1 ,3 ]
Greene, Derek [1 ,2 ,3 ]
Keane, Mark T. [1 ,3 ]
机构
[1] Univ Coll Dublin, Sch Comp Sci, Dublin, Ireland
[2] Insight Ctr Data Analyt, Dublin, Ireland
[3] VistaMilk SFI Res Ctr, Dublin, Ireland
基金
爱尔兰科学基金会;
关键词
XAI; Counterfactual explanation; User testing; BLACK-BOX;
D O I
10.1016/j.artint.2023.103995
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Counterfactual explanations have emerged as a popular solution for the eXplainable AI (XAI) problem of elucidating the predictions of black-box deep-learning systems because people easily understand them, they apply across different problem domains and seem to be legally compliant. Although over 100 counterfactual methods exist in the XAI literature, each claiming to generate plausible explanations akin to those preferred by people, few of these methods have actually been tested on users (similar to 7%). Even fewer studies adopt a user-centered perspective; for instance, asking people for their counterfactual explanations to determine their perspective on a "good explanation". This gap in the literature is addressed here using a novel methodology that (i) gathers human-generated counterfactual explanations for misclassified images, in two user studies and, then, (ii) compares these human-generated explanations to computationally-generated explanations for the same misclassifications. Results indicate that humans do not "minimally edit" images when generating counterfactual explanations. Instead, they make larger, "meaningful" edits that better approximate prototypes in the counterfactual class. An analysis based on "explanation goals" is proposed to account for this divergence between human and machine explanations. The implications of these proposals for future work are discussed. (c) 2023 The Author(s). Published by Elsevier B.V. This is an open access article under the CC BY license (http://creativecommons .org /licenses /by /4 .0/).
引用
收藏
页数:25
相关论文
共 50 条
  • [21] DeltaExplainer: A Software Debugging Approach to Generating Counterfactual Explanations
    Shree, Sunny
    Chandrasekaran, Jaganmohan
    Lei, Yu
    Kacker, Raghu N.
    Kuhn, D. Richard
    2022 FOURTH IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE TESTING (AITEST 2022), 2022, : 103 - 110
  • [22] Convex Density Constraints for Computing Plausible Counterfactual Explanations
    Artelt, Andre
    Hammer, Barbara
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT I, 2020, 12396 : 353 - 365
  • [23] Finding Regions of Counterfactual Explanations via Robust Optimization
    Maragno, Donato
    Kurtz, Jannis
    Rober, Tabea E.
    Goedhart, Rob
    Birbil, S. Ilker
    den Hertog, Dick
    INFORMS JOURNAL ON COMPUTING, 2024, 36 (05) : 1316 - 1334
  • [24] Contrastive Learning with Counterfactual Explanations for Radiology Report Generation
    Li, Mingjie
    Lin, Haokun
    Qiu, Liang
    Liang, Xiaodan
    Chen, Ling
    Elsaddik, Abdulmotaleb
    Chang, Xiaojun
    COMPUTER VISION-ECCV 2024, PT XLIII, 2025, 15101 : 162 - 180
  • [25] Diffusion Models Based Unconditional Counterfactual Explanations Generation
    Zhong, Zhi
    Wang, Yu
    Zhu, Ziye
    Li, Yun
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2024, 37 (11): : 1010 - 1021
  • [26] Iterative Partial Fulfillment of Counterfactual Explanations: Benefits and Risks
    Zhou, Yilun
    PROCEEDINGS OF THE 2023 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, AIES 2023, 2023, : 248 - 258
  • [27] Efficient computation of counterfactual explanations and counterfactual metrics of prototype-based classifiers
    Artelt, Andre
    Hammer, Barbara
    NEUROCOMPUTING, 2022, 470 : 304 - 317
  • [28] Model-Agnostic Counterfactual Explanations in Credit Scoring
    Dastile, Xolani
    Celik, Turgay
    Vandierendonck, Hans
    IEEE ACCESS, 2022, 10 : 69543 - 69554
  • [29] FCE: Feedback Based Counterfactual Explanations for Explainable AI
    Suffian, Muhammad
    Graziani, Pierluigi
    Alonso, Jose M.
    Bogliolo, Alessandro
    IEEE ACCESS, 2022, 10 : 72363 - 72372
  • [30] Evaluating the Usefulness of Counterfactual Explanations from Bayesian Networks
    Raphaela Butz
    Arjen Hommersom
    Renée Schulz
    Hans van Ditmarsch
    Human-Centric Intelligent Systems, 2024, 4 (2): : 286 - 298