Categorical and Continuous Features in Counterfactual Explanations of AI Systems

被引:9
作者
Warren, Greta [1 ]
Byrne, Ruth M. J. [2 ,3 ]
Keane, Mark T. [4 ]
机构
[1] Univ Coll Dublin, Sch Comp Sci, Dublin, Ireland
[2] Univ Dublin, Trinity Coll Dublin, Sch Psychol, Dublin, Ireland
[3] Univ Dublin, Trinity Coll Dublin, Inst Neurosci, Dublin, Ireland
[4] Univ Coll Dublin, Insight SFI Ctr Data Analyt, VistaMilk SFI Res Ctr, Sch Comp Sci, Dublin, Ireland
来源
PROCEEDINGS OF 2023 28TH ANNUAL CONFERENCE ON INTELLIGENT USER INTERFACES, IUI 2023 | 2023年
基金
爱尔兰科学基金会;
关键词
XAI; explanation; counterfactual; user study; THINKING; REALITY;
D O I
10.1145/3581641.3584090
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, eXplainable AI (XAI) research has focused on the use of counterfactual explanations to address interpretability, algorithmic recourse, and bias in AI system decision-making. The proponents of these algorithms claim they meet users' requirements for counterfactual explanations. For instance, many claim that the output of their algorithms work as explanations because they prioritise "plausible", "actionable" or "causally important" features in their generated counterfactuals. However, very few of these claims have been tested in controlled psychological studies, and we know very little about which aspects of counterfactual explanations help users to understand AI system decisions. Furthermore, we do not know whether counterfactual explanations are an advance on more traditional causal explanations that have a much longer history in AI (in explaining expert systems and decision trees). Accordingly, we carried out two user studies to (i) test a fundamental distinction in feature-types, between categorical and continuous features, and (ii) compare the relative effectiveness of counterfactual and causal explanations. The studies used a simulated, automated decision-making app that determined safe driving limits after drinking alcohol, based on predicted blood alcohol content, and user responses were measured objectively (users' predictive accuracy) and subjectively (users' satisfaction and trust judgments). Study 1 (N=127) showed that users understand explanations referring to categorical features more readily than those referring to continuous features. It also discovered a dissociation between objective and subjective measures: counterfactual explanations elicited higher accuracy of predictions than no-explanation control descriptions but no higher accuracy than causal explanations, yet counterfactual explanations elicited greater satisfaction and trust judgments than causal explanations. Study 2 (N=211) found that users were more accurate for categorically-transformed features compared to continuous ones, and also replicated the results of Study 1. The findings delineate important boundary conditions for current and future counterfactual explanation methods in XAI.
引用
收藏
页码:171 / 187
页数:17
相关论文
共 70 条
  • [1] The Hidden Assumptions Behind Counterfactual Explanations and Principal Reasons
    Barocas, Solon
    Selbst, Andrew D.
    Raghavan, Manish
    [J]. FAT* '20: PROCEEDINGS OF THE 2020 CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, 2020, : 80 - 89
  • [2] Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
    Barredo Arrieta, Alejandro
    Diaz-Rodriguez, Natalia
    Del Ser, Javier
    Bennetot, Adrien
    Tabik, Siham
    Barbado, Alberto
    Garcia, Salvador
    Gil-Lopez, Sergio
    Molina, Daniel
    Benjamins, Richard
    Chatila, Raja
    Herrera, Francisco
    [J]. INFORMATION FUSION, 2020, 58 : 82 - 115
  • [3] 'It's Reducing a Human Being to a Percentage'; Perceptions of Justice in Algorithmic Decisions
    Binns, Reuben
    Van Kleek, Max
    Veale, Michael
    Lyngs, Ulrik
    Zhao, Jun
    Shadbolt, Nigel
    [J]. PROCEEDINGS OF THE 2018 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI 2018), 2018,
  • [4] Buchanan B. G., 1984, Rule-Based Expert Systems: The MYCIN Experiments of the Stanford Heuristic Programming Project
  • [5] Proxy Tasks and Subjective Measures Can Be Misleading in Evaluating Explainable AI Systems
    Bucinca, Zana
    Lin, Phoebe
    Gajos, Krzysztof Z.
    Glassman, Elena L.
    [J]. PROCEEDINGS OF THE 25TH INTERNATIONAL CONFERENCE ON INTELLIGENT USER INTERFACES, IUI 2020, 2020, : 454 - 464
  • [6] Deductive reasoning with factual, possible, and counterfactual conditionals
    Byrne, RMJ
    Tasso, A
    [J]. MEMORY & COGNITION, 1999, 27 (04) : 726 - 740
  • [7] Byrne RMJ, 2005, RATIONAL IMAGINATION: HOW PEOPLE CREATE ALTERNATIVES TO REALITY, P1
  • [8] Byrne RMJ, 2019, PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P6276
  • [9] Dhurandhar A, 2019, Arxiv, DOI arXiv:1906.00117
  • [10] Explaining Models: An Empirical Study of How Explanations Impact Fairness Judgment
    Dodge, Jonathan
    Liao, Q. Vera
    Zhang, Yunfeng
    Bellamy, Rachel K. E.
    Dugan, Casey
    [J]. PROCEEDINGS OF IUI 2019, 2019, : 275 - 285