Effects of AI and Logic-Style Explanations on Users' Decisions Under Different Levels of Uncertainty

被引:7
作者
Cau, Federico Maria [1 ]
Hauptmann, Hanna [2 ]
Spano, Lucio Davide [1 ]
Tintarev, Nava [3 ]
机构
[1] Univ Cagliari, Palazzo Sci,Via Osped 72, I-09124 Cagliari Ca, Sardegna, Italy
[2] Univ Utrecht, Heidelberglaan 8, NL-3584 CS Utrecht, Netherlands
[3] Maastricht Univ, Minderbroedersberg 4-6, NL-6211 LK Maastricht, Netherlands
关键词
Explainable AI; user uncertainty; AI uncertainty; AI correctness; explanations; logical reasoning; MNIST; Yelp Reviews; neural networks; CNNs; intelligent user interfaces; TASK COMPLEXITY; TRUST;
D O I
10.1145/3588320
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing eXplainable Artificial Intelligence (XAI) techniques support people in interpreting AI advice. However, although previous work evaluates the users' understanding of explanations, factors influencing the decision support are largely overlooked in the literature. This article addresses this gap by studying the impact of user uncertainty, AI correctness, and the interaction between AI uncertainty and explanation logic-styles for classification tasks. We conducted two separate studies: one requesting participants to recognize handwritten digits and one to classify the sentiment of reviews. To assess the decision making, we analyzed the task performance, agreement with the AI suggestion, and the user's reliance on the XAI interface elements. Participants make their decision relying on three pieces of information in the XAI interface (image or text instance, AI prediction, and explanation). Participants were shown one explanation style (between-participants design) according to three styles of logical reasoning (inductive, deductive, and abductive). This allowed us to study how different levels of AI uncertainty influence the effectiveness of different explanation styles. The results show that user uncertainty and AI correctness on predictions significantly affected users' classification decisions considering the analyzedmetrics. In both domains (images and text), users reliedmainly on the instance to decide. Users were usually overconfident about their choices, and this evidence was more pronounced for text. Furthermore, the inductive style explanations led to overreliance on the AI advice in both domains-it was the most persuasive, even when the AI was incorrect. The abductive and deductive styles have complex effects depending on the domain and the AI uncertainty levels.
引用
收藏
页数:42
相关论文
共 93 条
[1]  
Adebayo J, 2018, ADV NEUR IN, V31
[2]  
Alipour K., 2020, arXiv, DOI DOI 10.48550/ARXIV.2003.00431
[3]   Evaluating Saliency Map Explanations for Convolutional Neural Networks: A User Study [J].
Alqaraawi, Ahmed ;
Schuessler, Martin ;
Weiss, Philipp ;
Costanza, Enrico ;
Berthouze, Nadia .
PROCEEDINGS OF THE 25TH INTERNATIONAL CONFERENCE ON INTELLIGENT USER INTERFACES, IUI 2020, 2020, :275-285
[4]  
Amini Alexander, 2020, Deep Evidential Uncertainty
[5]  
Arya V, 2019, Arxiv, DOI arXiv:1909.03012
[6]  
Asghar N, 2016, Arxiv, DOI arXiv:1605.05362
[7]   On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation [J].
Bach, Sebastian ;
Binder, Alexander ;
Montavon, Gregoire ;
Klauschen, Frederick ;
Mueller, Klaus-Robert ;
Samek, Wojciech .
PLOS ONE, 2015, 10 (07)
[8]   Why Attentions May Not Be Interpretable? [J].
Bai, Bing ;
Liang, Jian ;
Zhang, Guanhua ;
Li, Hao ;
Bai, Kun ;
Wang, Fei .
KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, :25-34
[9]   Proxy Tasks and Subjective Measures Can Be Misleading in Evaluating Explainable AI Systems [J].
Bucinca, Zana ;
Lin, Phoebe ;
Gajos, Krzysztof Z. ;
Glassman, Elena L. .
PROCEEDINGS OF THE 25TH INTERNATIONAL CONFERENCE ON INTELLIGENT USER INTERFACES, IUI 2020, 2020, :454-464
[10]  
Cau Federico Maria, 2020, P 2020 IT WORKSH EXP