Let's go to the Alien Zoo: Introducing an experimental framework to study usability of counterfactual explanations for machine learning

被引:10
作者
Kuhl, Ulrike [1 ,2 ]
Artelt, Andre [2 ]
Hammer, Barbara [2 ]
机构
[1] Bielefeld Univ, Res Inst Cognit & Robot CoR Lab, Bielefeld, Germany
[2] Bielefeld Univ, Fac Technol, Machine Learning Grp, Bielefeld, Germany
来源
FRONTIERS IN COMPUTER SCIENCE | 2023年 / 5卷
基金
欧洲研究理事会;
关键词
explainable AI; human-grounded evaluation; user study; experimental framework; counterfactual explanations; usability; human-computer interaction; FUNCTIONAL THEORY; CAUSABILITY; THINKING;
D O I
10.3389/fcomp.2023.1087929
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Introduction: To foster usefulness and accountability of machine learning (ML), it is essential to explain a model's decisions in addition to evaluating its performance. Accordingly, the field of explainable artificial intelligence (XAI) has resurfaced as a topic of active research, offering approaches to address the "how" and "why" of automated decision-making. Within this domain, counterfactual explanations (CFEs) have gained considerable traction as a psychologically grounded approach to generate post-hoc explanations. To do so, CFEs highlight what changes to a model's input would have changed its prediction in a particular way. However, despite the introduction of numerous CFE approaches, their usability has yet to be thoroughly validated at the human level.Methods: To advance the field of XAI, we introduce the Alien Zoo, an engaging, web-based and game-inspired experimental framework. The Alien Zoo provides the means to evaluate usability of CFEs for gaining new knowledge from an automated system, targeting novice users in a domain-general context. As a proof of concept, we demonstrate the practical efficacy and feasibility of this approach in a user study.Results: Our results suggest the efficacy of the Alien Zoo framework for empirically investigating aspects of counterfactual explanations in a game-type scenario and a low-knowledge domain. The proof of concept study reveals that users benefit from receiving CFEs compared to no explanation, both in terms of objective performance in the proposed iterative learning task, and subjective usability.Discussion: With this work, we aim to equip research groups and practitioners with the means to easily run controlled and well-powered user studies to complement their otherwise often more technology-oriented work. Thus, in the interest of reproducible research, we provide the entire code, together with the underlying models and user data:.
引用
收藏
页数:19
相关论文
共 70 条
[1]   Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI) [J].
Adadi, Amina ;
Berrada, Mohammed .
IEEE ACCESS, 2018, 6 :52138-52160
[2]  
Akula AR, 2020, AAAI CONF ARTIF INTE, V34, P2594
[3]   CLEVR-XAI: A benchmark dataset for the ground truth evaluation of neural network explanations [J].
Arras, Leila ;
Osman, Ahmed ;
Samek, Wojciech .
INFORMATION FUSION, 2022, 81 :14-40
[4]  
Artelt A., 2019, On the computation of counterfactual explanations-A survey
[5]   Evaluating Robustness of Counterfactual Explanations [J].
Artelt, Andre ;
Vaquet, Valerie ;
Velioglu, Riza ;
Hinder, Fabian ;
Brinkrolf, Johannes ;
Schilling, Malte ;
Hammer, Barbara .
2021 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2021), 2021,
[6]   Convex Density Constraints for Computing Plausible Counterfactual Explanations [J].
Artelt, Andre ;
Hammer, Barbara .
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT I, 2020, 12396 :353-365
[7]  
Artelt Andre, 2019, CEML: Counterfactuals for Explaining Machine Learning models-A Python toolbox
[8]  
Bansal Gagan, 2019, AAAI CONF ARTIF INTE, V33, P2429
[9]   Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI [J].
Barredo Arrieta, Alejandro ;
Diaz-Rodriguez, Natalia ;
Del Ser, Javier ;
Bennetot, Adrien ;
Tabik, Siham ;
Barbado, Alberto ;
Garcia, Salvador ;
Gil-Lopez, Sergio ;
Molina, Daniel ;
Benjamins, Richard ;
Chatila, Raja ;
Herrera, Francisco .
INFORMATION FUSION, 2020, 58 :82-115
[10]   Fitting Linear Mixed-Effects Models Using lme4 [J].
Bates, Douglas ;
Maechler, Martin ;
Bolker, Benjamin M. ;
Walker, Steven C. .
JOURNAL OF STATISTICAL SOFTWARE, 2015, 67 (01) :1-48