Notions of explainability and evaluation approaches for explainable artificial intelligence

被引:303
作者
Vilone, Giulia [1 ]
Longo, Luca [1 ]
机构
[1] Technol Univ Dublin, Coll Sci & Hlth, Sch Comp Sci, Dublin, Ireland
关键词
Explainable artificial intelligence; Notions of explainability; Evaluation methods; MACHINE LEARNING-MODELS; BLACK-BOX; EXPLANATION FACILITIES; NEURAL-NETWORK; SYSTEM; INTERPRETABILITY; DECISIONS;
D O I
10.1016/j.inffus.2021.05.009
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Explainable Artificial Intelligence (XAI) has experienced a significant growth over the last few years. This is due to the widespread application of machine learning, particularly deep learning, that has led to the development of highly accurate models that lack explainability and interpretability. A plethora of methods to tackle this problem have been proposed, developed and tested, coupled with several studies attempting to define the concept of explainability and its evaluation. This systematic review contributes to the body of knowledge by clustering all the scientific studies via a hierarchical system that classifies theories and notions related to the concept of explainability and the evaluation approaches for XAI methods. The structure of this hierarchy builds on top of an exhaustive analysis of existing taxonomies and peer-reviewed scientific material. Findings suggest that scholars have identified numerous notions and requirements that an explanation should meet in order to be easily understandable by end-users and to provide actionable information that can inform decision making. They have also suggested various approaches to assess to what degree machine-generated explanations meet these demands. Overall, these approaches can be clustered into human-centred evaluations and evaluations with more objective metrics. However, despite the vast body of knowledge developed around the concept of explainability, there is not a general consensus among scholars on how an explanation should be defined, and how its validity and reliability assessed. Eventually, this review concludes by critically discussing these gaps and limitations, and it defines future research directions with explainability as the starting component of any artificial intelligent system.
引用
收藏
页码:89 / 106
页数:18
相关论文
共 186 条
[1]   Trends and Trajectories for Explainable, Accountable and Intelligible Systems: An HCI Research Agenda [J].
Abdul, Ashraf ;
Vermeulen, Jo ;
Wang, Danding ;
Lim, Brian ;
Kankanhalli, Mohan .
PROCEEDINGS OF THE 2018 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI 2018), 2018,
[2]   Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI) [J].
Adadi, Amina ;
Berrada, Mohammed .
IEEE ACCESS, 2018, 6 :52138-52160
[3]  
Adebayo J., 2018, NEURIPS
[4]  
Adebayo J, 2018, ADV NEUR IN, V31
[5]   An effective metacognitive strategy: learning by doing and explaining with a computer-based Cognitive Tutor [J].
Aleven, VAWMM ;
Koedinger, KR .
COGNITIVE SCIENCE, 2002, 26 (02) :147-179
[6]   User-oriented Assessment of Classification Model Understandability [J].
Allahyari, Hiva ;
Lavesson, Niklas .
ELEVENTH SCANDINAVIAN CONFERENCE ON ARTIFICIAL INTELLIGENCE (SCAI 2011), 2011, 227 :11-19
[7]   A Bibliometric Analysis of the Explainable Artificial Intelligence Research Field [J].
Alonso, Jose M. ;
Castiello, Ciro ;
Mencar, Corrado .
INFORMATION PROCESSING AND MANAGEMENT OF UNCERTAINTY IN KNOWLEDGE-BASED SYSTEMS: THEORY AND FOUNDATIONS, IPMU 2018, PT I, 2018, 853 :3-15
[8]  
Alvarez-Melis D., 2018, ICML Workshop on Human Interpretability in Machine Learning
[9]  
Alvarez-Melis D, 2018, ADV NEUR IN, V31
[10]  
Ancona M., 2018, Towards better understanding of gradient-based attribution methods for deep neural networks, DOI DOI 10.1109/TNSE.2020.2996738