Explainable Artificial Intelligence (XAI): What we know and what is left to attain Trustworthy Artificial Intelligence

被引:558
作者
Ali, Sajid [1 ]
Abuhmed, Tamer [2 ]
El-Sappagh, Shaker [2 ,3 ,4 ]
Muhammad, Khan [5 ]
Alonso-Moral, Jose M. [6 ]
Confalonieri, Roberto [7 ]
Guidotti, Riccardo [8 ]
Del Ser, Javier [9 ,10 ]
Diaz-Rodriguez, Natalia [11 ]
Herrera, Francisco [11 ]
机构
[1] Sungkyunkwan Univ, Coll Informat & Commun Engn, Dept Elect & Comp Engn, Informat Lab InfoLab, Suwon 16419, South Korea
[2] Sungkyunkwan Univ, Coll Comp & Informat, Dept Comp Sci & Engn, Informat Lab InfoLab, Suwon 16419, South Korea
[3] Galala Univ, Fac Comp Sci & Engn, Suez, Egypt
[4] Benha Univ, Fac Comp & Artificial Intelligence, Informat Syst Dept, Banha 13518, Egypt
[5] Sungkyunkwan Univ, Coll Comp & Informat, Dept Appl Artificial Intelligence, Visual Analyt Knowledge Lab VIS2KNOW Lab, Seoul 03063, South Korea
[6] Univ Santiago De Compostela, Ctr Singular Invest Tecnoloxias Intelixentes CiTIU, Rua Jenaro Fuente Dominguez S-N, Santiago De Compostela 15782, A Coruna, Spain
[7] Univ Padua, Dept Math Tullio Levi Civita, I-35121 Civita, Padova, Italy
[8] Univ Pisa, Dept Comp Sci, I-56127 Pisa, Italy
[9] Basque Res & Technol Alliance BRTA, TECNALIA, Derio 48160, Spain
[10] Univ Basque Country UPV EHU, Dept Commun Engn, Bilbao 48013, Spain
[11] Univ Granada, Andalusian Res Inst Data Sci & Computat Intelligen, Dept Comp Sci & Artificial Intelligence, Granada 18071, Spain
基金
新加坡国家研究基金会;
关键词
Explainable Artificial Intelligence; Interpretable machine learning; Trustworthy AI; AI principles; Post-hoc explainability; XAI assessment; Data Fusion; Deep Learning; RULE INDUCTION ALGORITHM; DEEP NEURAL-NETWORKS; BLACK-BOX MODELS; EXTRACTING RULES; CLASSIFICATION PROBLEMS; VISUAL ANALYTICS; EXPLANATION; PREDICTION; DECISIONS; SELECTION;
D O I
10.1016/j.inffus.2023.101805
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Artificial intelligence (AI) is currently being utilized in a wide range of sophisticated applications, but the outcomes of many AI models are challenging to comprehend and trust due to their black-box nature. Usually, it is essential to understand the reasoning behind an AI mode ľs decision-making. Thus, the need for eXplainable AI (XAI) methods for improving trust in AI models has arisen. XAI has become a popular research subject within the AI field in recent years. Existing survey papers have tackled the concepts of XAI, its general terms, and post-hoc explainability methods but there have not been any reviews that have looked at the assessment methods, available tools, XAI datasets, and other related aspects. Therefore, in this comprehensive study, we provide readers with an overview of the current research and trends in this rapidly emerging area with a case study example. The study starts by explaining the background of XAI, common definitions, and summarizing recently proposed techniques in XAI for supervised machine learning. The review divides XAI techniques into four axes using a hierarchical categorization system: (i) data explainability, (ii) model explainability, (iii) post-hoc explainability, and (iv) assessment of explanations. We also introduce available evaluation metrics as well as open-source packages and datasets with future research directions. Then, the significance of explainability in terms of legal demands, user viewpoints, and application orientation is outlined, termed as XAI concerns. This paper advocates for tailoring explanation content to specific user types. An examination of XAI techniques and evaluation was conducted by looking at 410 critical articles, published between January 2016 and October 2022, in reputed journals and using a wide range of research databases as a source of information. The article is aimed at XAI researchers who are interested in making their AI models more trustworthy, as well as towards researchers from other disciplines who are looking for effective XAI methods to complete tasks with confidence while communicating meaning from data.
引用
收藏
页数:52
相关论文
共 511 条
[1]   Trends and Trajectories for Explainable, Accountable and Intelligible Systems: An HCI Research Agenda [J].
Abdul, Ashraf ;
Vermeulen, Jo ;
Wang, Danding ;
Lim, Brian ;
Kankanhalli, Mohan .
PROCEEDINGS OF THE 2018 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI 2018), 2018,
[2]   Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI) [J].
Adadi, Amina ;
Berrada, Mohammed .
IEEE ACCESS, 2018, 6 :52138-52160
[3]  
Adam R., 2021, CREATES SHINY APPL D
[4]  
Adebayoj, 2017, GITHUB AD FAIRML
[5]   Auditing black-box models for indirect influence [J].
Adler, Philip ;
Falk, Casey ;
Friedler, Sorelle A. ;
Nix, Tionney ;
Rybeck, Gabriel ;
Scheidegger, Carlos ;
Smith, Brandon ;
Venkatasubramanian, Suresh .
KNOWLEDGE AND INFORMATION SYSTEMS, 2018, 54 (01) :95-122
[6]  
Agarwal A, 2018, 35 INT C MACHINE LEA, V80
[7]  
Ahern I., 2020, ICLR 2020 C
[8]   FairSight: Visual Analytics for Fairness in Decision Making [J].
Ahn, Yongsu ;
Lin, Yu-Ru .
IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2020, 26 (01) :1086-1095
[9]   A new algorithm for automatic knowledge acquisition in inductive learning [J].
Akgobek, Omer ;
Aydin, Yavuz Selim ;
Oztemel, Ercan ;
Aksoy, Mehmet Sabih .
KNOWLEDGE-BASED SYSTEMS, 2006, 19 (06) :388-395
[10]  
Al-Shedivat M, 2018, Arxiv, DOI arXiv:1801.09808