An Overview of the Empirical Evaluation of Explainable AI (XAI): A Comprehensive Guideline for User-Centered Evaluation in XAI

被引:0
作者
Naveed, Sidra [1 ]
Stevens, Gunnar [1 ]
Robin-Kern, Dean [2 ]
机构
[1] Univ Siegen, Informat Syst, D-57072 Siegen, Germany
[2] Bikar Metalle GmbH, D-57319 Bad Berleburg, Germany
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 23期
关键词
AI transparency; explainable artificial intelligence (XAI); XAI evaluation procedure; user-centered evaluation; MENTAL WORKLOAD; EXPLANATIONS; SATISFACTION; FRAMEWORK; MODEL;
D O I
10.3390/app142311288
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Recent advances in technology have propelled Artificial Intelligence (AI) into a crucial role in everyday life, enhancing human performance through sophisticated models and algorithms. However, the focus on predictive accuracy has often resulted in opaque black-box models that lack transparency in decision-making. To address this issue, significant efforts have been made to develop explainable AI (XAI) systems that make outcomes comprehensible to users. Various approaches, including new concepts, models, and user interfaces, aim to improve explainability, build user trust, enhance satisfaction, and increase task performance. Evaluation research has emerged to define and measure the quality of these explanations, differentiating between formal evaluation methods and empirical approaches that utilize techniques from psychology and human-computer interaction. Despite the importance of empirical studies, evaluations remain underutilized, with literature reviews indicating a lack of rigorous evaluations from the user perspective. This review aims to guide researchers and practitioners in conducting effective empirical user-centered evaluations by analyzing several studies; categorizing their objectives, scope, and evaluation metrics; and offering an orientation map for research design and metric measurement.
引用
收藏
页数:53
相关论文
共 153 条
  • [1] COGAM: Measuring and Moderating Cognitive Load in Machine Learning Model Explanations
    Abdul, Ashraf
    von der Weth, Christian
    Kankanhalli, Mohan
    Lim, Brian Y.
    [J]. PROCEEDINGS OF THE 2020 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI'20), 2020,
  • [2] Trends and Trajectories for Explainable, Accountable and Intelligible Systems: An HCI Research Agenda
    Abdul, Ashraf
    Vermeulen, Jo
    Wang, Danding
    Lim, Brian
    Kankanhalli, Mohan
    [J]. PROCEEDINGS OF THE 2018 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI 2018), 2018,
  • [3] Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
    Adadi, Amina
    Berrada, Mohammed
    [J]. IEEE ACCESS, 2018, 6 : 52138 - 52160
  • [4] I Don't Know, Is AI Also Used in Airbags? An Empirical Study of Folk Concepts and People's Expectations of Current and Future Artificial Intelligence
    Alizadeh F.
    Stevens G.
    Esau M.
    [J]. i-com, 2021, 20 (01): : 3 - 17
  • [5] Alufaisan Y, 2021, AAAI CONF ARTIF INTE, V35, P6618
  • [6] Anik A. I., 2021, P 2021 CHI C HUMAN F, P1, DOI DOI 10.1145/3411764.3445736
  • [7] Anjomshoae S, 2019, AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, P1078
  • [8] Arksey H., 2005, INT J SOC RES METHOD, V8, P19, DOI [10.1080/1364557032000119616, DOI 10.1080/1364557032000119616]
  • [9] Bansal G., 2021, P 2021 CHI C HUM FAC, P1, DOI 10
  • [10] Bansal Gagan, 2019, P AAAI C HUMAN COMPU, P2