Measures for explainable AI: Explanation goodness, user satisfaction, mental models, curiosity, trust, and human-AI performance

被引:40
作者
Hoffman, Robert R. [1 ]
Mueller, Shane T. [2 ]
Klein, Gary [3 ]
Litman, Jordan [4 ]
机构
[1] Inst Human & Machine Cognit, Pensacola, FL 32502 USA
[2] Michigan Technol Univ, Dept Psychol, Houghton, MI USA
[3] MacroCognit LLC, Dayton, OH USA
[4] Univ Maine Machias, Dept Psychol, Machias, ME USA
来源
FRONTIERS IN COMPUTER SCIENCE | 2023年 / 5卷
关键词
explanatory reasoning; machine-generated explanations; measurement; explanation goodness; mental models; trust; performance; SELF-EXPLANATION; AUTOMATION; CATEGORIZATION; INFORMATION; CALIBRATION; MACHINES; ILLUSION; SUPPORT; EXPERTS; LIMITS;
D O I
10.3389/fcomp.2023.1096257
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
If a user is presented an AI system that portends to explain how it works, how do we know whether the explanation works and the user has achieved a pragmatic understanding of the AI? This question entails some key concepts of measurement such as explanation goodness and trust. We present methods for enabling developers and researchers to: (1) Assess the a priori goodness of explanations, (2) Assess users' satisfaction with explanations, (3) Reveal user's mental model of an AI system, (4) Assess user's curiosity or need for explanations, (5) Assess whether the user's trust and reliance on the AI are appropriate, and finally, (6) Assess how the human-XAI work system performs. The methods we present derive from our integration of extensive research literatures and our own psychometric evaluations. We point to the previous research that led to the measurement scales which we aggregated and tailored specifically for the XAI context. Scales are presented in sufficient detail to enable their use by XAI researchers. For Mental Model assessment and Work System Performance, XAI researchers have choices. We point to a number of methods, expressed in terms of methods' strengths and weaknesses, and pertinent measurement issues.
引用
收藏
页数:15
相关论文
共 163 条
  • [1] Adams BD, 2003, TRUST AUTOMATED SYST
  • [2] Alang N., 2017, TURNS OUT ALGORITHMS
  • [3] Mental Models of Mere Mortals with Explanations of Reinforcement Learning
    Anderson, Andrew
    Dodge, Jonathan
    Sadarangani, Amrita
    Juozapaitis, Zoe
    Newman, Evan
    Irvine, Jed
    Chattopadhyay, Souti
    Olson, Matthew
    Fern, Alan
    Burnett, Margaret
    [J]. ACM TRANSACTIONS ON INTERACTIVE INTELLIGENT SYSTEMS, 2020, 10 (02)
  • [4] COGNITIVE MODELING AND INTELLIGENT TUTORING
    ANDERSON, JR
    BOYLE, CF
    CORBETT, AT
    LEWIS, MW
    [J]. ARTIFICIAL INTELLIGENCE, 1990, 42 (01) : 7 - 49
  • [5] [Anonymous], 1996, HUM FAC TRANSP, DOI 10.1201/9781315137957
  • [6] [Anonymous], 2017, N.Y. TIMES
  • [7] [Anonymous], 2006, Joint cognitive systems: Patterns in cognitive systems engineering, DOI DOI 10.1201/9781420005684
  • [8] [Anonymous], 2008, Naturalistic Decision Making and Macrocognition, DOI DOI 10.1080/00461520.2014.965823
  • [9] VERBAL REPORTS AS EVIDENCE OF THE PROCESS OPERATORS KNOWLEDGE
    BAINBRIDGE, L
    [J]. INTERNATIONAL JOURNAL OF MAN-MACHINE STUDIES, 1979, 11 (04): : 411 - 436
  • [10] Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
    Barredo Arrieta, Alejandro
    Diaz-Rodriguez, Natalia
    Del Ser, Javier
    Bennetot, Adrien
    Tabik, Siham
    Barbado, Alberto
    Garcia, Salvador
    Gil-Lopez, Sergio
    Molina, Daniel
    Benjamins, Richard
    Chatila, Raja
    Herrera, Francisco
    [J]. INFORMATION FUSION, 2020, 58 : 82 - 115