Human performance consequences of normative and contrastive explanations: An experiment in machine learning for reliability maintenance

被引:3
作者
Gentile, Davide [1 ]
Donmez, Birsen [1 ]
Jamieson, Greg A. [1 ]
机构
[1] Univ Toronto, Dept Mech & Ind Engn, 5 Kings Coll Rd, Toronto, ON M5S 3G8, Canada
关键词
Human-AI interaction; Explainable AI; Automation reliance behavior; Automation transparency; AGENT TRANSPARENCY; AUTOMATION; FEEDBACK; IMPACT; TRUST;
D O I
10.1016/j.artint.2023.103945
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Decision aids based on artificial intelligence and machine learning can benefit human decisions and system performance, but can also provide incorrect advice, and invite operators to inappropriately rely on automation. This paper examined the extent to which example-based explanations could improve reliance on a decision aid that is based on machine learning. Participants engaged in a preventive maintenance task by providing their diagnosis of the conditions of three components of a hydraulic system. A decision aid based on machine learning provided advice but was not always reliable. Three explanation displays (baseline, normative, normative plus contrastive) were manipulated within-participants. With the normative explanation display, we found improvements in participants' decision time and subjective workload. With the addition of contrastive explanations, we found improvements in participants' hit rate and sensitivity in discriminating between correct and incorrect ML advice. Implications for the design of explainable interfaces to support human-AI interaction in data intensive environments are discussed.& COPY; 2023 Elsevier B.V. All rights reserved.
引用
收藏
页数:17
相关论文
共 53 条
  • [11] Christoffersen K, 2002, ADV HUM PER, V2, P1, DOI 10.1016/S1479-3601(02)02003-9
  • [12] Chromik M., 2020, P IUI WORKSH EXPL SM, V7
  • [13] A POWER PRIMER
    COHEN, J
    [J]. PSYCHOLOGICAL BULLETIN, 1992, 112 (01) : 155 - 159
  • [14] Doshi-Velez F, 2017, Arxiv, DOI [arXiv:1702.08608, DOI 10.48550/ARXIV.1702.08608]
  • [15] G*Power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences
    Faul, Franz
    Erdfelder, Edgar
    Lang, Albert-Georg
    Buchner, Axel
    [J]. BEHAVIOR RESEARCH METHODS, 2007, 39 (02) : 175 - 191
  • [16] Green David Marvin, 1966, Signal Detection Theory and Psychophysics, V1
  • [17] DARPA's Explainable Artificial Intelligence Program
    Gunning, David
    Aha, David W.
    [J]. AI MAGAZINE, 2019, 40 (02) : 44 - 58
  • [18] Helwig N, 2015, IEEE IMTC P, P210, DOI 10.1109/I2MTC.2015.7151267
  • [19] Human Factors in Model Interpretability: Industry Practices, Challenges, and Needs
    Hong S.R.
    Hullman J.
    Bertini E.
    [J]. Proceedings of the ACM on Human-Computer Interaction, 2020, 4 (CSCW1)
  • [20] A review on machinery diagnostics and prognostics implementing condition-based maintenance
    Jardine, Andrew K. S.
    Lin, Daming
    Banjevic, Dragan
    [J]. MECHANICAL SYSTEMS AND SIGNAL PROCESSING, 2006, 20 (07) : 1483 - 1510