What Clinicians Want: Contextualizing Explainable Machine Learning for Clinical End Use

被引:0
作者
Tonekaboni, Sana [1 ,2 ,4 ]
Joshi, Shalmali [2 ]
McCradden, Melissa D. [2 ,3 ,4 ]
Goldenberg, Anna [1 ,2 ,4 ]
机构
[1] Univ Toronto, Dept Comp Sci, Toronto, ON, Canada
[2] Vector Inst Artificial Intelligence, Toronto, ON, Canada
[3] Hosp Sick Children, Dept Bioeth, Toronto, ON, Canada
[4] Hosp Sick Children, Dept Genet & Genome Biol, Toronto, ON, Canada
来源
MACHINE LEARNING FOR HEALTHCARE CONFERENCE, VOL 106 | 2019年 / 106卷
关键词
SCORE; MODEL; SATURATION; INTERVIEWS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Translating machine learning (ML) models effectively to clinical practice requires establishing clinicians' trust. Explainability, or the ability of an ML model to justify its outcomes and assist clinicians in rationalizing the model prediction, has been generally understood to be critical to establishing trust. However, the field suffers from the lack of concrete definitions for usable explanations in different settings. To identify specific aspects of explainability that may catalyze building trust in ML models, we surveyed clinicians from two distinct acute care specialties (Intenstive Care Unit and Emergency Department). We use their feedback to characterize when explainability helps to improve clinicians' trust in ML models. We further identify the classes of explanations that clinicians identified as most relevant and crucial for effective translation to clinical practice. Finally, we discern concrete metrics for rigorous evaluation of clinical explainability methods. By integrating perceptions of explainability between clinicians and ML researchers we hope to facilitate the endorsement and broader adoption and sustained use of ML systems in healthcare.
引用
收藏
页数:21
相关论文
共 79 条
  • [1] Adebayo J., 2018, Advances in Neural Information Processing Systems, P9505, DOI DOI 10.5555/3327546.3327621
  • [2] Ahmad MA, 2018, IEEE INT CONF HEALT, P447, DOI [10.1109/ICHI.2018.00095, 10.1145/3233547.3233667]
  • [3] On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation
    Bach, Sebastian
    Binder, Alexander
    Montavon, Gregoire
    Klauschen, Frederick
    Mueller, Klaus-Robert
    Samek, Wojciech
    [J]. PLOS ONE, 2015, 10 (07):
  • [4] Minimal Impact of Implemented Early Warning Score and Best Practice Alert for Patient Deterioration
    Bedoya, Armando D.
    Clement, Meredith E.
    Phelan, Matthew
    Steorts, Rebecca C.
    O'Brien, Cara
    Goldstein, Benjamin A.
    [J]. CRITICAL CARE MEDICINE, 2019, 47 (01) : 49 - 55
  • [5] Lipton ZC, 2017, Arxiv, DOI arXiv:1606.03490
  • [6] Cai Carrie J., 2019, arXiv, DOI [DOI 10.48550/ARXIV.1902.02960, 10.48550/ARXIV.1902.02960]
  • [7] Intelligible Models for HealthCare: Predicting Pneumonia Risk and Hospital 30-day Readmission
    Caruana, Rich
    Lou, Yin
    Gehrke, Johannes
    Koch, Paul
    Sturm, Marc
    Elhadad, Noemie
    [J]. KDD'15: PROCEEDINGS OF THE 21ST ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2015, : 1721 - 1730
  • [8] Che ZP, 2015, Arxiv, DOI arXiv:1512.03542
  • [9] Choi E, 2016, ADV NEUR IN, V29
  • [10] Perceptions of geoengineering: public attitudes, stakeholder perspectives, and the challenge of 'upstream' engagement
    Corner, Adam
    Pidgeon, Nick
    Parkhill, Karen
    [J]. WILEY INTERDISCIPLINARY REVIEWS-CLIMATE CHANGE, 2012, 3 (05) : 451 - 466