Explainable machine learning practices: opening another black box for reliable medical AI

被引:0
|
作者
Emanuele Ratti
Mark Graves
机构
[1] Johannes Kepler University Linz,Institute of PHilosophy and Scientific Method
[2] Parexel AI Labs,Department of Humanities and Arts
[3] Technion Israel Institute of Technology,undefined
来源
AI and Ethics | 2022年 / 2卷 / 4期
关键词
Black box; Machine learning; Medical AI; Reliable AI; Values; Trustworthiness;
D O I
10.1007/s43681-022-00141-z
中图分类号
学科分类号
摘要
In the past few years, machine learning (ML) tools have been implemented with success in the medical context. However, several practitioners have raised concerns about the lack of transparency—at the algorithmic level—of many of these tools; and solutions from the field of explainable AI (XAI) have been seen as a way to open the ‘black box’ and make the tools more trustworthy. Recently, Alex London has argued that in the medical context we do not need machine learning tools to be interpretable at the algorithmic level to make them trustworthy, as long as they meet some strict empirical desiderata. In this paper, we analyse and develop London’s position. In particular, we make two claims. First, we claim that London’s solution to the problem of trust can potentially address another problem, which is how to evaluate the reliability of ML tools in medicine for regulatory purposes. Second, we claim that to deal with this problem, we need to develop London’s views by shifting the focus from the opacity of algorithmic details to the opacity of the way in which ML tools are trained and built. We claim that to regulate AI tools and evaluate their reliability, agencies need an explanation of how ML tools have been built, which requires documenting and justifying the technical choices that practitioners have made in designing such tools. This is because different algorithmic designs may lead to different outcomes, and to the realization of different purposes. However, given that technical choices underlying algorithmic design are shaped by value-laden considerations, opening the black box of the design process means also making transparent and motivating (technical and ethical) values and preferences behind such choices. Using tools from philosophy of technology and philosophy of science, we elaborate a framework showing how an explanation of the training processes of ML tools in medicine should look like.
引用
收藏
页码:801 / 814
页数:13
相关论文
共 50 条
  • [1] Opening the black box: interpretability of machine learning algorithms in electrocardiography
    Bodini, Matteo
    Rivolta, Massimo W.
    Sassi, Roberto
    PHILOSOPHICAL TRANSACTIONS OF THE ROYAL SOCIETY A-MATHEMATICAL PHYSICAL AND ENGINEERING SCIENCES, 2021, 379 (2212):
  • [2] Explainable Machine Learning for Trustworthy AI
    Giannotti, Fosca
    ARTIFICIAL INTELLIGENCE RESEARCH AND DEVELOPMENT, 2022, 356 : 3 - 3
  • [3] Opening the black box of AI-Medicine
    Poon, Aaron I. F.
    Sung, Joseph J. Y.
    JOURNAL OF GASTROENTEROLOGY AND HEPATOLOGY, 2021, 36 (03) : 581 - 584
  • [4] Reliable Autism Spectrum Disorder Diagnosis for Pediatrics Using Machine Learning and Explainable AI
    Jeon, Insu
    Kim, Minjoong
    So, Dayeong
    Kim, Eun Young
    Nam, Yunyoung
    Kim, Seungsoo
    Shim, Sehoon
    Kim, Joungmin
    Moon, Jihoon
    DIAGNOSTICS, 2024, 14 (22)
  • [5] Explainable AI: A Review of Machine Learning Interpretability Methods
    Linardatos, Pantelis
    Papastefanopoulos, Vasilis
    Kotsiantis, Sotiris
    ENTROPY, 2021, 23 (01) : 1 - 45
  • [6] Guaranteeing Correctness in Black-Box Machine Learning: A Fusion of Explainable AI and Formal Methods for Healthcare Decision-Making
    Khan, Nadia
    Nauman, Muhammad
    Almadhor, Ahmad S.
    Akhtar, Nadeem
    Alghuried, Abdullah
    Alhudhaif, Adi
    IEEE ACCESS, 2024, 12 : 90299 - 90316
  • [7] Opening the black box: interpretable machine learning for predictor finding of metabolic syndrome
    Yan Zhang
    Xiaoxu Zhang
    Jaina Razbek
    Deyang Li
    Wenjun Xia
    Liangliang Bao
    Hongkai Mao
    Mayisha Daken
    Mingqin Cao
    BMC Endocrine Disorders, 22
  • [8] Opening the black box: interpretable machine learning for predictor finding of metabolic syndrome
    Zhang, Yan
    Zhang, Xiaoxu
    Razbek, Jaina
    Li, Deyang
    Xia, Wenjun
    Bao, Liangliang
    Mao, Hongkai
    Daken, Mayisha
    Cao, Mingqin
    BMC ENDOCRINE DISORDERS, 2022, 22 (01)
  • [9] Opening the black box: Uncovering the leader trait paradigm through machine learning
    Doornenbal, Brian M.
    Spisak, Brian R.
    van der Laken, Paul A.
    LEADERSHIP QUARTERLY, 2022, 33 (05)
  • [10] Explainable AI: Machine Learning Interpretation in Blackcurrant Powders
    Przybyl, Krzysztof
    SENSORS, 2024, 24 (10)