Why did AI get this one wrong? - Tree-based explanations of machine learning model predictions

被引:25
作者
Parimbelli, Enea [1 ,3 ,5 ]
Buonocore, Tommaso Mario [1 ]
Nicora, Giovanna [1 ,2 ]
Michalowski, Wojtek [3 ]
Wilk, Szymon [4 ]
Bellazzi, Riccardo [1 ]
机构
[1] Univ Pavia, Dept Elect Comp & Biomed Engn, Pavia, Italy
[2] enGenome srl, Pavia, Italy
[3] Univ Ottawa, Telfer Sch Management, Ottawa, ON, Canada
[4] Poznan Univ Tech, Inst Comp Sci, Div Intelligent Decis Support Syst, Poznan, Poland
[5] Dept Elect Comp & Biomed Engn, Via Ferrata 5, I-27100 Pavia, Italy
基金
欧盟地平线“2020”;
关键词
XAI; Black-box; Explanation; Local explanation; Interpretable; Explainable; Fidelity; Reliability; Post-hoc; Model agnostic; Surrogate model; DATASET SHIFT;
D O I
10.1016/j.artmed.2022.102471
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Increasingly complex learning methods such as boosting, bagging and deep learning have made ML models more accurate, but harder to interpret and explain, culminating in black-box machine learning models. Model developers and users alike are often presented with a trade-off between performance and intelligibility, especially in high-stakes applications like medicine. In the present article we propose a novel methodological approach for generating explanations for the predictions of a generic machine learning model, given a specific instance for which the prediction has been made. The method, named AraucanaXAI, is based on surrogate, locally-fitted classification and regression trees that are used to provide post-hoc explanations of the prediction of a generic machine learning model. Advantages of the proposed XAI approach include superior fidelity to the original model, ability to deal with non-linear decision boundaries, and native support to both classification and regression problems. We provide a packaged, open-source implementation of the AraucanaXAI method and evaluate its behaviour in a number of different settings that are commonly encountered in medical applications of AI. These include potential disagreement between the model prediction and physician's expert opinion and low reliability of the prediction due to data scarcity.
引用
收藏
页数:9
相关论文
共 52 条
  • [1] Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
    Adadi, Amina
    Berrada, Mohammed
    [J]. IEEE ACCESS, 2018, 6 : 52138 - 52160
  • [2] Adhikari A, 2019, IEEE INT FUZZY SYST
  • [3] Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
    Barredo Arrieta, Alejandro
    Diaz-Rodriguez, Natalia
    Del Ser, Javier
    Bennetot, Adrien
    Tabik, Siham
    Barbado, Alberto
    Garcia, Salvador
    Gil-Lopez, Sergio
    Molina, Daniel
    Benjamins, Richard
    Chatila, Raja
    Herrera, Francisco
    [J]. INFORMATION FUSION, 2020, 58 : 82 - 115
  • [4] Bastani O, 2018, Arxiv, DOI arXiv:1706.09773
  • [5] Carlo Combi, 2022, ARTIF INTELL MED
  • [6] Intelligible and Explainable Machine Learning: Best Practices and Practical Challenges
    Caruana, Rich
    Lundberg, Scott
    Ribeiro, Marco Tulio
    Nori, Harsha
    Jenkins, Samuel
    [J]. KDD '20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2020, : 3511 - 3512
  • [7] SMOTE: Synthetic minority over-sampling technique
    Chawla, Nitesh V.
    Bowyer, Kevin W.
    Hall, Lawrence O.
    Kegelmeyer, W. Philip
    [J]. 2002, American Association for Artificial Intelligence (16)
  • [8] XGBoost: A Scalable Tree Boosting System
    Chen, Tianqi
    Guestrin, Carlos
    [J]. KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, : 785 - 794
  • [9] Chou YL, 2021, Arxiv, DOI [arXiv:2103.04244, 10.48550/arXiv.2103.04244, DOI 10.48550/ARXIV.2103.04244]
  • [10] Zeiler MD, 2013, Arxiv, DOI [arXiv:1311.2901, DOI 10.48550/ARXIV.1311.2901]