An Empirical Comparison of Interpretable Models to Post-Hoc Explanations

被引:2
|
作者
Mahya, Parisa [1 ]
Fuernkranz, Johannes [1 ]
机构
[1] Johannes Kepler Univ Linz, Inst Applicat Oriented Knowledge Proc FAW, A-4040 Linz, Austria
关键词
explainable AI; interpretable machine learning; interpretable models; black-box explanation; white-box models;
D O I
10.3390/ai4020023
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, some effort went into explaining intransparent and black-box models, such as deep neural networks or random forests. So-called model-agnostic methods typically approximate the prediction of the intransparent black-box model with an interpretable model, without considering any specifics of the black-box model itself. It is a valid question whether direct learning of interpretable white-box models should not be preferred over post-hoc approximations of intransparent and black-box models. In this paper, we report the results of an empirical study, which compares post-hoc explanations and interpretable models on several datasets for rule-based and feature-based interpretable models. The results seem to underline that often directly learned interpretable models approximate the black-box models at least as well as their post-hoc surrogates, even though the former do not have direct access to the black-box model.
引用
收藏
页码:426 / 436
页数:11
相关论文
共 50 条
  • [1] Comparing Strategies for Post-Hoc Explanations in Machine Learning Models
    Vij, Aabhas
    Nanjundan, Preethi
    MOBILE COMPUTING AND SUSTAINABLE INFORMATICS, 2022, 68 : 585 - 592
  • [2] A Study on Trust in Black Box Models and Post-hoc Explanations
    El Bekri, Nadia
    Kling, Jasmin
    Huber, Marco F.
    14TH INTERNATIONAL CONFERENCE ON SOFT COMPUTING MODELS IN INDUSTRIAL AND ENVIRONMENTAL APPLICATIONS (SOCO 2019), 2020, 950 : 35 - 46
  • [3] When are Post-hoc Conceptual Explanations Identifiable?
    Leemann, Tobias
    Kirchhof, Michael
    Rong, Yao
    Kasneci, Enkelejda
    Kasneci, Gjergji
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 1207 - 1218
  • [4] Generating Recommendations with Post-Hoc Explanations for Citizen Science
    Ben Zaken, Daniel
    Shani, Guy
    Segal, Avi
    Cavalier, Darlene
    Gal, Kobi
    PROCEEDINGS OF THE 30TH ACM CONFERENCE ON USER MODELING, ADAPTATION AND PERSONALIZATION, UMAP 2022, 2022, : 69 - 78
  • [5] The Dangers of Post-hoc Interpretability: Unjustified Counterfactual Explanations
    Laugel, Thibault
    Lesot, Marie-Jeanne
    Marsala, Christophe
    Renard, Xavier
    Detyniecki, Marcin
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 2801 - 2807
  • [6] A Responsible Machine Learning Workflow with Focus on Interpretable Models, Post-hoc Explanation, and Discrimination Testing
    Gill, Navdeep
    Hall, Patrick
    Montgomery, Kim
    Schmidt, Nicholas
    INFORMATION, 2020, 11 (03)
  • [7] Evaluating Stability of Post-hoc Explanations for Business Process Predictions
    Velmurugan, Mythreyi
    Ouyang, Chun
    Moreira, Catarina
    Sindhgatta, Renuka
    SERVICE-ORIENTED COMPUTING (ICSOC 2021), 2021, 13121 : 49 - 64
  • [8] Problems With SHAP and LIME in Interpretable AI for Education: A Comparative Study of Post-Hoc Explanations and Neural-Symbolic Rule Extraction
    Hooshyar, Danial
    Yang, Yeongwook
    IEEE ACCESS, 2024, 12 : 137472 - 137490
  • [9] Exploring Antimicrobial Resistance Prediction Using Post-hoc Interpretable Methods
    Canovas-Segura, Bernardo
    Morales, Antonio
    Lopez Martinez-Carrasco, Antonio
    Campos, Manuel
    Juarez, Jose M.
    Lopez Rodriguez, Lucia
    Palacios, Francisco
    ARTIFICIAL INTELLIGENCE IN MEDICINE: KNOWLEDGE REPRESENTATION AND TRANSPARENT AND EXPLAINABLE SYSTEMS, AIME 2019, 2019, 11979 : 93 - 107
  • [10] Using ontologies to enhance human understandability of global post-hoc explanations of black-box models
    Confalonieri, Roberto
    Weyde, Tillman
    Besold, Tarek R.
    Martin, Fermin Moscoso del Prado
    ARTIFICIAL INTELLIGENCE, 2021, 296