Specific-Input LIME Explanations for Tabular Data Based on Deep Learning Models

被引:14
作者
An, Junkang [1 ]
Zhang, Yiwan [1 ]
Joe, Inwhee [1 ]
机构
[1] Hanyang Univ, Dept Comp Sci, Seoul 04763, South Korea
来源
APPLIED SCIENCES-BASEL | 2023年 / 13卷 / 15期
关键词
explainable AI; interpretability; machine learning; tabular data;
D O I
10.3390/app13158782
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Deep learning researchers believe that as deep learning models evolve, they can perform well on many tasks. However, the complex parameters of deep learning models make it difficult for users to understand how deep learning models make predictions. In this paper, we propose the specific-input local interpretable model-agnostic explanations (LIME) model, a novel interpretable artificial intelligence (XAI) method that interprets deep learning models of tabular data. The specific-input process uses feature importance and partial dependency plots (PDPs) to select the "what" and "how". In our experiments, we first obtain a basic interpretation of the data by simulating user behaviour. Second, we use our approach to understand "which" features deep learning models focus on and how these features affect the model's predictions. From the experimental results, we find that this approach improves the stability of LIME interpretations, compensates for the problem of LIME only focusing on local interpretations, and achieves a balance between global and local interpretations.
引用
收藏
页数:19
相关论文
共 21 条
[1]  
Baehrens D, 2010, J MACH LEARN RES, V11, P1803
[2]   Intelligible Models for HealthCare: Predicting Pneumonia Risk and Hospital 30-day Readmission [J].
Caruana, Rich ;
Lou, Yin ;
Gehrke, Johannes ;
Koch, Paul ;
Sturm, Marc ;
Elhadad, Noemie .
KDD'15: PROCEEDINGS OF THE 21ST ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2015, :1721-1730
[3]   Modeling wine preferences by data mining from physicochemical properties [J].
Cortez, Paulo ;
Cerdeira, Antonio ;
Almeida, Fernando ;
Matos, Telmo ;
Reis, Jose .
DECISION SUPPORT SYSTEMS, 2009, 47 (04) :547-553
[4]  
Doshi-Velez F, 2017, Arxiv, DOI arXiv:1702.08608
[5]   Greedy function approximation: A gradient boosting machine [J].
Friedman, JH .
ANNALS OF STATISTICS, 2001, 29 (05) :1189-1232
[6]   A Survey of Methods for Explaining Black Box Models [J].
Guidotti, Riccardo ;
Monreale, Anna ;
Ruggieri, Salvatore ;
Turin, Franco ;
Giannotti, Fosca ;
Pedreschi, Dino .
ACM COMPUTING SURVEYS, 2019, 51 (05)
[7]   INTERPRETABLE CLASSIFIERS USING RULES AND BAYESIAN ANALYSIS: BUILDING A BETTER STROKE PREDICTION MODEL [J].
Letham, Benjamin ;
Rudin, Cynthia ;
McCormick, Tyler H. ;
Madigan, David .
ANNALS OF APPLIED STATISTICS, 2015, 9 (03) :1350-1371
[8]   The Mythos of Model Interpretability [J].
Lipton, Zachary C. .
COMMUNICATIONS OF THE ACM, 2018, 61 (10) :36-43
[9]  
Lundberg SM, 2017, ADV NEUR IN, V30
[10]  
Molnar C., 2020, Interpretable Machine Learning A Guide for Making Black Box Models Explainable