Stable local interpretable model-agnostic explanations based on a variational autoencoder

被引:9
作者
Xiang, Xu [1 ]
Yu, Hong [1 ]
Wang, Ye [1 ]
Wang, Guoyin [1 ]
机构
[1] Chongqing Univ Posts & Telecommun, Chongqing Key Lab Computat Intelligence, Chongqing, Peoples R China
基金
中国国家自然科学基金;
关键词
LIME; Stability; Local fidelity; Variational autoencoder;
D O I
10.1007/s10489-023-04942-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For humans to trust in artificial intelligence (AI) systems, it is essential for machine learning (ML) models to be interpretable to users. For example, the judicial process requires that AI conclusions must be rigorous and absolutely interpretable. In this paper, we propose a novel approach, VAE-SLIME, for providing stable local interpretable model-agnostic explanations (SLIME) based on a variational autoencoder (VAE). LIME is a technique that explains the predictions of any classifier in an interpretable and faithful manner. Despite the great success of LIME, the most popular method in this category, it has several disadvantages due to its random perturbation-based sampling method. The VAE-SLIME proposed in this paper is specifically designed to address the lack of stability and local fidelity exhibited by LIME for tabular data. VAE-SLIME first employs fixed noise to replace the random Gaussian noise used by the reparameterization trick of the VAE. Then, it uses this new VAE model instead of random perturbation method to generate stable samples. By considering the sequential relationship and flipping of features, a novel explanation stability evaluation metric, the feature sequence stability index (FSSI), is introduced to accurately evaluate the stability of explanations. In a comparison with 6 state-of-the-art approaches on 7 commonly used tabular datasets, the experimental results show beyond doubt that the explanations produced by our approach are most stable, and its local fidelity is 65.17% higher than that of other approaches on average.
引用
收藏
页码:28226 / 28240
页数:15
相关论文
共 34 条
[1]   Injury severity on traffic crashes: A text mining with an interpretable machine-learning approach [J].
Arteaga, Cristian ;
Paz, Alexander ;
Park, JeeWoong .
SAFETY SCIENCE, 2020, 132
[2]   Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI [J].
Barredo Arrieta, Alejandro ;
Diaz-Rodriguez, Natalia ;
Del Ser, Javier ;
Bennetot, Adrien ;
Tabik, Siham ;
Barbado, Alberto ;
Garcia, Salvador ;
Gil-Lopez, Sergio ;
Molina, Daniel ;
Benjamins, Richard ;
Chatila, Raja ;
Herrera, Francisco .
INFORMATION FUSION, 2020, 58 :82-115
[3]   Toward better prediction of recurrence for Cushing's disease: a factorization-machine based neural approach [J].
Fan, Yanghua ;
Li, Dongfang ;
Liu, Yifan ;
Feng, Ming ;
Chen, Qingcai ;
Wang, Renzhi .
INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2021, 12 (03) :625-633
[4]  
Garreau D, 2023, EXPLAINABLE DEEP LEA, ppp293
[5]  
Garreau D, 2021, PR MACH LEARN RES, V139
[6]  
Garreau D, 2020, PR MACH LEARN RES, V108, P1287
[7]   Visual diagnostics of an explainer model: Tools for the assessment of LIME explanations [J].
Goode, Katherine ;
Hofmann, Heike .
STATISTICAL ANALYSIS AND DATA MINING, 2021, 14 (02) :185-200
[8]   SHAP and LIME: An Evaluation of Discriminative Power in Credit Risk [J].
Gramegna, Alex ;
Giudici, Paolo .
FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2021, 4
[9]   Remote Sensing Scene Classification and Explanation Using RSSCNet and LIME [J].
Hung, Sheng-Chieh ;
Wu, Hui-Ching ;
Tseng, Ming-Hseng .
APPLIED SCIENCES-BASEL, 2020, 10 (18)
[10]  
King DB, 2015, ACS SYM SER, V1214, P1, DOI 10.1021/bk-2015-1214.ch001