Approximate Inverse Model Explanations (AIME): Unveiling Local and Global Insights in Machine Learning Models

被引:4
作者
Nakanishi, Takafumi [1 ]
机构
[1] Musashino Univ, Dept Data Sci, Tokyo 1358181, Japan
关键词
Approximate inverse models; explainable artificial intelligence (XAI); feature importance; generalized inverse matrices; interpretability; model explanation techniques;
D O I
10.1109/ACCESS.2023.3314336
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Data-driven decision-making has become pervasive in the fields of interpretive machine learning and Explainable AI (XAI). While both fields aim to improve human comprehension of machine learning models, they differ in focus. Interpretive machine learning centers on deciphering outcomes in transparent, or 'glass-box,' models, whereas XAI focuses on creating tools for explaining complex 'black-box' models in a human-understandable way. Some existing interpretable machine learning and explainable AI methods have utilized a forward problem to derive how the prediction and estimation output results of a black-box model change with respect to the input. However, methods adopting the forward problem lead to non-intuitive explanations. Therefore, hypothesizing that the inverse problem can yield more intuitive explanations, we propose approximate inverse model explanations (AIME), which offer unified global and local feature importance by deriving approximate inverse operators for black-box models. Additionally, we introduce a representative instance similarity distribution plot, aiding comprehension of the predictive behavior of the model and target dataset. In our experiments with LightGBM, AIME proved effective across diverse data types, from tabular and handwritten digit images to text data. Results demonstrate that AIME's explanations are not only simpler but more intuitive than those generated by well-established methods like LIME and SHAP. It also visualizes similarity distribution with the target dataset, illustrating the relation between different predictions. Furthermore, AIME estimates local and global feature importance and provides fresh insights by visualizing the similarity distribution between representative estimation instances and the target dataset.
引用
收藏
页码:101020 / 101044
页数:25
相关论文
共 44 条
  • [1] Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
    Adadi, Amina
    Berrada, Mohammed
    [J]. IEEE ACCESS, 2018, 6 : 52138 - 52160
  • [2] [Anonymous], 2019, LOFO Importance
  • [3] Shapley Chains: Extending Shapley Values to Classifier Chains
    Ayad, Celia Wafa
    Bonnier, Thomas
    Bosch, Benjamin
    Read, Jesse
    [J]. DISCOVERY SCIENCE (DS 2022), 2022, 13601 : 541 - 555
  • [4] A Generic and Model-Agnostic Exemplar Synthetization Framework for Explainable AI
    Barbalau, Antonio
    Cosma, Adrian
    Ionescu, Radu Tudor
    Popescu, Marius
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2020, PT II, 2021, 12458 : 190 - 205
  • [5] Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
    Barredo Arrieta, Alejandro
    Diaz-Rodriguez, Natalia
    Del Ser, Javier
    Bennetot, Adrien
    Tabik, Siham
    Barbado, Alberto
    Garcia, Salvador
    Gil-Lopez, Sergio
    Molina, Daniel
    Benjamins, Richard
    Chatila, Raja
    Herrera, Francisco
    [J]. INFORMATION FUSION, 2020, 58 : 82 - 115
  • [6] Bramhall S., 2020, SMU Data Sci. Rev., V3, P4
  • [7] Bykov K, 2021, Arxiv, DOI [arXiv:2108.10346, 10.48550/arXiv.2108.10346]
  • [8] Machine Learning Interpretability: A Survey on Methods and Metrics
    Carvalho, Diogo, V
    Pereira, Eduardo M.
    Cardoso, Jaime S.
    [J]. ELECTRONICS, 2019, 8 (08)
  • [9] Algorithmic Transparency via Quantitative Input Influence: Theory and Experiments with Learning Systems
    Datta, Anupam
    Sen, Shayak
    Zick, Yair
    [J]. 2016 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2016, : 598 - 617
  • [10] Dwork C., 2012, P 3 INNOVATIONS C, P214, DOI DOI 10.1145/2090236.2090255