Model-contrastive explanations through symbolic reasoning

被引:4
|
作者
Malandri, Lorenzo
Mercorio, Fabio [1 ]
Mezzanzanica, Mario
Seveso, Andrea
机构
[1] Univ Milano Bicocca, Dept Stat & Quantitat Methods, Milan, Italy
关键词
eXplainable AI; Contrastive explanation methods for XAI; Post -hoc explainability; XAI Interpretability;
D O I
10.1016/j.dss.2023.114040
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Explaining how two machine learning classification models differ in their behaviour is gaining significance in eXplainable AI, given the increasing diffusion of learning-based decision support systems. Human decisionmakers deal with more than one machine learning model in several practical situations. Consequently, the importance of understanding how two machine learning models work beyond their prediction performances is key to understanding their behaviour, differences, and likeness. Some attempts have been made to address these problems, for instance, by explaining text classifiers in a timecontrastive fashion. In this paper, we present MERLIN, a novel eXplainable AI approach that provides contrastive explanations of two machine learning models, introducing the concept of model-contrastive explanations. We propose an encoding that allows MERLIN to work with both text and tabular data and with mixed continuous and discrete features. To show the effectiveness of our approach, we evaluate it on an extensive set of benchmark datasets. MERLIN is also implemented as a python-pip package.
引用
收藏
页数:13
相关论文
共 50 条
  • [21] SLISEMAP: supervised dimensionality reduction through local explanations
    Bjorklund, Anton
    Makela, Jarmo
    Puolamaki, Kai
    MACHINE LEARNING, 2023, 112 (01) : 1 - 43
  • [22] SLISEMAP: supervised dimensionality reduction through local explanations
    Anton Björklund
    Jarmo Mäkelä
    Kai Puolamäki
    Machine Learning, 2023, 112 : 1 - 43
  • [23] Knowledge Graph Embedding in E-commerce Applications: Attentive Reasoning, Explanations, and Transferable Rules
    Zhang, Wen
    Deng, Shumin
    Chen, Mingyang
    Wang, Liang
    Chen, Qiang
    Xiong, Feiyu
    Liu, Xiangwen
    Chen, Huajun
    PROCEEDINGS OF THE 10TH INTERNATIONAL JOINT CONFERENCE ON KNOWLEDGE GRAPHS (IJCKG 2021), 2021, : 71 - 79
  • [24] Plan Explanations as Model Reconciliation - An Empirical Study
    Chakraborti, Tathagata
    Sreedharan, Sarath
    Grover, Sachin
    Kambhampati, Subbarao
    HRI '19: 2019 14TH ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, 2019, : 258 - 266
  • [25] Model-based explanations of concept drift
    Hinder, Fabian
    Vaquet, Valerie
    Brinkrolf, Johannes
    Hammer, Barbara
    NEUROCOMPUTING, 2023, 555
  • [26] PCAIME: Principal Component Analysis-Enhanced Approximate Inverse Model Explanations Through Dimensional Decomposition and Expansion
    Nakanishi, Takafumi
    IEEE ACCESS, 2024, 12 : 121093 - 121113
  • [27] ChemInformatics Model Explorer (CIME): exploratory analysis of chemical model explanations
    Humer, Christina
    Heberle, Henry
    Montanari, Floriane
    Wolf, Thomas
    Huber, Florian
    Henderson, Ryan
    Heinrich, Julian
    Streit, Marc
    JOURNAL OF CHEMINFORMATICS, 2022, 14 (01)
  • [28] ChemInformatics Model Explorer (CIME): exploratory analysis of chemical model explanations
    Christina Humer
    Henry Heberle
    Floriane Montanari
    Thomas Wolf
    Florian Huber
    Ryan Henderson
    Julian Heinrich
    Marc Streit
    Journal of Cheminformatics, 14
  • [29] BELLATREX: Building Explanations Through a LocaLly AccuraTe Rule EXtractor
    Dedja, Klest
    Nakano, Felipe Kenji
    Pliakos, Konstantinos
    Vens, Celine
    IEEE ACCESS, 2023, 11 : 41348 - 41367
  • [30] Model-Agnostic Counterfactual Explanations in Credit Scoring
    Dastile, Xolani
    Celik, Turgay
    Vandierendonck, Hans
    IEEE ACCESS, 2022, 10 : 69543 - 69554