Post-hoc Counterfactual Generation with Supervised Autoencoder

被引:3
作者
Guyomard, Victor [1 ,2 ]
Fessant, Francoise [1 ]
Bouadi, Tassadit [2 ]
Guyet, Thomas [3 ]
机构
[1] Orange Labs, Lannion, France
[2] Univ Rennes, IRISA, CNRS, INRIA, Rennes, France
[3] Inst Agro IRISA UMR6074, Rennes, France
来源
MACHINE LEARNING AND PRINCIPLES AND PRACTICE OF KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2021, PT I | 2021年 / 1524卷
关键词
Counterfactual explanation; Interpretability; Prototypes;
D O I
10.1007/978-3-030-93736-2_10
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Nowadays, AI is increasingly being used in many fields to automate decisions that largely affect the daily lives of humans. The inherent complexity of these systems makes them so-called black-box models. Explainable Artificial Intelligence (XAI) aims at solving this issue by providing methods to overcome this lack of transparency. Counterfactual explanation is a common and well-known class of explanations that produces actionable and understandable explanations for end-users. However, generating realistic and useful counterfactuals remains a challenge. In this work, we investigate the problem of generating counterfactuals that are both close to the data distribution, and to the distribution of the target class. Our objective is to obtain counterfactuals with likely values (i.e. realistic). We propose a model agnostic method for generating realistic counterfactuals by using class prototypes. The novelty of this approach is that these class prototypes are obtained using a supervised auto-encoder. Then, we performed an empirical evaluation across several interpretability metrics, that shows competitive results with the state-of-the-art method.
引用
收藏
页码:105 / 114
页数:10
相关论文
共 10 条
[1]  
[Anonymous], 2017, ABS171100399 CORR
[2]  
Dhurandhar A, 2018, ADV NEUR IN, V31
[3]   NONLINEAR PRINCIPAL COMPONENT ANALYSIS USING AUTOASSOCIATIVE NEURAL NETWORKS [J].
KRAMER, MA .
AICHE JOURNAL, 1991, 37 (02) :233-243
[4]  
Labaien J., ARXIV210409062
[5]  
Le L, 2018, ADV NEUR IN, V31
[6]  
Li O, 2018, AAAI CONF ARTIF INTE, P3530
[7]  
Looveren A.V., 2020, ARXIV190702584
[8]   Explanation in artificial intelligence: Insights from the social sciences [J].
Miller, Tim .
ARTIFICIAL INTELLIGENCE, 2019, 267 :1-38
[9]  
Nemirovsky D., 2020, ARXIV200905199
[10]  
Sharma A., 2019, P WORKSH MICR NIPS C