Interpreting deep learning models with marginal attribution by conditioning on quantiles

被引:0
作者
Michael Merz
Ronald Richman
Andreas Tsanakas
Mario V. Wüthrich
机构
[1] University of Hamburg,Faculty of Business Administration
[2] Old Mutual Insure,Bayes Business School
[3] University of the Witwatersrand,RiskLab, Department of Mathematics
[4] City,undefined
[5] University of London,undefined
[6] ETH Zurich,undefined
来源
Data Mining and Knowledge Discovery | 2022年 / 36卷
关键词
Explainable AI (XAI); Model-agnostic tools; Deep learning; Attribution; Accumulated local effects (ALE); Partial dependence plot (PDP); Locally interpretable model-agnostic explanation (LIME); Variable importance; Post-hoc analysis; Interaction;
D O I
暂无
中图分类号
学科分类号
摘要
A vast and growing literature on explaining deep learning models has emerged. This paper contributes to that literature by introducing a global gradient-based model-agnostic method, which we call Marginal Attribution by Conditioning on Quantiles (MACQ). Our approach is based on analyzing the marginal attribution of predictions (outputs) to individual features (inputs). Specifically, we consider variable importance by fixing (global) output levels, and explaining how features marginally contribute to these fixed global output levels. MACQ can be seen as a marginal attribution counterpart to approaches such as accumulated local effects, which study the sensitivities of outputs by perturbing inputs. Furthermore, MACQ allows us to separate marginal attribution of individual features from interaction effects and to visualize the 3-way relationship between marginal attribution, output level, and feature value.
引用
收藏
页码:1335 / 1370
页数:35
相关论文
共 40 条
  • [1] Acerbi C(2002)Spectral measures of risk: a coherent representation of subjective risk aversion J Bank Finance 7 1505-1518
  • [2] Apley DW(2020)Visualizing the effects of predictor variables in black box supervised learning models J R Stat Soc Ser B 82 1059-1086
  • [3] Zhu J(2003)A neural probabilistic language model J Mach Learn Res 3 1137-1155
  • [4] Bengio Y(2001)Random forests Mach Learn 45 5-32
  • [5] Ducharme R(2020)Prediction, estimation and attribution Int Stat Rev 88 S28-S59
  • [6] Vincent P(2014)Event labeling combining ensemble detectors and background knowledge Prog Artif Intell 2 113-127
  • [7] Jauvin C(2001)Greedy function approximation: a gradient boosting machine Ann Stat 29 1189-1232
  • [8] Breiman L(2008)Predictive learning via rule ensembles Ann Appl Stat 2 916-954
  • [9] Efron B(2015)Peeking inside the black box: visualizing statistical learning with plots of individual conditional expectation J Comput Graph Stat 24 44-65
  • [10] Fanaee-T H(2000)Sensitivity analysis of values at risk J Empir Finance 7 225-245