Algorithms to estimate Shapley value feature attributions

被引:111
作者
Chen, Hugh [1 ]
Covert, Ian C. [1 ]
Lundberg, Scott M. [2 ]
Lee, Su-In [1 ]
机构
[1] Univ Washington, Paul G Allen Sch Comp Sci & Engn, Seattle, WA 98195 USA
[2] Microsoft Res, New York, NY USA
基金
美国国家科学基金会; 美国国家卫生研究院;
关键词
EXPLAINABLE AI; GAME; CLASSIFICATIONS; LEVEL;
D O I
10.1038/s42256-023-00657-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Feature attributions based on the Shapley value are popular for explaining machine learning models. However, their estimation is complex from both theoretical and computational standpoints. We disentangle this complexity into two main factors: the approach to removing feature information and the tractable estimation strategy. These two factors provide a natural lens through which we can better understand and compare 24 distinct algorithms. Based on the various feature-removal approaches, we describe the multiple types of Shapley value feature attributions and the methods to calculate each one. Then, based on the tractable estimation strategies, we characterize two distinct families of approaches: model-agnostic and model-specific approximations. For the model-agnostic approximations, we benchmark a wide class of estimation approaches and tie them to alternative yet equivalent characterizations of the Shapley value. For the model-specific approximations, we clarify the assumptions crucial to each method's tractability for linear, tree and deep models. Finally, we identify gaps in the literature and promising future research directions. There are numerous algorithms for generating Shapley value explanations. The authors provide a comprehensive survey of Shapley value feature attribution algorithms by disentangling and clarifying the fundamental challenges underlying their computation.
引用
收藏
页码:590 / 601
页数:12
相关论文
共 75 条
  • [1] Explaining individual predictions when features are dependent: More accurate approximations to Shapley values
    Aas, Kjersti
    Jullum, Martin
    Loland, Anders
    [J]. ARTIFICIAL INTELLIGENCE, 2021, 298
  • [2] Explaining predictive models using Shapley values and non-parametric vine copulas
    Aas, Kjersti
    Nagler, Thomas
    Jullum, Martin
    Loland, Anders
    [J]. DEPENDENCE MODELING, 2021, 9 (01): : 62 - 81
  • [3] Ancona M, 2019, PR MACH LEARN RES, V97
  • [4] Benard C., 2022, PMLR, P5563
  • [5] Layer-Wise Relevance Propagation for Neural Networks with Local Renormalization Layers
    Binder, Alexander
    Montavon, Gregoire
    Lapuschkin, Sebastian
    Mueller, Klaus-Robert
    Samek, Wojciech
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2016, PT II, 2016, 9887 : 63 - 71
  • [6] Random forests
    Breiman, L
    [J]. MACHINE LEARNING, 2001, 45 (01) : 5 - 32
  • [7] Improving polynomial estimation of the Shapley value by stratified random sampling with optimum allocation
    Castro, Javier
    Gomez, Daniel
    Molina, Elisenda
    Tejada, Juan
    [J]. COMPUTERS & OPERATIONS RESEARCH, 2017, 82 : 180 - 188
  • [8] Polynomial calculation of the Shapley value based on sampling
    Castro, Javier
    Gomez, Daniel
    Tejada, Juan
    [J]. COMPUTERS & OPERATIONS RESEARCH, 2009, 36 (05) : 1726 - 1730
  • [9] Charnes A., 1988, ECONOMETRICS PLANNIN, P123, DOI DOI 10.1007/978-94-009-3677-5_7
  • [10] Chen H, 2020, PREPRINT