Representer Point Selection via Local Jacobian Expansion for Post-hoc Classifier Explanation of Deep Neural Networks and Ensemble Models

被引:0
作者
Sui, Yi [1 ]
Wu, Ga [1 ,2 ]
Sanner, Scott [1 ,3 ]
机构
[1] Univ Toronto, Toronto, ON, Canada
[2] Borealis AI, Toronto, ON, Canada
[3] Vector Inst Artificial Intelligence, Toronto, ON, Canada
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021) | 2021年 / 34卷
基金
加拿大自然科学与工程研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Explaining the influence of training data on machine learning model predictions is a critical tool for debugging models through data curation. A recent appealing and efficient approach for this task was provided via the concept of Representer Point Selection (RPS), i.e. a method the leverages the dual form of l(2) regularized optimization in the last layer of the neural network to identify the contribution of training points to the prediction. However, two key drawbacks of RPS-l(2) are that they (i) lead to disagreement between the originally trained network and the RPS-l(2) regularized network modification and (ii) often yield a static ranking of training data for test points in the same class, independent of the test point being classified. Inspired by the RPS-l(2) approach, we propose an alternative method based on a local Jacobian Taylor expansion (LJE). We empirically compared RPS-LJE with the original RPS-l(2) on image classification (with ResNet), text classification recurrent neural networks (with Bi-LSTM), and tabular classification (with XGBoost) tasks. Quantitatively, we show that RPS-LJE slightly outperforms RPS-l(2) and other state-of-the-art data explanation methods by up to 3% on a data debugging task. More critically, we qualitatively observe that RPS-LJE provides stable and individualized explanations that are more coherent to each test data point. Overall, RPS-LJE represents a novel approach to RPS-l(2) that provides a powerful tool for sample-based model explanation and debugging.
引用
收藏
页数:12
相关论文
共 23 条
  • [1] PROTOTYPE SELECTION FOR INTERPRETABLE CLASSIFICATION
    Bien, Jacob
    Tibshirani, Robert
    [J]. ANNALS OF APPLIED STATISTICS, 2011, 5 (04) : 2403 - 2424
  • [2] Chen T., 2015, R package version 0.4-2, V1, P1, DOI DOI 10.1145/2939672.2939785
  • [3] Dabkowski P, 2017, ADV NEUR IN, V30
  • [4] Dua D, 2017, UCI MACHINE LEARNING, DOI DOI 10.1016/J.DSS.2009.05.016
  • [5] Interpretable Explanations of Black Boxes by Meaningful Perturbation
    Fong, Ruth C.
    Vedaldi, Andrea
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 3449 - 3457
  • [6] Graves A, 2012, STUD COMPUT INTELL, V385, P1, DOI [10.1162/neco.1997.9.1.1, 10.1007/978-3-642-24797-2]
  • [7] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [8] Khanna R., 2019, AISTATS
  • [9] Kim B, 2016, ADV NEUR IN, V29
  • [10] Kim H, 2014, IEEE INT SYMP INFO, P1952, DOI 10.1109/ISIT.2014.6875174