DeepVID: Deep Visual Interpretation and Diagnosis for Image Classifiers via Knowledge Distillation

被引:83
作者
Wang, Junpeng [1 ]
Gou, Liang [2 ]
Zhang, Wei [2 ]
Yang, Hao [3 ]
Shen, Han-Wei [1 ]
机构
[1] Ohio State Univ, Dept Comp Sci & Engn, Columbus, OH 43210 USA
[2] Visa Res, Data Analyt Team, Palo Alto, CA 94306 USA
[3] Visa Res, Data Analyt, Palo Alto, CA 94306 USA
关键词
Deep neural networks; model interpretation; knowledge distillation; generative model; visual analytics;
D O I
10.1109/TVCG.2019.2903943
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Deep Neural Networks (DNNs) have been extensively used in multiple disciplines due to their superior performance. However, in most cases, DNNs are considered as black-boxes and the interpretation of their internal working mechanism is usually challenging. Given that model trust is often built on the understanding of how a model works, the interpretation of DNNs becomes more important, especially in safety-critical applications (e.g., medical diagnosis, autonomous driving). In this paper, we propose DeepVID, a Deep learning approach to Visually Interpret and Diagnose DNN models, especially image classifiers. In detail, we train a small locally-faithful model to mimic the behavior of an original cumbersome DNN around a particular data instance of interest, and the local model is sufficiently simple such that it can be visually interpreted (e.g., a linear model). Knowledge distillation is used to transfer the knowledge from the cumbersome DNN to the small model, and a deep generative model (i.e., variational auto-encoder) is used to generate neighbors around the instance of interest. Those neighbors, which come with small feature variances and semantic meanings, can effectively probe the DNN's behaviors around the interested instance and help the small model to learn those behaviors. Through comprehensive evaluations, as well as case studies conducted together with deep learning experts, we validate the effectiveness of DeepVID.
引用
收藏
页码:2168 / 2180
页数:13
相关论文
共 33 条
  • [1] On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation
    Bach, Sebastian
    Binder, Alexander
    Montavon, Gregoire
    Klauschen, Frederick
    Mueller, Klaus-Robert
    Samek, Wojciech
    [J]. PLOS ONE, 2015, 10 (07):
  • [2] Chattopadhyay A., 2017, ARXIV171011063
  • [3] SUPPORT-VECTOR NETWORKS
    CORTES, C
    VAPNIK, V
    [J]. MACHINE LEARNING, 1995, 20 (03) : 273 - 297
  • [4] Robust Physical-World Attacks on Deep Learning Visual Classification
    Eykholt, Kevin
    Evtimov, Ivan
    Fernandes, Earlence
    Li, Bo
    Rahmati, Amir
    Xiao, Chaowei
    Prakash, Atul
    Kohno, Tadayoshi
    Song, Dawn
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 1625 - 1634
  • [5] Interpretable Explanations of Black Boxes by Meaningful Perturbation
    Fong, Ruth C.
    Vedaldi, Andrea
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 3449 - 3457
  • [6] An Interactive Node-Link Visualization of Convolutional Neural Networks
    Harley, Adam W.
    [J]. ADVANCES IN VISUAL COMPUTING, PT I (ISVC 2015), 2015, 9474 : 867 - 877
  • [7] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [8] Hinton Geoffrey, 2015, Distilling the knowledge in a neural network
  • [9] Deep Feature Consistent Variational Autoencoder
    Hou, Xianxu
    Shen, Linlin
    Sun, Ke
    Qiu, Guoping
    [J]. 2017 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2017), 2017, : 1133 - 1141
  • [10] GAN Lab: Understanding Complex Deep Generative Models using Interactive Visual Experimentation
    Kahng, Minsuk
    Thorat, Nikhil
    Chau, Duen Horng
    Viegas, Fernanda B.
    Wattenberg, Martin
    [J]. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2019, 25 (01) : 310 - 320