Improving the Interpretability of Deep Neural Networks with Knowledge Distillation

被引:79
作者
Liu, Xuan [1 ]
Wang, Xiaoguang [1 ,2 ]
Matwin, Stan [1 ,3 ]
机构
[1] Dalhousie Univ, Fac Comp Sci, Inst Big Data Analyt, Halifax, NS, Canada
[2] Alibaba Grp, Hangzhou, Zhejiang, Peoples R China
[3] Polish Acad Sci, Inst Comp Sci, Warsaw, Poland
来源
2018 18TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW) | 2018年
关键词
interpretation; Neural Networks; Decision Tree; TensorFlow; dark knowledge; knowledge distillation;
D O I
10.1109/ICDMW.2018.00132
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep Neural Networks have achieved huge success at a wide spectrum of applications from language modeling, computer vision to speech recognition. However, nowadays, good performance alone is not enough to satisfy the needs of practical deployment where interpretability is demanded for cases involving ethics and mission critical applications. The complex models of Deep Neural Networks make it hard to understand and reason the predictions, which hinders its further progress. To tackle this problem, we apply the Knowledge Distillation technique to distill Deep Neural Networks into decision trees in order to attain good performance and interpretability simultaneously. We formulate the problem at hand as a multi-output regression problem and the experiments demonstrate that the student model achieves significantly better accuracy performance (about 1% to 5%) than vanilla decision trees at the same level of tree depth. The experiments are implemented on the TensorFlow platform to make it scalable to big datasets. To the best of our knowledge, we are the first to distill Deep Neural Networks into vanilla decision trees on multi-class datasets.
引用
收藏
页码:905 / 912
页数:8
相关论文
共 62 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]  
[Anonymous], NIPS DEEP LEARN WORK
[3]  
[Anonymous], 2017, KDD 17 WORKSH
[4]  
[Anonymous], 1993, MORGAN KAUFMANN SERI
[5]  
[Anonymous], ICML 2018
[6]  
[Anonymous], 2018, AAAI C ART INT
[7]  
[Anonymous], AAAI ACM AIES 2018
[8]  
[Anonymous], 31 INT C MACH LEARN
[9]  
[Anonymous], 2018 INT JOINT C NEU
[10]  
[Anonymous], 2017, DIGITAL SIGNAL PROCE