Explanations for Neural Networks by Neural Networks

被引:9
作者
Marton, Sascha [1 ]
Luedtke, Stefan [1 ]
Bartelt, Christian [1 ]
机构
[1] Univ Mannheim, Inst Enterprise Syst, D-68131 Mannheim, Germany
来源
APPLIED SCIENCES-BASEL | 2022年 / 12卷 / 03期
关键词
Explainable AI (XAI); interpretability; explainability; neural networks; machine learning; symbolic representations;
D O I
10.3390/app12030980
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Understanding the function learned by a neural network is crucial in many domains, e.g., to detect a model's adaption to concept drift in online learning. Existing global surrogate model approaches generate explanations by maximizing the fidelity between the neural network and a surrogate model on a sample-basis, which can be very time-consuming. Therefore, these approaches are not applicable in scenarios where timely or frequent explanations are required. In this paper, we introduce a real-time approach for generating a symbolic representation of the function learned by a neural network. Our idea is to generate explanations via another neural network (called the Interpretation Network, or I-Net), which maps network parameters to a symbolic representation of the network function. We show that the training of an I-Net for a family of functions can be performed up-front and subsequent generation of an explanation only requires querying the I-Net once, which is computationally very efficient and does not require training data. We empirically evaluate our approach for the case of low-order polynomials as explanations, and show that it achieves competitive results for various data and function complexities. To the best of our knowledge, this is the first approach that attempts to learn mapping from neural networks to symbolic representations.
引用
收藏
页数:14
相关论文
共 25 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]  
Alaa AM, 2019, ADV NEUR IN, V32
[3]  
Andoni A, 2014, PR MACH LEARN RES, V32, P1908
[4]  
Beals R., 2013, NOTICES OF THE AMS, V60, P866
[5]  
Fisher A, 2019, J MACH LEARN RES, V20
[6]   Greedy function approximation: A gradient boosting machine [J].
Friedman, JH .
ANNALS OF STATISTICS, 2001, 29 (05) :1189-1232
[7]   PREDICTIVE LEARNING VIA RULE ENSEMBLES [J].
Frieman, Jerome H. ;
Popescu, Bogdan E. .
ANNALS OF APPLIED STATISTICS, 2008, 2 (03) :916-954
[8]  
Frosst Nicholas., 2017, CORR
[9]  
Glorot X., P 14 INT C ARTIFICIA, DOI DOI 10.1177/1753193410395357
[10]  
Greenwell B.M., 2018, 180504755 ARXIV