TimeSHAP: Explaining Recurrent Models through Sequence Perturbations

被引:46
作者
Bento, Joao [1 ,2 ]
Saleiro, Pedro [1 ]
Cruz, Andre F. [1 ]
Figueiredo, Mario A. T. [2 ,3 ]
Bizarro, Pedro [1 ]
机构
[1] Feedzai, San Mateo, CA 94402 USA
[2] ULisboa, Inst Super Tecn, Lisbon, Portugal
[3] Inst Telecomunicacoes, Lisbon, Portugal
来源
KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING | 2021年
关键词
TimeSHAP; RNN; SHAP; Shapley Values; XAI; Explainability; FRAUD;
D O I
10.1145/3447548.3467166
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Although recurrent neural networks (RNNs) are state-of-the-art in numerous sequential decision-making tasks, there has been little research on explaining their predictions. In this work, we present TimeSHAP, a model-agnostic recurrent explainer that builds upon KernelSHAP and extends it to the sequential domain. TimeSHAP computes feature-, timestep-, and cell-level attributions. As sequences may be arbitrarily long, we further propose a pruning method that is shown to dramatically decrease both its computational cost and the variance of its attributions. We use TimeSHAP to explain the predictions of a real-world bank account takeover fraud detection RNN model, and draw key insights from its explanations: i) the model identifies important features and events aligned with what fraud analysts consider cues for account takeover; ii) positive predicted sequences can be pruned to only 10% of the original length, as older events have residual attribution values; iii) the most recent input event of positive predictions only contributes on average to 41% of the model's score; iv) notably high attribution to client's age, suggesting a potential discriminatory reasoning, later confirmed as higher false positive rates for older clients.
引用
收藏
页码:2565 / 2573
页数:9
相关论文
共 36 条
[1]   Fraud detection system: A survey [J].
Abdallah, Aisha ;
Maarof, Mohd Aizaini ;
Zainal, Anazida .
JOURNAL OF NETWORK AND COMPUTER APPLICATIONS, 2016, 68 :90-113
[2]   On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation [J].
Bach, Sebastian ;
Binder, Alexander ;
Montavon, Gregoire ;
Klauschen, Frederick ;
Mueller, Klaus-Robert ;
Samek, Wojciech .
PLOS ONE, 2015, 10 (07)
[3]  
BENGIO Y, 1993, 1993 IEEE INTERNATIONAL CONFERENCE ON NEURAL NETWORKS, VOLS 1-3, P1183, DOI 10.1109/ICNN.1993.298725
[4]  
Brown C.E., 1998, Springer eBooks, P155, DOI [10.1007/978-3-642-80328-4_13, DOI 10.1007/978-3-642-80328-4_13, DOI 10.1007/978-3-642-80328-4]
[5]  
Cho K., 2014, P 8 WORKSH SYNT SEM, DOI [10.3115/v1/W14-4012, DOI 10.3115/V1/W14-4012]
[6]  
Choi E, 2016, ADV NEUR IN, V29
[7]  
Denil Misha, 2014, ARXIV14126815
[8]  
Federal Bureau of Investigation, 2016, CRED CARD FRAUD
[9]   A Survey of Methods for Explaining Black Box Models [J].
Guidotti, Riccardo ;
Monreale, Anna ;
Ruggieri, Salvatore ;
Turin, Franco ;
Giannotti, Fosca ;
Pedreschi, Dino .
ACM COMPUTING SURVEYS, 2019, 51 (05)
[10]   Interpreting a recurrent neural network's predictions of ICU mortality risk [J].
Ho, Long, V ;
Aczon, Melissa ;
Ledbetter, David ;
Wetzel, Randall .
JOURNAL OF BIOMEDICAL INFORMATICS, 2021, 114