Explaining the Performance of Black Box Regression Models

被引:4
作者
Areosa, Ines [1 ]
Torgo, Luis [2 ]
机构
[1] INESC TEC, LIAAD, Porto, Portugal
[2] Dalhousie Univ, Fac Comp Sci, Halifax, NS, Canada
来源
2019 IEEE INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (DSAA 2019) | 2019年
关键词
D O I
10.1109/DSAA.2019.00025
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The widespread usage of Machine Learning and Data Mining models in several key areas of our societies has raised serious concerns in terms of accountability and ability to justify and interpret the decisions of these models. This is even more relevant when models are too complex and often regarded as black boxes. In this paper we present several tools designed to help in understanding and explaining the reasons for the observed predictive performance of black box regression models. We describe, evaluate and propose several variants of Error Dependence Plots. These plots provide a visual display of the expected relationship between the prediction error of any model and the values of a predictor variable. They allow the end user to understand what to expect from the models given some concrete values of the predictor variables. These tools allow more accurate explanations on the conditions that may lead to some failures of the models. Moreover, our proposed extensions also provide a multivariate perspective of this analysis, and the ability to compare the behaviour of multiple models under different conditions. This comparative analysis empowers the end user with the ability to have a case-based analysis of the risks associated with different models, and thus select the model with lower expected risk for each test case, or even decide not to use any model because the expected error is unacceptable.
引用
收藏
页码:110 / 118
页数:9
相关论文
共 22 条
[1]  
[Anonymous], 2011, INT C MACHINE LEARNI
[2]  
[Anonymous], 2011, e1071: Misc Functions of the Department of Statistics (e1071)
[3]  
[Anonymous], 2018, Model class reliance: Variable importance measures for any machine learning model class, from the 'rashomon' perspective'
[4]  
Areosa L., 2019, P 19 EPIA C ART INT
[5]  
Bi J, 2003, P 20 INT C MACH LEAR, P43
[6]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[7]   Accountability in Algorithmic Decision Making [J].
Diakopoulos, Nicholas .
COMMUNICATIONS OF THE ACM, 2016, 59 (02) :56-62
[8]   Greedy function approximation: A gradient boosting machine [J].
Friedman, JH .
ANNALS OF STATISTICS, 2001, 29 (05) :1189-1232
[9]   Peeking Inside the Black Box: Visualizing Statistical Learning With Plots of Individual Conditional Expectation [J].
Goldstein, Alex ;
Kapelner, Adam ;
Bleich, Justin ;
Pitkin, Emil .
JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2015, 24 (01) :44-65
[10]   ROC curves for regression [J].
Hernandez-Orallo, Jose .
PATTERN RECOGNITION, 2013, 46 (12) :3395-3411