Interpretable generalized additive neural networks

被引:23
作者
Kraus, Mathias [1 ]
Tschernutter, Daniel [2 ]
Weinzierl, Sven [1 ]
Zschech, Patrick [1 ]
机构
[1] FAU Erlangen Nurnberg, Lange Gasse 20, D-90403 Nurnberg, Germany
[2] Swiss Fed Inst Technol, Weinbergstr 56-58, CH-8092 Zurich, Switzerland
关键词
Analytics; Generalized additive models; Gradient boosting; Interpretable machine learning; Neural networks; LOGISTIC-REGRESSION; MACHINE; ANALYTICS;
D O I
10.1016/j.ejor.2023.06.032
中图分类号
C93 [管理学];
学科分类号
12 ; 1201 ; 1202 ; 120202 ;
摘要
We propose Interpretable Generalized Additive Neural Networks (IGANN), a novel machine learning model that uses gradient boosting and tailored neural networks to obtain high predictive performance while being interpretable to humans. We derive an efficient training algorithm based on the theory of extreme learning machines, that allows reducing the training process to solving a sequence of regularized linear regressions. We analyze the algorithm theoretically, provide insights into the rate of change of so-called shape functions, and show that the computational complexity of the training process scales linearly with the number of samples in the training dataset. We implement IGANN in PyTorch, which allows the model to be trained on graphics processing units (GPUs) to speed up training. We demonstrate favorable results in a variety of numerical experiments and showcase IGANN's value in three real-world case studies for productivity prediction, credit scoring, and criminal recidivism prediction. (c) 2023 The Author(s). Published by Elsevier B.V. This is an open access article under the CC BY license ( http://creativecommons.org/licenses/by/4.0/ )
引用
收藏
页码:303 / 316
页数:14
相关论文
共 56 条
[1]  
Agarwal Rishabh, 2021, Advances in Neural Information Processing Systems, V34
[2]  
Al Imran A, 2019, INT C CONTROL DECISI, P1402, DOI [10.1109/codit.2019.8820486, 10.1109/CoDIT.2019.8820486]
[3]  
Al-Ebbini L, 2017, IEEE JORDAN CONF APP
[4]  
Angwin Julia., 2016, Machine bias: Risk assessments in criminal sentencing
[5]  
[Anonymous], 2013, J. Mach. Learn. Res.
[6]  
Badirli S, 2020, Arxiv, DOI arXiv:2002.07971
[7]   Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI [J].
Barredo Arrieta, Alejandro ;
Diaz-Rodriguez, Natalia ;
Del Ser, Javier ;
Bennetot, Adrien ;
Tabik, Siham ;
Barbado, Alberto ;
Garcia, Salvador ;
Gil-Lopez, Sergio ;
Molina, Daniel ;
Benjamins, Richard ;
Chatila, Raja ;
Herrera, Francisco .
INFORMATION FUSION, 2020, 58 :82-115
[8]   Explainable models of credit losses [J].
Bastos, Joao A. ;
Matos, Sara M. .
EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2022, 301 (01) :386-394
[9]   Extending business failure prediction models with textual website content using deep learning [J].
Borchert, Philipp ;
Coussement, Kristof ;
De Caigny, Arno ;
De Weerdt, Jochen .
EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2023, 306 (01) :348-357
[10]   Application of machine learning techniques for supply chain demand forecasting [J].
Carbonneau, Real ;
Laframboise, Kevin ;
Vahidov, Rustam .
EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2008, 184 (03) :1140-1154