Optimal deep neural networks by maximization of the approximation power

被引:6
作者
Calvo-Pardo, Hector [1 ,2 ,3 ,4 ]
Mancini, Tullio [1 ]
Olmo, Jose [1 ,5 ,6 ]
机构
[1] Univ Southampton, Southampton, England
[2] CFS, Osnabruck, Germany
[3] CPC, London, England
[4] ILB, Paris, France
[5] Univ Zaragoza, Zaragoza, Spain
[6] Univ Southampton, Dept Econ, Highfield Campus, Southampton SO17 1BJ, England
关键词
Machine learning; Artificial intelligence; Data science; Forecasting; Feedforward neural networks; PERFORMANCE;
D O I
10.1016/j.cor.2023.106264
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
We propose an optimal architecture for deep neural networks of given size. The optimal architecture obtains from maximizing the lower bound of the maximum number of linear regions approximated by a deep neural network with a ReLu activation function. The accuracy of the approximation function relies on the neural network structure characterized by the number, dependence and hierarchy between the nodes within and across layers. We show how the accuracy of the approximation improves as we optimally choose the width and depth of the network. A Monte-Carlo simulation exercise illustrates the outperformance of the optimized architecture against cross-validation methods and gridsearch for linear and nonlinear prediction models. The application of this methodology to the Boston Housing dataset confirms empirically the outperformance of our method against state-of the-art machine learning models.
引用
收藏
页数:15
相关论文
共 59 条
[31]  
Lu Z, 2017, ADV NEUR IN, V30
[32]   Multilayer neural networks:: an experimental evaluation of on-line training methods [J].
Martí, R ;
El-Fallahi, A .
COMPUTERS & OPERATIONS RESEARCH, 2004, 31 (09) :1491-1513
[33]   A mean field view of the landscape of two-layer neural networks [J].
Mei, Song ;
Montanari, Andrea ;
Phan-Minh Nguyen .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2018, 115 (33) :E7665-E7671
[34]  
Mellor H., 2021, arXiv
[35]  
Montufar G.G., 2017, Notes on the Number of Linear Regions of Deep Neural Networks. SampTA
[36]  
Montúfar G, 2014, ADV NEUR IN, V27
[37]  
MYSHKOV P, 2016, WORKSH BAYES DEEP LE
[38]  
Nado Z, 2018, INT C LEARN REPR WOR
[39]  
Nair V., 2010, P 27 INT C MACH LEAR, P807
[40]   A SIMPLE, POSITIVE SEMIDEFINITE, HETEROSKEDASTICITY AND AUTOCORRELATION CONSISTENT COVARIANCE-MATRIX [J].
NEWEY, WK ;
WEST, KD .
ECONOMETRICA, 1987, 55 (03) :703-708