Approximation Spaces of Deep Neural Networks

被引:56
作者
Gribonval, Remi [1 ]
Kutyniok, Gitta [2 ]
Nielsen, Morten [3 ]
Voigtlaender, Felix [4 ]
机构
[1] Univ Lyon, EnsL, UCBL, CNRS,Inria,LIP, F-69342 Lyon 07, France
[2] Ludwig Maximilians Univ Munchen, Dept Math, Munich, Germany
[3] Aalborg Univ, Dept Math Sci, Aalborg, Denmark
[4] Katholische Univ Eichstatt Ingolstadt, Dept Comp Sci, Eichstatt, Germany
基金
欧盟地平线“2020”;
关键词
Deep neural networks; Sparsely connected networks; Approximation spaces; Besov spaces; Direct estimates; Inverse estimates; Piecewise polynomials; ReLU activation function; MULTILAYER FEEDFORWARD NETWORKS; BOUNDS; SMOOTH;
D O I
10.1007/s00365-021-09543-4
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
We study the expressivity of deep neural networks. Measuring a network's complexity by its number of connections or by its number of neurons, we consider the class of functions for which the error of best approximation with networks of a given complexity decays at a certain rate when increasing the complexity budget. Using results from classical approximation theory, we show that this class can be endowed with a (quasi)-norm that makes it a linear function space, called approximation space. We establish that allowing the networks to have certain types of "skip connections" does not change the resulting approximation spaces. We also discuss the role of the network's nonlinearity (also known as activation function) on the resulting spaces, as well as the role of depth. For the popular ReLU nonlinearity and its powers, we relate the newly constructed spaces to classical Besov spaces. The established embeddings highlight that some functions of very low Besov smoothness can nevertheless be well approximated by neural networks, if these networks are sufficiently deep.
引用
收藏
页码:259 / 367
页数:109
相关论文
共 69 条
[1]  
Adams R., 2003, Sobolev Spaces
[2]   Solving ill-posed inverse problems using iterative deep neural networks [J].
Adler, Jonas ;
Oktem, Ozan .
INVERSE PROBLEMS, 2017, 33 (12)
[3]   Generalized approximation spaces and applications [J].
Almira, JM ;
Luther, U .
MATHEMATISCHE NACHRICHTEN, 2004, 263 :3-35
[4]  
[Anonymous], 2017, ARXIV
[5]  
[Anonymous], 1998, Thesis
[6]   UNIVERSAL APPROXIMATION BOUNDS FOR SUPERPOSITIONS OF A SIGMOIDAL FUNCTION [J].
BARRON, AR .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1993, 39 (03) :930-945
[7]   APPROXIMATION AND ESTIMATION BOUNDS FOR ARTIFICIAL NEURAL NETWORKS [J].
BARRON, AR .
MACHINE LEARNING, 1994, 14 (01) :115-133
[8]  
Beck C., 2018, ARXIV PREPRINT ARXIV
[9]   Optimal Approximation with Sparsely Connected Deep Neural Networks [J].
Boelcskei, Helmut ;
Grohs, Philipp ;
Kutyniok, Gitta ;
Petersen, Philipp .
SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2019, 1 (01) :8-45
[10]   Learning the invisible: a hybrid deep learning-shearlet framework for limited angle computed tomography [J].
Bubba, Tatiana A. ;
Kutyniok, Gitta ;
Lasses, Matti ;
Maerz, Maximilian ;
Samek, Wojciech ;
Siltanen, Samuli ;
Srinivasan, Vignesh .
INVERSE PROBLEMS, 2019, 35 (06)