Approximating smooth functions by deep neural networks with sigmoid activation function

被引:44
作者
Langer, Sophie [1 ]
机构
[1] Tech Univ Darmstadt, Fachbereich Math, Schlossgartenstr 7, D-64289 Darmstadt, Germany
关键词
Deep learning; Full connectivity; Neural networks; Uniform approximation;
D O I
10.1016/j.jmva.2020.104696
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We study the power of deep neural networks (DNNs) with sigmoid activation function. Recently, it was shown that DNNs approximate any d-dimensional, smooth function on a compact set with a rate of order W-p/d, where W is the number of nonzero weights in the network and p is the smoothness of the function. Unfortunately, these rates only hold for a special class of sparsely connected DNNs. We ask ourselves if we can show the same approximation rate for a simpler and more general class, i.e., DNNs which are only defined by its width and depth. In this article we show that DNNs with fixed depth and a width of order M-d achieve an approximation rate of M-2p. As a conclusion we quantitatively characterize the approximation power of DNNs in terms of the overall weights W-0 in the network and show an approximation rate of W-0(-p/d). This more general result finally helps us to understand which network topology guarantees a special target accuracy. (C) 2020 Elsevier Inc. All rights reserved.
引用
收藏
页数:21
相关论文
共 21 条
  • [21] YAROTSKY D, 2019, ARXIV190609477