A deep network construction that adapts to intrinsic dimensionality beyond the domain

被引:15
作者
Cloninger, Alexander [1 ,2 ]
Klock, Timo [1 ,3 ]
机构
[1] Univ Calif San Diego, Dept Math, 9500 Gilman Dr, La Jolla, CA 92093 USA
[2] Univ Calif San Diego, Haliciglu Data Sci Inst, 10100 Hopkins Dr, La Jolla, CA 92093 USA
[3] Dept Numer Anal & Sci Comp, Simula Res Lab, Martin Linges Vei 25, N-1364 Fornebu, Norway
基金
美国国家科学基金会;
关键词
Deep neural networks; Approximation theory; Curse of dimensionality; Composite functions; Noisy manifold models; MULTILAYER FEEDFORWARD NETWORKS; NEURAL-NETWORKS; OPTIMAL APPROXIMATION; BOUNDS; SMOOTH; RATES;
D O I
10.1016/j.neunet.2021.06.004
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study the approximation of two-layer compositions f(x) = g(phi(x)) via deep networks with ReLU activation, where phi is a geometrically intuitive, dimensionality reducing feature map. We focus on two intuitive and practically relevant choices for phi the projection onto a low-dimensional embedded submanifold and a distance to a collection of low-dimensional sets. We achieve near optimal approximation rates, which depend only on the complexity of the dimensionality reducing map phi rather than the ambient dimension. Since phi encapsulates all nonlinear features that are material to the function f, this suggests that deep nets are faithful to an intrinsic dimension governed by f rather than the complexity of the domain of f. In particular, the prevalent assumption of approximating functions on low-dimensional manifolds can be significantly relaxed using functions of type f(x) = g(phi(x)) with phi representing an orthogonal projection onto the same manifold. (C) 2021 Elsevier Ltd. All rights reserved.
引用
收藏
页码:404 / 419
页数:16
相关论文
共 68 条
[1]  
[Anonymous], 2008, Computer Vision and Pattern Recognition
[2]   Random Projections of Smooth Manifolds [J].
Baraniuk, Richard G. ;
Wakin, Michael B. .
FOUNDATIONS OF COMPUTATIONAL MATHEMATICS, 2009, 9 (01) :51-77
[3]   UNIVERSAL APPROXIMATION BOUNDS FOR SUPERPOSITIONS OF A SIGMOIDAL FUNCTION [J].
BARRON, AR .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1993, 39 (03) :930-945
[4]   APPROXIMATION AND ESTIMATION BOUNDS FOR ARTIFICIAL NEURAL NETWORKS [J].
BARRON, AR .
MACHINE LEARNING, 1994, 14 (01) :115-133
[5]   ON DEEP LEARNING AS A REMEDY FOR THE CURSE OF DIMENSIONALITY IN NONPARAMETRIC REGRESSION [J].
Bauer, Benedikt ;
Kohler, Michael .
ANNALS OF STATISTICS, 2019, 47 (04) :2261-2285
[6]  
Bickel Peter J, 2007, COMPLEX DATASETS INV, P177, DOI DOI 10.1214/074921707000000148
[7]   Optimal Approximation with Sparsely Connected Deep Neural Networks [J].
Boelcskei, Helmut ;
Grohs, Philipp ;
Kutyniok, Gitta ;
Petersen, Philipp .
SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2019, 1 (01) :8-45
[8]   The reach, metric distortion, geodesic convexity and the variation of tangent spaces [J].
Boissonnat J.-D. ;
Lieutier A. ;
Wintraecken M. .
Journal of Applied and Computational Topology, 2019, 3 (1-2) :29-58
[9]   Manifold Reconstruction Using Tangential Delaunay Complexes [J].
Boissonnat, Jean-Daniel ;
Ghosh, Arijit .
DISCRETE & COMPUTATIONAL GEOMETRY, 2014, 51 (01) :221-267
[10]  
Bolcskei H., 2019, ARXIV PREPRINT ARXIV