Sobolev-Type Embeddings for Neural Network Approximation Spaces

被引:0
作者
Philipp Grohs
Felix Voigtlaender
机构
[1] University of Vienna,Faculty of Mathematics
[2] Research Network Data Science @ Uni Vienna,Department of Mathematics
[3] Johann Radon Institute,undefined
[4] Technical University of Munich,undefined
来源
Constructive Approximation | 2023年 / 57卷
关键词
Deep neural networks; Approximation spaces; Hölder spaces; Embedding theorems; Optimal learning algorithms; Primary: 68T07; 46E35; Secondary: 65D05; 46E30;
D O I
暂无
中图分类号
学科分类号
摘要
We consider neural network approximation spaces that classify functions according to the rate at which they can be approximated (with error measured in Lp\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L^p$$\end{document}) by ReLU neural networks with an increasing number of coefficients, subject to bounds on the magnitude of the coefficients and the number of hidden layers. We prove embedding theorems between these spaces for different values of p. Furthermore, we derive sharp embeddings of these approximation spaces into Hölder spaces. We find that, analogous to the case of classical function spaces (such as Sobolev spaces, or Besov spaces) it is possible to trade “smoothness” (i.e., approximation rate) for increased integrability. Combined with our earlier results in Grohs and Voigtlaender (Proof of the theory-to-practice gap in deep learning via sampling complexity bounds for neural network approximation spaces, 2021. arXiv preprint arXiv:2104.02746), our embedding theorems imply a somewhat surprising fact related to “learning” functions from a given neural network space based on point samples: if accuracy is measured with respect to the uniform norm, then an optimal “learning” algorithm for reconstructing functions that are well approximable by ReLU neural networks is simply given by piecewise constant interpolation on a tensor product grid.
引用
收藏
页码:579 / 599
页数:20
相关论文
共 50 条
  • [21] On approximation spaces and Greedy-type bases
    Berna, Pablo M.
    Chu, Hung Viet
    Hernandez, Eugenio
    ANNALS OF FUNCTIONAL ANALYSIS, 2025, 16 (01)
  • [22] Moser-type trace inequalities for generalized Lorentz–Sobolev spaces
    Robert Černý
    Revista Matemática Complutense, 2015, 28 : 303 - 357
  • [23] MOSER-TYPE INEQUALITIES FOR GENERALIZED LORENTZ-SOBOLEV SPACES
    Cerny, Robert
    Gurka, Petr
    HOUSTON JOURNAL OF MATHEMATICS, 2014, 40 (04): : 1225 - 1269
  • [24] Wavelet Approximation in Weighted Sobolev Spaces of Mixed Order with Applications to the Electronic Schrödinger Equation
    Andreas Zeiser
    Constructive Approximation, 2012, 35 : 293 - 322
  • [25] Commutators for approximation spaces and Marcinkiewicz-type multipliers
    Cerdà, J
    Martín, J
    Krugljak, NY
    JOURNAL OF APPROXIMATION THEORY, 1999, 100 (02) : 251 - 265
  • [26] Memory-optimal neural network approximation
    Bolcskei, Helmut
    Grohs, Philipp
    Kutyniok, Gitta
    Petersen, Philipp
    WAVELETS AND SPARSITY XVII, 2017, 10394
  • [27] SignReLU neural network and its approximation ability
    Li, Jianfei
    Feng, Han
    Zhou, Ding-Xuan
    JOURNAL OF COMPUTATIONAL AND APPLIED MATHEMATICS, 2024, 440
  • [28] Simultaneous neural network approximation for smooth functions
    Hon, Sean
    Yang, Haizhao
    NEURAL NETWORKS, 2022, 154 : 152 - 164
  • [29] Sharpness and non-compactness of embeddings of Bessel-potential-type spaces
    Gogatishvili, Amiran
    Neves, Julio Severino
    Opic, Bohumir
    MATHEMATISCHE NACHRICHTEN, 2007, 280 (9-10) : 1083 - 1093
  • [30] Deep Neural Network Embeddings for Text-Independent Speaker Verification
    Snyder, David
    Garcia-Romero, Daniel
    Povey, Daniel
    Khudanpur, Sanjeev
    18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION, 2017, : 999 - 1003