On minimal representations of shallow ReLU networks

被引:6
作者
Dereich, Steffen [1 ]
Kassing, Sebastian [1 ]
机构
[1] Westfal Wilhelms Univ Munster, Inst Math Stochast, Math & Informat, Fachbereich 10,Orleans Ring 10, D-48149 Munster, Germany
关键词
Neural networks; Shallow networks; Minimal representations; ReLU activation; MULTILAYER FEEDFORWARD NETWORKS;
D O I
10.1016/j.neunet.2022.01.006
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The realization function of a shallow ReLU network is a continuous and piecewise affine function f : R-d ->& nbsp;R, where the domain Rd is partitioned by a set of n hyperplanes into cells on which f is affine. We show that the minimal representation for f uses either n, n + 1 or n + 2 neurons and we characterize each of the three cases. In the particular case, where the input layer is one-dimensional, minimal representations always use at most n+1 neurons but in all higher dimensional settings there are functions for which n+2 neurons are needed. Then we show that the set of minimal networks representing f forms a C-infinity-submanifold M and we derive the dimension and the number of connected components of M. Additionally, we give a criterion for the hyperplanes that guarantees that a continuous, piecewise affine function is the realization function of an appropriate shallow ReLU network.(c) 2022 Elsevier Ltd. All rights reserved.
引用
收藏
页码:121 / 128
页数:8
相关论文
共 50 条
  • [1] Rates of approximation by ReLU shallow neural networks
    Mao, Tong
    Zhou, Ding-Xuan
    JOURNAL OF COMPLEXITY, 2023, 79
  • [2] Weighted variation spaces and approximation by shallow ReLU networks
    Devore, Ronald
    Nowak, Robert D.
    Parhi, Rahul
    Siegel, Jonathan W.
    APPLIED AND COMPUTATIONAL HARMONIC ANALYSIS, 2025, 74
  • [3] Near-Minimax Optimal Estimation With Shallow ReLU Neural Networks
    Parhi, Rahul
    Nowak, Robert D. D.
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2023, 69 (02) : 1125 - 1140
  • [4] ON THE EXISTENCE OF MINIMIZERS IN SHALLOW RESIDUAL RELU NEURAL NETWORK OPTIMIZATION LANDSCAPES
    Dereich, Steffen
    Jentzen, Arnulf
    Kassing, Sebastian
    SIAM JOURNAL ON NUMERICAL ANALYSIS, 2024, 62 (06) : 2640 - 2666
  • [5] Gradient Descent Provably Escapes Saddle Points in the Training of Shallow ReLU Networks
    Cheridito, Patrick
    Jentzen, Arnulf
    Rossmannek, Florian
    JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 2024, 203 (03) : 2617 - 2648
  • [6] Nonparametric Regression Using Over-parameterized Shallow ReLU Neural Networks
    Yang, Yunfei
    Zhou, Ding-Xuan
    JOURNAL OF MACHINE LEARNING RESEARCH, 2024, 25 : 1 - 35
  • [7] Gradient flow dynamics of shallow ReLU networks for square loss and orthogonal inputs
    Boursier, Etienne
    Pillaud-Vivien, Loucas
    Flammarion, Nicolas
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [8] Shallow Univariate ReLU Networks as Splines: Initialization, Loss Surface, Hessian, and Gradient Flow Dynamics
    Sahs, Justin
    Pyle, Ryan
    Damaraju, Aneel
    Caro, Josue Ortega
    Tavaslioglu, Onur
    Lu, Andy
    Anselmi, Fabio
    Patel, Ankit B.
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2022, 5
  • [9] On the Stability of Analog ReLU Networks
    Elfadel, Ibrahim M.
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2021, 40 (11) : 2426 - 2430
  • [10] Approximation of Nonlinear Functionals Using Deep ReLU Networks
    Song, Linhao
    Fan, Jun
    Chen, Di-Rong
    Zhou, Ding-Xuan
    JOURNAL OF FOURIER ANALYSIS AND APPLICATIONS, 2023, 29 (04)