On minimal representations of shallow ReLU networks

被引:6
作者
Dereich, Steffen [1 ]
Kassing, Sebastian [1 ]
机构
[1] Westfal Wilhelms Univ Munster, Inst Math Stochast, Math & Informat, Fachbereich 10,Orleans Ring 10, D-48149 Munster, Germany
关键词
Neural networks; Shallow networks; Minimal representations; ReLU activation; MULTILAYER FEEDFORWARD NETWORKS;
D O I
10.1016/j.neunet.2022.01.006
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The realization function of a shallow ReLU network is a continuous and piecewise affine function f : R-d ->& nbsp;R, where the domain Rd is partitioned by a set of n hyperplanes into cells on which f is affine. We show that the minimal representation for f uses either n, n + 1 or n + 2 neurons and we characterize each of the three cases. In the particular case, where the input layer is one-dimensional, minimal representations always use at most n+1 neurons but in all higher dimensional settings there are functions for which n+2 neurons are needed. Then we show that the set of minimal networks representing f forms a C-infinity-submanifold M and we derive the dimension and the number of connected components of M. Additionally, we give a criterion for the hyperplanes that guarantees that a continuous, piecewise affine function is the realization function of an appropriate shallow ReLU network.(c) 2022 Elsevier Ltd. All rights reserved.
引用
收藏
页码:121 / 128
页数:8
相关论文
共 50 条
  • [31] Algebraic Analysis of Minimal Representations
    Kobayashi, Toshiyuki
    PUBLICATIONS OF THE RESEARCH INSTITUTE FOR MATHEMATICAL SCIENCES, 2011, 47 (02) : 585 - 611
  • [32] Convergent Time-Stepping Schemes for Analog ReLU Networks
    Elfadel, Ibrahim M.
    2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2021,
  • [33] Probabilistic Verification of ReLU Neural Networks via Characteristic Functions
    Pilipovsky, Joshua
    Sivaramakrishnan, Vignesh
    Oishi, Meeko M. K.
    Tsiotras, Panagiotis
    LEARNING FOR DYNAMICS AND CONTROL CONFERENCE, VOL 211, 2023, 211
  • [34] ReLU deep neural networks from the hierarchical basis perspective
    He, Juncai
    Li, Lin
    Xu, Jinchao
    COMPUTERS & MATHEMATICS WITH APPLICATIONS, 2022, 120 : 105 - 114
  • [35] HOW DO NOISE TAILS IMPACT ON DEEP RELU NETWORKS?
    Fan, Jianqian
    Gu, Yihong
    Zhou, Wen-Xin
    ANNALS OF STATISTICS, 2024, 52 (04) : 1845 - 1871
  • [36] Width is Less Important than Depth in ReLU Neural Networks
    Vardi, Gal
    Yehudai, Gilad
    Shamir, Ohad
    CONFERENCE ON LEARNING THEORY, VOL 178, 2022, 178
  • [37] On the uniform approximation estimation of deep ReLU networks via frequency decomposition
    Chen, Liang
    Liu, Wenjun
    AIMS MATHEMATICS, 2022, 7 (10): : 19018 - 19025
  • [38] EV-GAN: Simulation of extreme events with ReLU neural networks
    Allouche, Michael
    Girard, Stephane
    Gobet, Emmanuel
    JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [39] Gradient descent optimizes over-parameterized deep ReLU networks
    Zou, Difan
    Cao, Yuan
    Zhou, Dongruo
    Gu, Quanquan
    MACHINE LEARNING, 2020, 109 (03) : 467 - 492
  • [40] New Error Bounds for Deep ReLU Networks Using Sparse Grids
    Montanelli, Hadrien
    Du, Qiang
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2019, 1 (01): : 78 - 92