Neural network with unbounded activation functions is universal approximator

被引:172
|
作者
Sonoda, Sho [1 ]
Murata, Noboru [1 ]
机构
[1] Waseda Univ, Fac Sci & Engn, Shinjuku Ku, 3-4-1 Okubo, Tokyo 1698555, Japan
关键词
Neural network; Integral representation; Rectified linear unit (ReLU); Universal approximation; Ridgelet transform; Admissibility condition; Lizorkin distribution; Radon transform; Backprojection filter; Bounded extension to L-2; TRANSFORM; REPRESENTATION; SUPERPOSITIONS; RATES;
D O I
10.1016/j.acha.2015.12.005
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
This paper presents an investigation of the approximation property of neural networks with unbounded activation functions, such as the rectified linear unit (ReLU), which is the new de-facto standard of deep learning. The ReLU network can be analyzed by the ridgelet transform with respect to Lizorkin distributions. By showing three reconstruction formulas by using the Fourier slice theorem, the Radon transform, and Parseval's relation, it is shown that a neural network with unbounded activation functions still satisfies the universal approximation property. As an additional consequence, the ridgelet transform, or the backprojection filter in the Radon domain, is what the network learns after backpropagation. Subject to a constructive admissibility condition, the trained network can be obtained by simply discretizing the ridgelet transform, without backpropagation. Numerical examples not only support the consistency of the admissibility condition but also imply that some non-admissible cases result in low-pass filtering. (C) 2015 Elsevier Inc. All rights reserved.
引用
收藏
页码:233 / 268
页数:36
相关论文
共 50 条
  • [31] Cubic approximation neural network for multivariate functions
    Stein, D
    Feuer, A
    NEURAL NETWORKS, 1998, 11 (02) : 235 - 248
  • [32] Neural network interpolation operators of multivariate functions
    Wang, Guoshun
    Yu, Dansheng
    Guan, Lingmin
    JOURNAL OF COMPUTATIONAL AND APPLIED MATHEMATICS, 2023, 431
  • [33] The GroupMax Neural Network Approximation of Convex Functions
    Warin, Xavier
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (08) : 11608 - 11612
  • [34] A Multicast Routing Scheme for a Universal Spiking Neural Network Architecture
    Wu, Jian
    Furber, Steve
    COMPUTER JOURNAL, 2010, 53 (03) : 280 - 288
  • [35] Simultaneous neural network approximation for smooth functions
    Hon, Sean
    Yang, Haizhao
    NEURAL NETWORKS, 2022, 154 : 152 - 164
  • [36] Neural networks as an approximator for a family of optimization algorithm solutions for online applications
    Lopez-Rojas, Arturo D.
    Cruz-Villar, Carlos A.
    NEURAL COMPUTING & APPLICATIONS, 2024, 36 (06) : 3125 - 3140
  • [37] A Review of Activation Function for Artificial Neural Network
    Rasamoelina, Andrinandrasana David
    Adjailia, Fouzia
    Sincak, Peter
    2020 IEEE 18TH WORLD SYMPOSIUM ON APPLIED MACHINE INTELLIGENCE AND INFORMATICS (SAMI 2020), 2020, : 281 - 286
  • [38] Modswish: a new activation function for neural network
    Kalim, Heena
    Chug, Anuradha
    Singh, Amit Prakash
    EVOLUTIONARY INTELLIGENCE, 2024, 17 (04) : 2637 - 2647
  • [39] REPRESENTATION OF FUNCTIONS BY SUPERPOSITIONS OF A STEP OR SIGMOID FUNCTION AND THEIR APPLICATIONS TO NEURAL NETWORK THEORY
    ITO, Y
    NEURAL NETWORKS, 1991, 4 (03) : 385 - 394
  • [40] Improving the Performance of Neural Networks with an Ensemble of Activation Functions
    Nandi, Arijit
    Jana, Nanda Dulal
    Das, Swagatam
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,