Neural network with unbounded activation functions is universal approximator

被引:172
|
作者
Sonoda, Sho [1 ]
Murata, Noboru [1 ]
机构
[1] Waseda Univ, Fac Sci & Engn, Shinjuku Ku, 3-4-1 Okubo, Tokyo 1698555, Japan
关键词
Neural network; Integral representation; Rectified linear unit (ReLU); Universal approximation; Ridgelet transform; Admissibility condition; Lizorkin distribution; Radon transform; Backprojection filter; Bounded extension to L-2; TRANSFORM; REPRESENTATION; SUPERPOSITIONS; RATES;
D O I
10.1016/j.acha.2015.12.005
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
This paper presents an investigation of the approximation property of neural networks with unbounded activation functions, such as the rectified linear unit (ReLU), which is the new de-facto standard of deep learning. The ReLU network can be analyzed by the ridgelet transform with respect to Lizorkin distributions. By showing three reconstruction formulas by using the Fourier slice theorem, the Radon transform, and Parseval's relation, it is shown that a neural network with unbounded activation functions still satisfies the universal approximation property. As an additional consequence, the ridgelet transform, or the backprojection filter in the Radon domain, is what the network learns after backpropagation. Subject to a constructive admissibility condition, the trained network can be obtained by simply discretizing the ridgelet transform, without backpropagation. Numerical examples not only support the consistency of the admissibility condition but also imply that some non-admissible cases result in low-pass filtering. (C) 2015 Elsevier Inc. All rights reserved.
引用
收藏
页码:233 / 268
页数:36
相关论文
共 50 条
  • [21] Wavelets as activation functions in Neural Networks
    Herrera, Oscar
    Priego, Belem
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2022, 42 (05) : 4345 - 4355
  • [22] Neural network interpolation operators activated by smooth ramp functions
    Qian, Yunyou
    Yu, Dansheng
    ANALYSIS AND APPLICATIONS, 2022, 20 (04) : 791 - 813
  • [23] Research on the Blind Equalization technology Based on the Complex BP Neural Network with Tunable Activation Functions
    Zhao, Juan
    Gao, Zheng-Ming
    2017 IEEE 2ND ADVANCED INFORMATION TECHNOLOGY, ELECTRONIC AND AUTOMATION CONTROL CONFERENCE (IAEAC), 2017, : 813 - 817
  • [24] Cascaded centralized TSK fuzzy system: universal approximator and high interpretation
    Wang, ST
    Chung, FL
    Shen, HB
    Hu, DW
    APPLIED SOFT COMPUTING, 2005, 5 (02) : 131 - 145
  • [25] Integration of residual network and convolutional neural network along with various activation functions and global pooling for time series classification
    Zou, Xiaowu
    Wang, Zidong
    Li, Qi
    Sheng, Weiguo
    NEUROCOMPUTING, 2019, 367 : 39 - 45
  • [26] Implementation of neural network with approximations functions
    Hnatiuc, M
    Lamarque, G
    SCS 2003: INTERNATIONAL SYMPOSIUM ON SIGNALS, CIRCUITS AND SYSTEMS, VOLS 1 AND 2, PROCEEDINGS, 2003, : 553 - 556
  • [27] Activation Functions and Their Characteristics in Deep Neural Networks
    Ding, Bin
    Qian, Huimin
    Zhou, Jun
    PROCEEDINGS OF THE 30TH CHINESE CONTROL AND DECISION CONFERENCE (2018 CCDC), 2018, : 1836 - 1841
  • [28] Neural network interpolation operators optimized by Lagrange polynomial?
    Wang, Guoshun
    Yu, Dansheng
    Zhou, Ping
    NEURAL NETWORKS, 2022, 153 : 179 - 191
  • [29] Neural network with NewSigmoid activation function
    Kumar, Arvind
    Sodhi, Sartaj Singh
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2022, 43 (01) : 545 - 559
  • [30] Approximation by network operators with logistic activation functions
    Chen, Zhixiang
    Cao, Feilong
    Hu, Jinjie
    APPLIED MATHEMATICS AND COMPUTATION, 2015, 256 : 565 - 571