Approximating functions with multi-features by deep convolutional neural networks

被引:36
作者
Mao, Tong [1 ]
Shi, Zhongjie [2 ]
Zhou, Ding-Xuan [3 ]
机构
[1] Claremont Grad Univ, Inst Math Sci, 710 N Coll Ave, Claremont, CA 91711 USA
[2] Katholieke Univ Leuven, Dept Elect Engn, ESAT STADIUS, Kasteelpk Arenberg 10, B-3001 Leuven, Belgium
[3] Univ Sydney, Sch Math & Stat, Sydney, NSW 2006, Australia
基金
美国国家科学基金会;
关键词
Deep learning; convolutional neural networks; rates of approximation; curse of dimensionality; feature extraction; ERROR-BOUNDS;
D O I
10.1142/S0219530522400085
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Deep convolutional neural networks (DCNNs) have achieved great empirical success in many fields such as natural language processing, computer vision, and pattern recognition. But there still lacks theoretical understanding of the flexibility and adaptivity of DCNNs in various learning tasks, and the power of DCNNs at feature extraction. We propose a generic DCNN structure consisting of two groups of convolutional layers associated with two downsampling operators, and a fully connected layer, which is determined only by three structural parameters. Our generic DCNNs are capable of extracting various features including not only polynomial features but also general smooth features. We also show that the curse of dimensionality can be circumvented by our DCNNs for target functions of the compositional form with (symmetric) polynomial features, spatially sparse smooth features, and interaction features. These demonstrate the expressive power of our DCNN structure, while the model selection can be relaxed comparing with other deep neural networks since there are only three hyperparameters controlling the architecture to tune.
引用
收藏
页码:93 / 125
页数:33
相关论文
共 41 条
  • [1] Bach F, 2017, J MACH LEARN RES, V18
  • [2] UNIVERSAL APPROXIMATION BOUNDS FOR SUPERPOSITIONS OF A SIGMOIDAL FUNCTION
    BARRON, AR
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 1993, 39 (03) : 930 - 945
  • [3] ON DEEP LEARNING AS A REMEDY FOR THE CURSE OF DIMENSIONALITY IN NONPARAMETRIC REGRESSION
    Bauer, Benedikt
    Kohler, Michael
    [J]. ANNALS OF STATISTICS, 2019, 47 (04) : 2261 - 2285
  • [4] Realization of Spatial Sparseness by Deep ReLU Nets With Massive Data
    Chui, Charles K.
    Lin, Shao-Bo
    Zhang, Bo
    Zhou, Ding-Xuan
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (01) : 229 - 243
  • [5] CHUI CK, 1994, MATH COMPUT, V63, P607, DOI 10.1090/S0025-5718-1994-1240656-2
  • [6] Cucker F, 2007, C MO AP C M, P1, DOI 10.1017/CBO9780511618796
  • [7] Cybenko G., 1989, Mathematics of Control, Signals, and Systems, V2, P303, DOI 10.1007/BF02551274
  • [8] Spatially Sparse Precoding in Millimeter Wave MIMO Systems
    El Ayach, Omar
    Rajagopal, Sridhar
    Abu-Surra, Shadi
    Pi, Zhouyue
    Heath, Robert W., Jr.
    [J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2014, 13 (03) : 1499 - 1513
  • [9] Theory of deep convolutional neural networks II: Spherical analysis
    Fang, Zhiying
    Feng, Han
    Huang, Shuo
    Zhou, Ding-Xuan
    [J]. NEURAL NETWORKS, 2020, 131 : 154 - 162
  • [10] CNN MODELS FOR READABILITY OF CHINESE TEXTS
    Feng, Han
    Hou, Sizai
    Wei, Le-Yin
    Zhou, Ding-Xuan
    [J]. MATHEMATICAL FOUNDATIONS OF COMPUTING, 2022, 5 (04): : 351 - 362