Function-Space Optimality of Neural Architectures with Multivariate Nonlinearities

被引:0
|
作者
Parhi, Rahul [1 ,2 ]
Unser, Michael [3 ]
机构
[1] Univ Calif San Diego, Dept Elect & Comp Engn, La Jolla, CA 92093 USA
[2] Ecole Polytech Fed Lausanne, Biomed Imaging Grp, CH-1015 Lausanne, Switzerland
[3] Ecole Polytech Fed Lausanne, Biomed Imaging Grp, CH-1015 Lausanne, Switzerland
来源
SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE | 2025年 / 7卷 / 01期
基金
瑞士国家科学基金会;
关键词
multi-index models; multivariate nonlinearities; neural networks; regularization; representer theorem; PLANE TRANSFORMS; RIDGE FUNCTIONS; APPROXIMATION; NETWORKS; REGRESSION; REDUCTION; INVERSION; SPLINES;
D O I
10.1137/23M1620971
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
We investigate the function-space optimality (specifically, the Banach-space optimality) of a large class of shallow neural architectures with multivariate nonlinearities/activation functions. To that end, we construct a new family of Banach spaces defined via a regularization operator, the k-plane transform, and a sparsity-promoting norm. We prove a representer theorem that states that the solution sets to learning problems posed over these Banach spaces are completely characterized by neural architectures with multivariate nonlinearities. These optimal architectures have skip connections and are tightly connected to orthogonal weight normalization and multi-index models, both of which have received recent interest in the neural network community. Our framework is compatible with a number of classical nonlinearities including the rectified linear unit activation function, the norm activation function, and the radial basis functions found in the theory of thin-plate/polyharmonic splines. We also show that the underlying spaces are special instances of reproducing kernel Banach spaces and variation spaces. Our results shed light on the regularity of functions learned by neural networks trained on data, particularly with multivariate nonlinearities, and provide new theoretical motivation for several architectural choices found in practice.
引用
收藏
页码:110 / 135
页数:26
相关论文
共 21 条
  • [1] Controlled learning of pointwise nonlinearities in neural-network-like architectures
    Unser, Michael
    Goujon, Alexis
    Ducotterd, Stanislas
    APPLIED AND COMPUTATIONAL HARMONIC ANALYSIS, 2025, 77
  • [2] Efficient Design Space Exploration for Sparse Mixed Precision Neural Architectures
    Chitty-Venkata, Krishna Teja
    Emani, Murali
    Vishwanath, Venkatram
    Somani, Arun K.
    PROCEEDINGS OF THE 31ST INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE PARALLEL AND DISTRIBUTED COMPUTING, HPDC 2022, 2022, : 265 - 276
  • [3] Regularising neural networks using flexible multivariate activation function
    Solazzi, M
    Uncini, A
    NEURAL NETWORKS, 2004, 17 (02) : 247 - 260
  • [4] Multivariate statistical inference in a radial basis function neural network
    de Leon-Delgado, Homero
    Praga-Alej, Rolando J.
    Gonzalez-Gonzalez, David S.
    Cantu-Sifuentes, Mario
    EXPERT SYSTEMS WITH APPLICATIONS, 2018, 93 : 313 - 321
  • [5] A New Function Space from Barron Class and Application to Neural Network Approximation
    Meng, Yan
    Ming, Pingbing
    COMMUNICATIONS IN COMPUTATIONAL PHYSICS, 2022, 32 (05) : 1361 - 1400
  • [6] A three layer neural network can represent any multivariate function
    Ismailov, Vugar E.
    JOURNAL OF MATHEMATICAL ANALYSIS AND APPLICATIONS, 2023, 523 (01)
  • [7] ExplainLFS: Explaining neural architectures for similarity learning from local perturbations in the latent feature space
    Bello, Marilyn
    Costa, Pablo
    Napoles, Gonzalo
    Mesejo, Pablo
    Cordon, Oscar
    INFORMATION FUSION, 2024, 108
  • [8] Prediction of Multivariate Chaotic Time Series Via Radial Basis Function Neural Network
    Chen, Diyi
    Han, Wenting
    COMPLEXITY, 2013, 18 (04) : 55 - 66
  • [9] Reducing the Dependence of the Neural Network Function to Systematic Uncertainties in the Input Space
    Wunsch S.
    Jörger S.
    Wolf R.
    Quast G.
    Computing and Software for Big Science, 2020, 4 (1)
  • [10] The Barron Space and the Flow-Induced Function Spaces for Neural Network Models
    Weinan, E.
    Ma, Chao
    Wu, Lei
    CONSTRUCTIVE APPROXIMATION, 2022, 55 (01) : 369 - 406