A Representer Theorem for Deep Neural Networks

被引:0
|
作者
Unser, Michael [1 ]
机构
[1] Ecole Polytech Fed Lausanne, Biomed Imaging Grp, CH-1015 Lausanne, Switzerland
基金
瑞士国家科学基金会;
关键词
splines; regularization; sparsity; learning; deep neural networks; activation functions; LINEAR INVERSE PROBLEMS; SPLINES; KERNELS;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We propose to optimize the activation functions of a deep neural network by adding a corresponding functional regularization to the cost function. We justify the use of a second-order total-variation criterion. This allows us to derive a general representer theorem for deep neural networks that makes a direct connection with splines and sparsity. Specifically, we show that the optimal network configuration can be achieved with activation functions that are nonuniform linear splines with adaptive knots. The bottom line is that the action of each neuron is encoded by a spline whose parameters (including the number of knots) are optimized during the training procedure. The scheme results in a computational structure that is compatible with existing deep-ReLU, parametric ReLU, APL (adaptive piecewise-linear) and MaxOut architectures. It also suggests novel optimization challenges and makes an explicit link with l(1) minimization and sparsity-promoting techniques.
引用
收藏
页数:30
相关论文
共 50 条
  • [21] Robust dimensionality reduction for data visualization with deep neural networks
    Becker, Martin
    Lippel, Jens
    Stuhlsatz, Andre
    Zielke, Thomas
    GRAPHICAL MODELS, 2020, 108
  • [22] DEEPKRIGING: SPATIALLY DEPENDENT DEEP NEURAL NETWORKS FOR SPATIAL PREDICTION
    Chen, Wanfang
    Li, Yuxiao
    Reich, Brian J.
    Sun, Ying
    STATISTICA SINICA, 2024, 34 (01) : 291 - 311
  • [23] Deep Neural Networks With Trainable Activations and Controlled Lipschitz Constant
    Aziznejad, Shayan
    Gupta, Harshit
    Campos, Joaquim
    Unser, Michael
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2020, 68 : 4688 - 4699
  • [24] Dynamic Slicing for Deep Neural Networks
    Zhang, Ziqi
    Li, Yuanchun
    Guo, Yao
    Chen, Xiangqun
    Liu, Yunxin
    PROCEEDINGS OF THE 28TH ACM JOINT MEETING ON EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING (ESEC/FSE '20), 2020, : 838 - 850
  • [25] Activation Ensembles for Deep Neural Networks
    Klabjan, Diego
    Harmon, Mark
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 206 - 214
  • [26] Conceptual alignment deep neural networks
    Dai, Yinglong
    Wang, Guojun
    Li, Kuan-Ching
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2018, 34 (03) : 1631 - 1642
  • [27] The Representation of Speech in Deep Neural Networks
    Scharenborg, Odette
    van der Gouw, Nikki
    Larson, Martha
    Marchiori, Elena
    MULTIMEDIA MODELING, MMM 2019, PT II, 2019, 11296 : 194 - 205
  • [28] Temporal Alignment for Deep Neural Networks
    Lin, Payton
    Lyu, Dau-Cheng
    Chang, Yun-Fan
    Tsao, Yu
    2015 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP), 2015, : 108 - 112
  • [29] Embedding Watermarks into Deep Neural Networks
    Uchida, Yusuke
    Nagai, Yuki
    Sakazawa, Shigeyuki
    Satoh, Shin'ichi
    PROCEEDINGS OF THE 2017 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL (ICMR'17), 2017, : 274 - 282
  • [30] Subaging in underparametrized deep neural networks
    Herrera Segura, Carolina
    Montoya, Edison
    Tapias, Diego
    MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2022, 3 (03):