A Representer Theorem for Deep Neural Networks

被引:0
作者
Unser, Michael [1 ]
机构
[1] Ecole Polytech Fed Lausanne, Biomed Imaging Grp, CH-1015 Lausanne, Switzerland
基金
瑞士国家科学基金会;
关键词
splines; regularization; sparsity; learning; deep neural networks; activation functions; LINEAR INVERSE PROBLEMS; SPLINES; KERNELS;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We propose to optimize the activation functions of a deep neural network by adding a corresponding functional regularization to the cost function. We justify the use of a second-order total-variation criterion. This allows us to derive a general representer theorem for deep neural networks that makes a direct connection with splines and sparsity. Specifically, we show that the optimal network configuration can be achieved with activation functions that are nonuniform linear splines with adaptive knots. The bottom line is that the action of each neuron is encoded by a spline whose parameters (including the number of knots) are optimized during the training procedure. The scheme results in a computational structure that is compatible with existing deep-ReLU, parametric ReLU, APL (adaptive piecewise-linear) and MaxOut architectures. It also suggests novel optimization challenges and makes an explicit link with l(1) minimization and sparsity-promoting techniques.
引用
收藏
页数:30
相关论文
共 50 条
  • [41] Deep Neural Networks in Semantic Analysis
    Averkin, Alexey
    Yarushev, Sergey
    10TH INTERNATIONAL CONFERENCE ON THEORY AND APPLICATION OF SOFT COMPUTING, COMPUTING WITH WORDS AND PERCEPTIONS - ICSCCW-2019, 2020, 1095 : 846 - 853
  • [42] Threshout Regularization for Deep Neural Networks
    Williams, Travis
    Li, Robert
    SOUTHEASTCON 2021, 2021, : 728 - 735
  • [43] Digital watermarking for deep neural networks
    Yuki Nagai
    Yusuke Uchida
    Shigeyuki Sakazawa
    Shin’ichi Satoh
    International Journal of Multimedia Information Retrieval, 2018, 7 : 3 - 16
  • [44] Attacking Neural Networks with Neural Networks: Towards Deep Synchronization for Backdoor Attacks
    Guan, Zihan
    Sun, Lichao
    Du, Mengnan
    Liu, Ninghao
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 608 - 618
  • [45] Digital watermarking for deep neural networks
    Nagai, Yuki
    Uchida, Yusuke
    Sakazawa, Shigeyuki
    Satoh, Shin'ichi
    INTERNATIONAL JOURNAL OF MULTIMEDIA INFORMATION RETRIEVAL, 2018, 7 (01) : 3 - 16
  • [46] Deep Neural Networks on Chip - A Survey
    Huo Yingge
    Ali, Imran
    Lee, Kang-Yoon
    2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING (BIGCOMP 2020), 2020, : 589 - 592
  • [47] Addressing Sparsity in Deep Neural Networks
    Zhou, Xuda
    Du, Zidong
    Zhang, Shijin
    Zhang, Lei
    Lan, Huiying
    Liu, Shaoli
    Li, Ling
    Guo, Qi
    Chen, Tianshi
    Chen, Yunji
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2019, 38 (10) : 1858 - 1871
  • [48] Polymorphic Accelerators for Deep Neural Networks
    Azizimazreah, Arash
    Chen, Lizhong
    IEEE TRANSACTIONS ON COMPUTERS, 2022, 71 (03) : 534 - 546
  • [49] TRACK FINDING WITH DEEP NEURAL NETWORKS
    Kucharczyk, Marcin
    Wolter, Marcin
    COMPUTER SCIENCE-AGH, 2019, 20 (04): : 477 - 493
  • [50] Federated Repair of Deep Neural Networks
    Li Calsi, Davide
    Laurent, Thomas
    Arcaini, Paolo
    Ishikawa, Fuyuki
    PROCEEDINGS OF THE 2024 IEEE/ACM INTERNATIONAL WORKSHOP ON DEEP LEARNING FOR TESTING AND TESTING FOR DEEP LEARNING, DEEPTEST 2024, 2024, : 17 - 24