Controlled learning of pointwise nonlinearities in neural-network-like architectures

被引:0
|
作者
Unser, Michael [1 ]
Goujon, Alexis [1 ]
Ducotterd, Stanislas [1 ]
机构
[1] Ecole Polytech Fed Lausanne EPFL, Biomed Imaging Grp, Stn 17, CH-1015 Lausanne, Switzerland
基金
欧洲研究理事会;
关键词
LINEAR INVERSE PROBLEMS; CONVEX REGULARIZERS; IMAGE; RECONSTRUCTION; ALGORITHM; SPLINES; MODELS;
D O I
10.1016/j.acha.2025.101764
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
We present a general variational framework for the training of freeform nonlinearities in layered computational architectures subject to some slope constraints. The regularization that we add to the traditional training loss penalizes the second-order total variation of each trainable activation. The slope constraints allow us to impose properties such as 1-Lipschitz stability, firm non-expansiveness, and monotonicity/invertibility. These properties are crucial to ensure the proper functioning of certain classes of signal-processing algorithms (e.g., plug-and-play schemes, unrolled proximal gradient, invertible flows). We prove that the global optimum of the stated constrained-optimization problem is achieved with nonlinearities that are adaptive nonuniform linear splines. We then show how to solve the resulting function-optimization problem numerically by representing the nonlinearities in a suitable (nonuniform) B-spline basis. Finally, we illustrate the use of our framework with the data-driven design of (weakly) convex regularizers for the denoising of images and the resolution of inverse problems.
引用
收藏
页数:19
相关论文
共 50 条
  • [1] A survey of deep neural network architectures and their applications
    Liu, Weibo
    Wang, Zidong
    Liu, Xiaohui
    Zeng, Nianyin
    Liu, Yurong
    Alsaadi, Fuad E.
    NEUROCOMPUTING, 2017, 234 : 11 - 26
  • [2] Function-Space Optimality of Neural Architectures with Multivariate Nonlinearities
    Parhi, Rahul
    Unser, Michael
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2025, 7 (01): : 110 - 135
  • [3] Learning in Memristive Neural Network Architectures Using Analog Backpropagation Circuits
    Krestinskaya, Olga
    Salama, Khaled Nabil
    James, Alex Pappachen
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2019, 66 (02) : 719 - 732
  • [4] Automated Design of Neural Network Architectures With Reinforcement Learning for Detection of Global Manipulations
    Chen, Yifang
    Wang, Zheng
    Wang, Z. Jane
    Kang, Xiangui
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2020, 14 (05) : 997 - 1011
  • [5] Deep transform and metric learning network: Wedding deep dictionary learning and neural network
    Tang, Wen
    Chouzenoux, Emilie
    Pesquet, Jean-Christophe
    Krim, Hamid
    NEUROCOMPUTING, 2022, 509 : 244 - 256
  • [6] Neural Network Architectures and Magnetic Hysteresis: Overview and Comparisons
    Licciardi, Silvia
    Ala, Guido
    Francomano, Elisa
    Viola, Fabio
    Lo Giudice, Michele
    Salvini, Alessandro
    Sargeni, Fausto
    Bertolini, Vittorio
    Di Schino, Andrea
    Faba, Antonio
    MATHEMATICS, 2024, 12 (21)
  • [7] Optimal Design of Convolutional Neural Network Architectures Using Teaching-Learning-Based Optimization for Image Classification
    Ang, Koon Meng
    El-kenawy, El-Sayed M.
    Abdelhamid, Abdelaziz A.
    Ibrahim, Abdelhameed
    Alharbi, Amal H.
    Khafaga, Doaa Sami
    Tiang, Sew Sun
    Lim, Wei Hong
    SYMMETRY-BASEL, 2022, 14 (11):
  • [8] One dimensional convolutional neural network architectures for wind prediction
    Harbola, Shubhi
    Coors, Volker
    ENERGY CONVERSION AND MANAGEMENT, 2019, 195 : 70 - 75
  • [9] Parallel Architectures for Learning the RTRN and Elman Dynamic Neural Networks
    Bilski, Jaroslaw
    Smolag, Jacek
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2015, 26 (09) : 2561 - 2570
  • [10] First break picking with deep learning - evaluation of network architectures
    Zwartjes, Paul
    Yoo, Jewoo
    GEOPHYSICAL PROSPECTING, 2022, 70 (02) : 318 - 342