DC Neural Networks avoid overfitting in one-dimensional nonlinear regression

被引:6
作者
Beltran-Royo, Cesar [1 ]
Llopis-Ibor, Laura [1 ]
Pantrigo, Juan J. [1 ]
Ramirez, Ivan [1 ]
机构
[1] Univ Rey Juan Carlos, Comp Sci & Stat Dept, C Tulipan s-n, Mostoles 28933, Madrid, Spain
关键词
DC neural network; Multilayer perceptron; Nonlinear regression; Overfitting; MODEL; FRAMEWORK;
D O I
10.1016/j.knosys.2023.111154
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we analyze Difference of Convex Neural Networks in the context of one-dimensional nonlinear regression. Specifically, we show the surprising ability of the Difference of Convex Multilayer Perceptron (DC-MLP) to avoid overfitting in nonlinear regression. Otherwise said, DC-MLPs self-regularize (do not require additional regularization techniques). Thus, DC-MLPs could result very useful for practical purposes based on one-dimensional nonlinear regression. It turns out that shallow MLPs with a convex activation (ReLU, softplus, etc.) fall in the class of DC-MLPs. On the other hand, we call SQ-MLP the shallow MLP with a Squashing activation (logistic, hyperbolic tangent, etc.). In the numerical experiments, we show that DC-MLPs used for nonlinear regression avoid overfitting, in contrast with SQ-MLPs. We also compare DC-MLPs and SQ-MLPs from a theoretical point of view.
引用
收藏
页数:24
相关论文
共 55 条
  • [1] Aceves-Fernandez M., 2018, EEG steady-state visual evoked potential signals
  • [2] Akbilgic O., 2013, Istanbul stock exchange
  • [3] Amos B, 2017, 34 INT C MACHINE LEA, V70
  • [4] [Anonymous], 2020, Ai4i 2020 predictive maintenance dataset data set
  • [5] [Anonymous], 2016, Springer Optimization and Its Applications
  • [6] [Anonymous], 2020, Productivity prediction of garment employees
  • [7] Apostol T.M., 1974, Addison-Wesley series in Mathematics, V2nd
  • [8] Fit without fear: remarkable mathematical phenomena of deep learning through the prism of interpolation
    Belkin, Mikhail
    [J]. ACTA NUMERICA, 2021, 30 : 203 - 248
  • [9] Reconciling modern machine-learning practice and the classical bias-variance trade-off
    Belkin, Mikhail
    Hsu, Daniel
    Ma, Siyuan
    Mandal, Soumik
    [J]. PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2019, 116 (32) : 15849 - 15854
  • [10] DYNAMIC PROGRAMMING
    BELLMAN, R
    [J]. SCIENCE, 1966, 153 (3731) : 34 - &