Robust nonparametric regression based on deep ReLU neural networks

被引:0
|
作者
Chen, Juntong [1 ]
机构
[1] Univ Twente, Dept Appl Math, Twente, Netherlands
基金
欧盟地平线“2020”;
关键词
Nonparametric; Robust estimation; Deep neural; Circumventing the curse of dimensionality; Supremum of an empirical; CONVERGENCE;
D O I
10.1016/j.jspi.2024.106182
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
In this paper, we consider robust nonparametric regression using deep neural networks with ReLU activation function. While several existing theoretically justified methods are geared towards robustness against identical heavy-tailed noise distributions, the rise of adversarial attacks has emphasized the importance of safeguarding estimation procedures against systematic contamination. We approach this statistical issue by shifting our focus towards estimating conditional distributions. To address it robustly, we introduce a novel estimation procedure based on l-estimation. Under a mild model assumption, we establish general non-asymptotic risk bounds for the resulting estimators, showcasing their robustness against contamination, outliers, and model misspecification. We then delve into the application of our approach using deep ReLU neural networks. When the model is well-specified and the regression function belongs to an alpha-Holder class, employing l-type estimation on suitable networks enables the resulting estimators to achieve the minimax optimal rate of convergence. Additionally, we demonstrate that deep l-type estimators can circumvent the curse of dimensionality by assuming the regression function closely resembles the composition of several Holder functions. To attain this, new deep fully-connected ReLU neural networks have been designed to approximate this composition class. This approximation result can be of independent interest.
引用
收藏
页数:25
相关论文
共 50 条
  • [21] Convergence of deep ReLU networks
    Xu, Yuesheng
    Zhang, Haizhang
    NEUROCOMPUTING, 2024, 571
  • [22] The Integration of Explicit MPC and ReLU based Neural Networks
    Katz, Justin
    Pappas, Iosif
    Avraamidou, Styliani
    Pistikopoulos, Efstratios N.
    IFAC PAPERSONLINE, 2020, 53 (02): : 11350 - 11355
  • [23] Nonparametric Estimation of Non-Crossing Quantile Regression Process with Deep ReQU Neural Networks
    Shen, Guohao
    Jiao, Yuling
    Lin, Yuanyuan
    Horowitz, Joel L.
    Huang, Jian
    JOURNAL OF MACHINE LEARNING RESEARCH, 2024, 25 : 1 - 75
  • [24] On the CVP for the root lattices via folding with deep ReLU neural networks
    Corlay, Vincent
    Boutros, Joseph J.
    Ciblat, Philippe
    Brunel, Loic
    2019 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2019, : 1622 - 1626
  • [25] Biased ReLU neural networks
    Liang, XingLong
    Xu, Jun
    NEUROCOMPUTING, 2021, 423 : 71 - 79
  • [26] Approximation in shift-invariant spaces with deep ReLU neural networks
    Yang, Yunfei
    Li, Zhen
    Wang, Yang
    NEURAL NETWORKS, 2022, 153 : 269 - 281
  • [27] A Regression Approach to Speech Enhancement Based on Deep Neural Networks
    Xu, Yong
    Du, Jun
    Dai, Li-Rong
    Lee, Chin-Hui
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2015, 23 (01) : 7 - 19
  • [28] Adaptive Robust Watermarking Method Based on Deep Neural Networks
    Li, Fan
    Wan, Chen
    Huang, Fangjun
    DIGITAL FORENSICS AND WATERMARKING, IWDW 2022, 2023, 13825 : 162 - 173
  • [29] Deep ReLU neural networks overcome the curse of dimensionality for partial integrodifferential equations
    Gonon, Lukas
    Schwab, Christoph
    ANALYSIS AND APPLICATIONS, 2023, 21 (01) : 1 - 47
  • [30] Provable Accelerated Convergence of Nesterov's Momentum for Deep ReLU Neural Networks
    Liao, Fangshuo
    Kyrillidis, Anastasios
    INTERNATIONAL CONFERENCE ON ALGORITHMIC LEARNING THEORY, VOL 237, 2024, 237