Learning under (1+ε)-moment conditions

被引:23
作者
Feng, Yunlong [1 ]
Wu, Qiang [2 ]
机构
[1] SUNY Albany, Dept Math & Stat, Albany, NY 12222 USA
[2] Middle Tennessee State Univ, Dept Math Sci, Murfreesboro, TN 37132 USA
关键词
Learning theory; Empirical risk minimization; Mean regression calibration; No-free lunch theorem; Weak moment condition; Robust regression; ROBUST VARIABLE SELECTION; EXPONENTIAL SQUARED LOSS; GAUSSIAN FIELDS; INDUCED LOSSES; CORRENTROPY; REGRESSION; REGISTRATION; MINIMIZATION; CRITERION; MIXTURE;
D O I
10.1016/j.acha.2020.05.009
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
We study the theoretical underpinning of a robust empirical risk minimization (RERM) scheme which has been finding numerous successful applications across various data science fields owing to its robustness to outliers and heavy-tailed noises. The specialties of RERM lie in its nonconvexity and that it is induced by a loss function with an integrated scale parameter trading off the robustness and the prediction accuracy. The nonconvexity of RERM and the integrated scale parameter also bring barriers when assessing its learning performance theoretically. In this paper, concerning the study of RERM, we make the following main contributions. First, we establish a no-free-lunch result, showing that there is no hope of distribution-free learning of the truth without adjusting the scale parameter. Second, by imposing the (1 + epsilon)-th (with epsilon > 0) order moment condition on the response variable, we establish a comparison theorem that characterizes the relation between the excess generalization error of RERM and its prediction error. Third, with a diverging scale parameter, we establish almost sure convergence rates for RERM under the (1 + epsilon)-moment condition. Notably, the (1 + epsilon)-moment condition allows the presence of noise with infinite variance. Last but not least, the learning theory analysis of RERM conducted in this study, on one hand, showcases the merits of RERM on robustness and the trade-off role that the scale parameter plays, and on the other hand, brings us inspirational insights into robust machine learning. Published by Elsevier Inc.
引用
收藏
页码:495 / 520
页数:26
相关论文
共 78 条
[1]   Target Uncertainty Mediates Sensorimotor Error Correction [J].
Acerbi, Luigi ;
Vijayakumar, Sethu ;
Wolpert, Daniel M. .
PLOS ONE, 2017, 12 (01)
[2]   On the Origins of Suboptimality in Human Probabilistic Inference [J].
Acerbi, Luigi ;
Vijayakumar, Sethu ;
Wolpert, Daniel M. .
PLOS COMPUTATIONAL BIOLOGY, 2014, 10 (06)
[3]  
Allen-Zhu Z., 2019, ADV NEURAL INFORM PR, P6155
[4]  
[Anonymous], PREPRINT
[5]  
[Anonymous], IFAC P
[6]  
[Anonymous], 2019, FRONTIERS APPL MATH
[7]  
[Anonymous], 2018, Robust Statistics: Theory and Methods
[8]  
[Anonymous], 2020, PREPRINT
[9]  
[Anonymous], STAT RISK MODEL
[10]  
[Anonymous], 1984, STAT 6 DECISIONS