Robust statistical learning with Lipschitz and convex loss functions

被引:0
作者
Geoffrey Chinot
Guillaume Lecué
Matthieu Lerasle
机构
来源
Probability Theory and Related Fields | 2020年 / 176卷
关键词
62G35;
D O I
暂无
中图分类号
学科分类号
摘要
We obtain estimation and excess risk bounds for Empirical Risk Minimizers (ERM) and minmax Median-Of-Means (MOM) estimators based on loss functions that are both Lipschitz and convex. Results for the ERM are derived under weak assumptions on the outputs and subgaussian assumptions on the design as in Alquier et al. (Estimation bounds and sharp oracle inequalities of regularized procedures with Lipschitz loss functions. arXiv:1702.01402, 2017). The difference with Alquier et al. (2017) is that the global Bernstein condition of this paper is relaxed here into a local assumption. We also obtain estimation and excess risk bounds for minmax MOM estimators under similar assumptions on the output and only moment assumptions on the design. Moreover, the dataset may also contains outliers in both inputs and outputs variables without deteriorating the performance of the minmax MOM estimators. Unlike alternatives based on MOM’s principle (Lecué and Lerasle in Ann Stat, 2017; Lugosi and Mendelson in JEMS, 2016), the analysis of minmax MOM estimators is not based on the small ball assumption (SBA) of Koltchinskii and Mendelson (Int Math Res Not IMRN 23:12991–13008, 2015). In particular, the basic example of non parametric statistics where the learning class is the linear span of localized bases, that does not satisfy SBA (Saumard in Bernoulli 24(3):2176–2203, 2018) can now be handled. Finally, minmax MOM estimators are analysed in a setting where the local Bernstein condition is also dropped out. It is shown to achieve excess risk bounds with exponentially large probability under minimal assumptions insuring only the existence of all objects.
引用
收藏
页码:897 / 940
页数:43
相关论文
共 63 条
[1]  
Alon N(1999)The space complexity of approximating the frequency moments J. Comput. Syst. Sci. 58 137-147
[2]  
Matias Y(2011)Robust linear least squares regression Ann. Stat. 39 2766-2794
[3]  
Szegedy M(2011)Convex optimization with sparsity-inducing norms Optim. Mach. Learn. 5 19-53
[4]  
Audibert J-Y(2017)A new method for estimation and model selection: Invent. Math. 207 425-517
[5]  
Catoni O(2005)-estimation Ann. Stat. 33 1497-1537
[6]  
Bach F(2005)Local Rademacher complexities Ann. Stat. 33 1497-1537
[7]  
Jenatton R(2006)Local rademacher complexities Probab. Theory Relat. Fields 135 311-334
[8]  
Mairal J(1984)Empirical minimization Ann. Inst. H. Poincaré Probab. Stat. 20 201-223
[9]  
Obozinski G(2005)Stabilité et instabilité du risque minimax pour des variables indépendantes équidistribuées ESAIM Probab. Stat. 9 323-375
[10]  
Baraud Y(2015)Theory of classification: a survey of some recent advances Found. Trends® Mach. Learn. 8 231-357