Approximate Leave-One-Out Cross Validation for Regression With l1 Regularizers

被引:0
作者
Auddy, Arnab [1 ]
Zou, Haolin [2 ]
Rad, Kamiar Rahnama [3 ]
Maleki, Arian [2 ]
机构
[1] Ohio State Univ, Dept Stat, Columbus, OH 43210 USA
[2] Columbia Univ, Dept Stat, New York, NY 10032 USA
[3] CUNY, Baruch Coll, New York, NY 10031 USA
基金
美国国家科学基金会;
关键词
Computational modeling; Signal to noise ratio; Perturbation methods; Measurement; Reviews; Linear regression; Computational efficiency; High dimensional statistics; empirical risk minimization; regularization; cross validation; approximate leave-one-out; elastic net; LASSO;
D O I
10.1109/TIT.2024.3450002
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The out-of-sample error (OO) is the main quantity of interest in risk estimation and model selection. Leave-one- out cross validation (LO) offers a (nearly) distribution-free yet computationally demanding approach to estimate OO. Recent theoretical work showed that approximate leave-one-out cross validation (ALO) is a computationally efficient and statistically reliable estimate of LO (and OO) for generalized linear models with differentiable regularizers. For problems involving non- differentiable regularizers, despite significant empirical evidence, the theoretical understanding of ALO's error remains unknown. In this paper, we present a novel theory for a wide class of problems in the generalized linear model family with non- differentiable regularizers. We bound the error |ALO - LO| in terms of intuitive metrics such as the size of leave-i-out perturbations in active sets, sample size n, number of features p and regularization parameters. As a consequence, for the l(1)-regularized problems, we show that |ALO - LO| ->(p ->infinity) 0 while n/p and signal-to-noise ratio (SNR) are bounded.
引用
收藏
页码:8040 / 8071
页数:32
相关论文
共 54 条