Asymptotic Analysis of Robust LASSOs in the Presence of Noise With Large Variance

被引:24
作者
Chen, Xiaohui [1 ]
Wang, Z. Jane [1 ]
McKeown, Martin J. [2 ]
机构
[1] Univ British Columbia, Dept Elect & Comp Engn, Vancouver, BC V6T 1Z4, Canada
[2] Univ British Columbia, Dept Med Neurol, Vancouver, BC V6T 2B5, Canada
关键词
Asymptotic normality; Huber loss; least absolute shrinkage and selection operator (LASSO); model selection consistency; random designs; robustness; signal recovery; sparse linear regression; LARGE UNDERDETERMINED SYSTEMS; VARIABLE SELECTION; REGRESSION; RECOVERY; REPRESENTATIONS; EQUATIONS;
D O I
10.1109/TIT.2010.2059770
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the context of linear regression, the least absolute shrinkage and selection operator (LASSO) is probably the most popular supervised-learning technique proposed to recover sparse signals from high-dimensional measurements. Prior literature has mainly concerned itself with independent, identically distributed noise with moderate variance. In many real applications, however, the measurement errors may have heavy-tailed distributions or suffer from severe outliers, making the LASSO poorly estimate the coefficients due to its sensitivity to large error variance. To address this concern, a robust version of the LASSO is proposed, and the limiting distribution of its estimator is derived. Model selection consistency is established for the proposed robust LASSO under an adaptation procedure of the penalty weight. A parallel asymptotic analysis is derived for the Huberized LASSO, a previously proposed robust LASSO, and it is shown that the Huberized LASSO estimator preserves similar asymptotics even with a Cauchy error distribution. We show that asymptotic variances of the two robust LASSO estimators are stabilized in the presence of large variance noise, compared with the unbounded asymptotic variance of the ordinary LASSO estimator. The asymptotic analysis from the non-stochastic design is extended to the case of random design. Simulations further confirm our theoretical results.
引用
收藏
页码:5131 / 5149
页数:19
相关论文
共 49 条
[1]   NEW LOOK AT STATISTICAL-MODEL IDENTIFICATION [J].
AKAIKE, H .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 1974, AC19 (06) :716-723
[2]  
[Anonymous], 2006, Journal of the Royal Statistical Society, Series B
[3]   LIMIT OF THE SMALLEST EIGENVALUE OF A LARGE DIMENSIONAL SAMPLE COVARIANCE-MATRIX [J].
BAI, ZD ;
YIN, YQ .
ANNALS OF PROBABILITY, 1993, 21 (03) :1275-1294
[4]  
BAI ZD, 1990, CHINESE SCI A, V33, P449
[5]   SIMULTANEOUS ANALYSIS OF LASSO AND DANTZIG SELECTOR [J].
Bickel, Peter J. ;
Ritov, Ya'acov ;
Tsybakov, Alexandre B. .
ANNALS OF STATISTICS, 2009, 37 (04) :1705-1732
[6]   Sparsity oracle inequalities for the Lasso [J].
Bunea, Florentina ;
Tsybakov, Alexandre ;
Wegkamp, Marten .
ELECTRONIC JOURNAL OF STATISTICS, 2007, 1 :169-194
[7]   On Recovery of Sparse Signals Via l1 Minimization [J].
Cai, T. Tony ;
Xu, Guangwu ;
Zhang, Jun .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2009, 55 (07) :3388-3397
[8]  
CANDES E, 2008, ARXIV08010345, P1
[9]   Robust uncertainty principles:: Exact signal reconstruction from highly incomplete frequency information [J].
Candès, EJ ;
Romberg, J ;
Tao, T .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2006, 52 (02) :489-509
[10]   Decoding by linear programming [J].
Candes, EJ ;
Tao, T .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2005, 51 (12) :4203-4215