ADAPTIVE ROBUST VARIABLE SELECTION

被引:171
作者
Fan, Jianqing [1 ]
Fan, Yingying [2 ]
Barut, Emre [3 ]
机构
[1] Princeton Univ, Dept Operat Res & Financial Engn, Princeton, NJ 08544 USA
[2] Univ So Calif, Data Sci & Operat Dept, Marshall Sch Business, Los Angeles, CA 90089 USA
[3] IBM Corp, TJ Watson Res Ctr, Yorktown Hts, NY 10598 USA
基金
美国国家科学基金会; 美国国家卫生研究院;
关键词
Adaptive weighted L-1; high dimensions; oracle properties; robust regularization; NONCONCAVE PENALIZED LIKELIHOOD; QUANTILE REGRESSION; MODEL SELECTION; QUASI-LIKELIHOOD; SHRINKAGE; LASSO;
D O I
10.1214/13-AOS1191
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Heavy-tailed high-dimensional data are commonly encountered in various scientific fields and pose great challenges to modern statistical analysis. A natural procedure to address this problem is to use penalized quantile regression with weighted L-1-penalty, called weighted robust Lasso (WR-Lasso), in which weights are introduced to ameliorate the bias problem induced by the L-1-penalty. In the ultra-high dimensional setting, where the dimensionality can grow exponentially with the sample size, we investigate the model selection oracle property and establish the asymptotic normality of the WR-Lasso. We show that only mild conditions on the model error distribution are needed. Our theoretical results also reveal that adaptive choice of the weight vector is essential for the WR-Lasso to enjoy these nice asymptotic properties. To make the WR-Lasso practically feasible, we propose a two-step procedure, called adaptive robust Lasso (AR-Lasso), in which the weight vector in the second step is constructed based on the L-1-penalized quantile regression estimate from the first step. This two-step procedure is justified theoretically to possess the oracle property and the asymptotic normality. Numerical studies demonstrate the favorable finite-sample performance of the AR-Lasso.
引用
收藏
页码:324 / 351
页数:28
相关论文
共 28 条
[1]  
[Anonymous], 2012, Stable distributions
[2]   l1-PENALIZED QUANTILE REGRESSION IN HIGH-DIMENSIONAL SPARSE MODELS [J].
Belloni, Alexandre ;
Chernozhukov, Victor .
ANNALS OF STATISTICS, 2011, 39 (01) :82-130
[3]   Regularization in statistics [J].
Bickel, Peter J. ;
Li, Bo .
TEST, 2006, 15 (02) :271-303
[4]   SIMULTANEOUS ANALYSIS OF LASSO AND DANTZIG SELECTOR [J].
Bickel, Peter J. ;
Ritov, Ya'acov ;
Tsybakov, Alexandre B. .
ANNALS OF STATISTICS, 2009, 37 (04) :1705-1732
[5]   Penalized composite quasi-likelihood for ultrahigh dimensional variable selection [J].
Bradic, Jelena ;
Fan, Jianqing ;
Wang, Weiwei .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2011, 73 :325-349
[6]  
Bühlmann P, 2011, SPRINGER SER STAT, P1, DOI 10.1007/978-3-642-20192-9
[7]  
Candes E, 2007, ANN STAT, V35, P2313, DOI 10.1214/009053606000001523
[8]  
FAN J., 2014, ADAPTIVE ROBUST VA S, DOI [10.1214/13-AOS1191SUPP, DOI 10.1214/13-A0S1191SUPP]
[9]   Sure independence screening for ultrahigh dimensional feature space [J].
Fan, Jianqing ;
Lv, Jinchi .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2008, 70 :849-883
[10]   Nonconcave Penalized Likelihood With NP-Dimensionality [J].
Fan, Jianqing ;
Lv, Jinchi .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2011, 57 (08) :5467-5484