Thresholding-based iterative selection procedures for model selection and shrinkage

被引:105
作者
She, Yiyuan [1 ]
机构
[1] Florida State Univ, Dept Stat, Tallahassee, FL 32306 USA
基金
美国国家科学基金会;
关键词
Sparsity; nonconvex penalties; thresholding; model selection & shrinkage; lasso; ridge; SCAD; VARIABLE SELECTION; LASSO; REGRESSION; RECOVERY; SPARSITY; REGULARIZATION;
D O I
10.1214/08-EJS348
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
This paper discusses a class of thresholding-based iterative selection procedures (TISP) for model selection and shrinkage. People have long before noticed the weakness of the convex l(1)-constraint (or the soft-thresholding) in wavelets and have designed many different forms of nonconvex penalties to increase model sparsity and accuracy. But for a nonorthogonal regression matrix, there is great difficulty in both investigating the performance in theory and solving the problem in computation. TISP Provides a simple and efficient way to tackle this so that we successfully borrow the rich results in the orthogonal design to solve the nonconvex penalized regression for a general design matrix. Our starting point is, however, thresholding rules rather than penalty functions. Indeed, there is a universal connection between them. But a drawback of the latter is its non-unique form, and our approach greatly facilitates the computation and the analysis. In fact, we are able to build the convergence theorem and explore theoretical properties of the selection and estimation via TISP nonasymptotically. More importantly, a novel Hybrid-TISP is proposed based on hard-thresholding and ridge-thresholding. It provides a fusion between the l(0)-penalty and the l(2)-penalty, and adaptively achieves the right balance between shrinkage and selection in statistical modeling. In practice, Hybrid-TISP shows superior performance in test-error and is parsimonious.
引用
收藏
页码:384 / 415
页数:32
相关论文
共 39 条
[1]  
[Anonymous], 1997, J. Italian Stat. Soc, DOI DOI 10.1007/BF03178905
[2]   Regularization of wavelet approximations - Rejoinder [J].
Antoniadis, A ;
Fan, J .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2001, 96 (455) :964-967
[3]   Wavelet methods in statistics: Some recent developments and their applications [J].
Antoniadis, Anestis .
STATISTICS SURVEYS, 2007, 1 :16-55
[4]   CONSTRUCTION OF FIXED POINTS OF NONLINEAR MAPPINGS IN HILBERT SPACE [J].
BROWDER, FE ;
PETRYSHY.WV .
JOURNAL OF MATHEMATICAL ANALYSIS AND APPLICATIONS, 1967, 20 (02) :197-&
[5]   Sparsity oracle inequalities for the Lasso [J].
Bunea, Florentina ;
Tsybakov, Alexandre ;
Wegkamp, Marten .
ELECTRONIC JOURNAL OF STATISTICS, 2007, 1 :169-194
[6]   Hazard models with varying coefficients for multivariate failure time data [J].
Cai, Jianwen ;
Fan, Jianqing ;
Zhou, Haibo ;
Zhou, Yong .
ANNALS OF STATISTICS, 2007, 35 (01) :324-354
[7]   Rejoinder:: The Dantzig selector:: Statistical estimation when p is much larger than n [J].
Candes, Emmanuel ;
Tao, Terence .
ANNALS OF STATISTICS, 2007, 35 (06) :2392-2404
[8]  
Candès EJ, 2006, ACT NUMERIC, V15, P257, DOI 10.1017/S0962492906230010
[9]   Stable signal recovery from incomplete and inaccurate measurements [J].
Candes, Emmanuel J. ;
Romberg, Justin K. ;
Tao, Terence .
COMMUNICATIONS ON PURE AND APPLIED MATHEMATICS, 2006, 59 (08) :1207-1223
[10]   An iterative thresholding algorithm for linear inverse problems with a sparsity constraint [J].
Daubechies, I ;
Defrise, M ;
De Mol, C .
COMMUNICATIONS ON PURE AND APPLIED MATHEMATICS, 2004, 57 (11) :1413-1457