Thresholding-based iterative selection procedures for model selection and shrinkage

被引:104
|
作者
She, Yiyuan [1 ]
机构
[1] Florida State Univ, Dept Stat, Tallahassee, FL 32306 USA
来源
ELECTRONIC JOURNAL OF STATISTICS | 2009年 / 3卷
基金
美国国家科学基金会;
关键词
Sparsity; nonconvex penalties; thresholding; model selection & shrinkage; lasso; ridge; SCAD; VARIABLE SELECTION; LASSO; REGRESSION; RECOVERY; SPARSITY; REGULARIZATION;
D O I
10.1214/08-EJS348
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
This paper discusses a class of thresholding-based iterative selection procedures (TISP) for model selection and shrinkage. People have long before noticed the weakness of the convex l(1)-constraint (or the soft-thresholding) in wavelets and have designed many different forms of nonconvex penalties to increase model sparsity and accuracy. But for a nonorthogonal regression matrix, there is great difficulty in both investigating the performance in theory and solving the problem in computation. TISP Provides a simple and efficient way to tackle this so that we successfully borrow the rich results in the orthogonal design to solve the nonconvex penalized regression for a general design matrix. Our starting point is, however, thresholding rules rather than penalty functions. Indeed, there is a universal connection between them. But a drawback of the latter is its non-unique form, and our approach greatly facilitates the computation and the analysis. In fact, we are able to build the convergence theorem and explore theoretical properties of the selection and estimation via TISP nonasymptotically. More importantly, a novel Hybrid-TISP is proposed based on hard-thresholding and ridge-thresholding. It provides a fusion between the l(0)-penalty and the l(2)-penalty, and adaptively achieves the right balance between shrinkage and selection in statistical modeling. In practice, Hybrid-TISP shows superior performance in test-error and is parsimonious.
引用
收藏
页码:384 / 415
页数:32
相关论文
共 50 条
  • [1] Group sparse recovery via group square-root elastic net and the iterative multivariate thresholding-based algorithm
    Xie, Wanling
    Yang, Hu
    ASTA-ADVANCES IN STATISTICAL ANALYSIS, 2023, 107 (03) : 469 - 507
  • [2] Model selection based on combined penalties for biomarker identification
    Vradi, Eleni
    Brannath, Werner
    Jaki, Thomas
    Vonk, Richardus
    JOURNAL OF BIOPHARMACEUTICAL STATISTICS, 2018, 28 (04) : 735 - 749
  • [3] FIRST: Combining forward iterative selection and shrinkage in high dimensional sparse linear regression
    Hwang, Wook Yeon
    Zhang, Hao Helen
    Ghosal, Subhashis
    STATISTICS AND ITS INTERFACE, 2009, 2 (03) : 341 - 348
  • [4] A Thresholding Algorithm for Order Selection in Finite Mixture Models
    Xu, Chen
    Chen, Jiahua
    COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 2015, 44 (02) : 433 - 453
  • [5] Generalized Cross Validation in variable selection with and without shrinkage
    Jansen, Maarten
    JOURNAL OF STATISTICAL PLANNING AND INFERENCE, 2015, 159 : 90 - 104
  • [6] Model selection via adaptive shrinkage with t priors
    Armagan, Artin
    Zaretzki, Russell L.
    COMPUTATIONAL STATISTICS, 2010, 25 (03) : 441 - 461
  • [7] Iterative hard thresholding for model selection in genome-wide association studies
    Keys, Kevin L.
    Chen, Gary K.
    Lange, Kenneth
    GENETIC EPIDEMIOLOGY, 2017, 41 (08) : 756 - 768
  • [8] Analysis of Testing-Based Forward Model Selection
    Kozbur, Damian
    ECONOMETRICA, 2020, 88 (05) : 2147 - 2173
  • [9] A Shrinkage-Thresholding Metropolis Adjusted Langevin Algorithm for Bayesian Variable Selection
    Schreck, Amandine
    Fort, Gersende
    Le Corff, Sylvain
    Moulines, Eric
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2016, 10 (02) : 366 - 375
  • [10] An Algorithm for Iterative Selection of Blocks of Features
    Alquier, Pierre
    ALGORITHMIC LEARNING THEORY, ALT 2010, 2010, 6331 : 35 - 49