NONPENALIZED VARIABLE SELECTION IN HIGH-DIMENSIONAL LINEAR MODEL SETTINGS VIA GENERALIZED FIDUCIAL INFERENCE

被引:9
|
作者
Williams, Jonathan P. [1 ]
Hannig, Jan [1 ]
机构
[1] Univ N Carolina, Dept Stat & Operat Res, Chapel Hill, NC 27599 USA
基金
美国国家科学基金会;
关键词
Best subset selection; high-dimensional regression; L-0; minimization; feature selection; REGRESSION;
D O I
10.1214/18-AOS1733
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Standard penalized methods of variable selection and parameter estimation rely on the magnitude of coefficient estimates to decide which variables to include in the final model. However, coefficient estimates are unreliable when the design matrix is collinear. To overcome this challenge, an entirely new perspective on variable selection is presented within a generalized fiducial inference framework. This new procedure is able to effectively account for linear dependencies among subsets of covariates in a high-dimensional setting where p can grow almost exponentially in n, as well as in the classical setting where p <= n. It is shown that the procedure very naturally assigns small probabilities to subsets of covariates which include redundancies by way of explicit L-0 minimization Furthermore, with a typical sparsity assumption, it is shown that the proposed method is consistent in the sense that the probability of the true sparse subset of covariates converges in probability to 1 as n -> infinity, or as n -> infinity and p -> infinity. Very reasonable conditions are needed, and little restriction is placed on the class of possible subsets of covariates to achieve this consistency result.
引用
收藏
页码:1723 / 1753
页数:31
相关论文
共 50 条
  • [31] High-dimensional feature selection via feature grouping: A Variable Neighborhood Search approach
    Garcia-Torres, Miguel
    Gomez-Vela, Francisco
    Melian-Batista, Belen
    Marcos Moreno-Vega, J.
    INFORMATION SCIENCES, 2016, 326 : 102 - 118
  • [32] Learning High-Dimensional Generalized Linear Autoregressive Models
    Hall, Eric C.
    Raskutti, Garvesh
    Willett, Rebecca M.
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2019, 65 (04) : 2401 - 2422
  • [33] A group adaptive elastic-net approach for variable selection in high-dimensional linear regression
    Jianhua Hu
    Jian Huang
    Feng Qiu
    Science China(Mathematics), 2018, 61 (01) : 173 - 188
  • [34] Cluster feature selection in high-dimensional linear models
    Lin, Bingqing
    Pang, Zhen
    Wang, Qihua
    RANDOM MATRICES-THEORY AND APPLICATIONS, 2018, 7 (01)
  • [35] A group adaptive elastic-net approach for variable selection in high-dimensional linear regression
    Jianhua Hu
    Jian Huang
    Feng Qiu
    Science China Mathematics, 2018, 61 : 173 - 188
  • [36] Orthogonality-projection-based penalized variable selection for high-dimensional partially linear models
    Yang, Yiping
    Zhao, Peixin
    Zhang, Jun
    APPLIED MATHEMATICAL MODELLING, 2025, 138
  • [37] A group adaptive elastic-net approach for variable selection in high-dimensional linear regression
    Hu, Jianhua
    Huang, Jian
    Qiu, Feng
    SCIENCE CHINA-MATHEMATICS, 2018, 61 (01) : 173 - 188
  • [38] A WEIGHT-RELAXED MODEL AVERAGING APPROACH FOR HIGH-DIMENSIONAL GENERALIZED LINEAR MODELS
    Ando, Tomohiro
    Li, Ker-Chau
    ANNALS OF STATISTICS, 2017, 45 (06) : 2654 - 2679
  • [39] High-dimensional Bayesian inference via the unadjusted Langevin algorithm
    Durmus, Alain
    Moulines, Eric
    BERNOULLI, 2019, 25 (4A) : 2854 - 2882
  • [40] Feature Augmentation via Nonparametrics and Selection (FANS) in High-Dimensional Classification
    Fan, Jianqing
    Feng, Yang
    Jiang, Jiancheng
    Tong, Xin
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2016, 111 (513) : 275 - 287