Honest Confidence Sets for High-Dimensional Regression by Projection and Shrinkage

被引:0
|
作者
Zhou, Kun [1 ]
Li, Ker-Chau [1 ,2 ]
Zhou, Qing [1 ]
机构
[1] Univ Calif Los Angeles, Dept Stat, Los Angeles, CA 90095 USA
[2] Acad Sinica, Inst Stat Sci, Nangang, Taiwan
基金
美国国家科学基金会;
关键词
Adaptive confidence set; High-dimensional inference; Sparse linear regression; Stein estimate; SIMULTANEOUS INFERENCE; INTERVALS; LASSO; ESTIMATORS; SELECTION; REGIONS; RATES;
D O I
10.1080/01621459.2021.1938581
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
The issue of honesty in constructing confidence sets arises in nonparametric regression. While optimal rate in nonparametric estimation can be achieved and utilized to construct sharp confidence sets, severe degradation of confidence level often happens after estimating the degree of smoothness. Similarly, for high-dimensional regression, oracle inequalities for sparse estimators could be utilized to construct sharp confidence sets. Yet, the degree of sparsity itself is unknown and needs to be estimated, which causes the honesty problem. To resolve this issue, we develop a novel method to construct honest confidence sets for sparse high-dimensional linear regression. The key idea in our construction is to separate signals into a strong and a weak group, and then construct confidence sets for each group separately. This is achieved by a projection and shrinkage approach, the latter implemented via Stein estimation and the associated Stein unbiased risk estimate. Our confidence set is honest over the full parameter space without any sparsity constraints, while its size adapts to the optimal rate of n(-1/4) when the true parameter is indeed sparse. Moreover, under some form of a separation assumption between the strong and weak signals, the diameter of our confidence set can achieve a faster rate than existing methods. Through extensive numerical comparisons on both simulated and real data, we demonstrate that our method outperforms other competitors with bigmargins for finite samples, including oracle methods built upon the true sparsity of the underlying model.
引用
收藏
页码:469 / 488
页数:20
相关论文
共 50 条
  • [21] REGULARIZED PROJECTION SCORE ESTIMATION OF TREATMENT EFFECTS IN HIGH-DIMENSIONAL QUANTILE REGRESSION
    Cheng, Chao
    Feng, Xingdong
    Huang, Jian
    Liu, Xu
    STATISTICA SINICA, 2022, 32 (01) : 23 - 41
  • [22] Bayesian stein-type shrinkage estimators in high-dimensional linear regression models
    Zanboori, Ahmadreza
    Zanboori, Ehsan
    Mousavi, Maryam
    Mirjalili, Sayyed Mahmoud
    SAO PAULO JOURNAL OF MATHEMATICAL SCIENCES, 2024, 18 (02): : 1889 - 1914
  • [23] Sparse signal shrinkage and outlier detection in high-dimensional quantile regression with variational Bayes
    Lim, Daeyoung
    Park, Beomjo
    Nott, David
    Wang, Xueou
    Choi, Taeryon
    STATISTICS AND ITS INTERFACE, 2020, 13 (02) : 237 - 249
  • [24] High-Dimensional LASSO-Based Computational Regression Models: Regularization, Shrinkage, and Selection
    Emmert-Streib, Frank
    Dehmer, Matthias
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2019, 1 (01): : 359 - 383
  • [25] Regression on High-dimensional Inputs
    Kuleshov, Alexander
    Bernstein, Alexander
    2016 IEEE 16TH INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW), 2016, : 732 - 739
  • [26] On inference in high-dimensional regression
    Battey, Heather S.
    Reid, Nancy
    JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2023, 85 (01) : 149 - 175
  • [28] Converting high-dimensional regression to high-dimensional conditional density estimation
    Izbicki, Rafael
    Lee, Ann B.
    ELECTRONIC JOURNAL OF STATISTICS, 2017, 11 (02): : 2800 - 2831
  • [29] Fixed-Size Confidence Regions in High-Dimensional Sparse Linear Regression Models
    Ing, Ching-Kang
    Lai, Tze Leung
    SEQUENTIAL ANALYSIS-DESIGN METHODS AND APPLICATIONS, 2015, 34 (03): : 324 - 335