Honest Confidence Sets for High-Dimensional Regression by Projection and Shrinkage

被引:0
|
作者
Zhou, Kun [1 ]
Li, Ker-Chau [1 ,2 ]
Zhou, Qing [1 ]
机构
[1] Univ Calif Los Angeles, Dept Stat, Los Angeles, CA 90095 USA
[2] Acad Sinica, Inst Stat Sci, Nangang, Taiwan
基金
美国国家科学基金会;
关键词
Adaptive confidence set; High-dimensional inference; Sparse linear regression; Stein estimate; SIMULTANEOUS INFERENCE; INTERVALS; LASSO; ESTIMATORS; SELECTION; REGIONS; RATES;
D O I
10.1080/01621459.2021.1938581
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
The issue of honesty in constructing confidence sets arises in nonparametric regression. While optimal rate in nonparametric estimation can be achieved and utilized to construct sharp confidence sets, severe degradation of confidence level often happens after estimating the degree of smoothness. Similarly, for high-dimensional regression, oracle inequalities for sparse estimators could be utilized to construct sharp confidence sets. Yet, the degree of sparsity itself is unknown and needs to be estimated, which causes the honesty problem. To resolve this issue, we develop a novel method to construct honest confidence sets for sparse high-dimensional linear regression. The key idea in our construction is to separate signals into a strong and a weak group, and then construct confidence sets for each group separately. This is achieved by a projection and shrinkage approach, the latter implemented via Stein estimation and the associated Stein unbiased risk estimate. Our confidence set is honest over the full parameter space without any sparsity constraints, while its size adapts to the optimal rate of n(-1/4) when the true parameter is indeed sparse. Moreover, under some form of a separation assumption between the strong and weak signals, the diameter of our confidence set can achieve a faster rate than existing methods. Through extensive numerical comparisons on both simulated and real data, we demonstrate that our method outperforms other competitors with bigmargins for finite samples, including oracle methods built upon the true sparsity of the underlying model.
引用
收藏
页码:469 / 488
页数:20
相关论文
共 50 条
  • [31] Converting high-dimensional regression to high-dimensional conditional density estimation
    Izbicki, Rafael
    Lee, Ann B.
    ELECTRONIC JOURNAL OF STATISTICS, 2017, 11 (02): : 2800 - 2831
  • [32] Confidence Intervals and Hypothesis Testing for High-dimensional Quantile Regression: Convolution Smoothing and Debiasing
    Yan, Yibo
    Wang, Xiaozhou
    Zhang, Riquan
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [33] HIGH-DIMENSIONAL REGRESSION WITH NOISY AND MISSING DATA: PROVABLE GUARANTEES WITH NONCONVEXITY
    Loh, Po-Ling
    Wainwright, Martin J.
    ANNALS OF STATISTICS, 2012, 40 (03) : 1637 - 1664
  • [34] SCAD-PENALIZED REGRESSION IN HIGH-DIMENSIONAL PARTIALLY LINEAR MODELS
    Xie, Huiliang
    Huang, Jian
    ANNALS OF STATISTICS, 2009, 37 (02) : 673 - 696
  • [35] Shrinkage priors for high-dimensional demand estimation
    Smith, Adam N.
    Griffin, Jim E.
    QME-QUANTITATIVE MARKETING AND ECONOMICS, 2023, 21 (01): : 95 - 146
  • [36] Confidence Intervals and Tests for High-Dimensional Models: A Compact Review
    Buhlmann, Peter
    MODELING AND STOCHASTIC LEARNING FOR FORECASTING IN HIGH DIMENSIONS, 2015, 217 : 21 - 34
  • [37] Confidence intervals for parameters in high-dimensional sparse vector autoregression
    Zhu, Ke
    Liu, Hanzhong
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2022, 168
  • [38] ON ASYMPTOTICALLY OPTIMAL CONFIDENCE REGIONS AND TESTS FOR HIGH-DIMENSIONAL MODELS
    Van de Geer, Sara
    Buehlmann, Peter
    Ritov, Ya'acov
    Dezeure, Ruben
    ANNALS OF STATISTICS, 2014, 42 (03) : 1166 - 1202
  • [39] Sparsified simultaneous confidence intervals for high-dimensional linear models
    Zhu, Xiaorui
    Qin, Yichen
    Wang, Peng
    METRIKA, 2024,
  • [40] Robust Estimation of High-Dimensional Linear Regression With Changepoints
    Cui, Xiaolong
    Geng, Haoyu
    Wang, Zhaojun
    Zou, Changliang
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2024, 70 (10) : 7297 - 7319