Multidimensional linear functional estimation in sparse Gaussian models and robust estimation of the mean

被引:5
作者
Collier, Olivier [1 ,2 ]
Dalalyan, Arnak S. [2 ]
机构
[1] Univ Paris Nanterre, ModalX, Nanterre, France
[2] ENSAE, CREST, Palaiseau, France
来源
ELECTRONIC JOURNAL OF STATISTICS | 2019年 / 13卷 / 02期
关键词
Column-sparsity; Minimax estimation; Group-sparsity; Linear transformation; High-dimensional inference; Robust estimation; ADAPTIVE ESTIMATION; MINIMAX ESTIMATION; SELECTION; REGRESSION; LASSO; SHARP;
D O I
10.1214/19-EJS1590
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We consider two problems of estimation in high-dimensional Gaussian models. The first problem is that of estimating a linear functional of the means of n independent p-dimensional Gaussian vectors, under the assumption that at most s of the means are nonzero. We show that, up to a logarithmic factor, the minimax rate of estimation in squared Euclidean norm is between (s(2) boolean AND n) + sp and (s(2) boolean AND np) + sp. The estimator that attains the upper bound being computationally demanding, we investigate suitable versions of group thresholding estimators that are efficiently computable even when the dimension and the sample size are very large. An interesting new phenomenon revealed by this investigation is that the group thresholding leads to a substantial improvement in the rate as compared to the element-wise thresholding. Thus, the rate of the group thresholding is s(2) root p + sp, while the element-wise thresholding has an error of order s(2)p + sp. To the best of our knowledge, this is the first known setting in which leveraging the group structure leads to a polynomial improvement in the rate. The second problem studied in this work is the estimation of the common p-dimensional mean of the inliers among n independent Gaussian vectors. We show that there is a strong analogy between this problem and the first one. Exploiting it, we propose new strategies of robust estimation that are computationally tractable and have better rates of convergence than the other computationally tractable robust (with respect to the presence of the outliers in the data) estimators studied in the literature. However, this tractability comes with a loss of the minimax-rate-optimality in some regimes.
引用
收藏
页码:2830 / 2864
页数:35
相关论文
共 50 条
  • [1] [Anonymous], 2006, J ROYAL STAT SOC B
  • [2] [Anonymous], 1997, Math. Methods Statist
  • [3] Balakrishnan Sivaraman, 2017, C LEARN THEOR, P169
  • [4] Balmand S., 2015, 151204734 ARXIV
  • [5] BICKEL PJ, 1988, SANKHYA SER A, V50, P381
  • [6] The Group Square-Root Lasso: Theoretical Properties and Fast Algorithms
    Bunea, Florentina
    Lederer, Johannes
    She, Yiyuan
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2014, 60 (02) : 1313 - 1325
  • [7] Adaptive estimation of linear functionals in the convolution model and applications
    Butucea, C.
    Comte, F.
    [J]. BERNOULLI, 2009, 15 (01) : 69 - 98
  • [8] Optimal adaptive estimation of a quadratic functional
    Cai, T. Tony
    Low, Mark G.
    [J]. ANNALS OF STATISTICS, 2006, 34 (05) : 2298 - 2325
  • [9] TESTING COMPOSITE HYPOTHESES, HERMITE POLYNOMIALS AND OPTIMAL ESTIMATION OF A NONSMOOTH FUNCTIONAL
    Cai, T. Tony
    Low, Mark G.
    [J]. ANNALS OF STATISTICS, 2011, 39 (02) : 1012 - 1041
  • [10] On adaptive estimation of linear functionals
    Cai, TT
    Low, MG
    [J]. ANNALS OF STATISTICS, 2005, 33 (05) : 2311 - 2343