Matrix means and a novel high-dimensional shrinkage phenomenon

被引:2
作者
Lodhia, Asad [1 ]
Levin, Keith [2 ]
Levina, Elizaveta [3 ]
机构
[1] Broad Inst MIT & Harvard, 415 Main St, Cambridge, MA 02142 USA
[2] Univ Wisconsin Madison, Dept Stat, 1220 Med Sci Ctr,1300 Univ Ave, Madison, WI 53706 USA
[3] Univ Michigan, Dept Stat, 323 West Hall,1085 S Univ Ave, Ann Arbor, MI 48109 USA
基金
美国国家科学基金会;
关键词
Matrix means; positive definite cone; covariance estimation; shrinkage Estimators; COVARIANCE-MATRIX; PRINCIPAL-COMPONENTS; OPTIMAL RATES; PRECISION MATRICES; CONVERGENCE; EIGENVALUE; ESTIMATORS; ALGORITHMS; GEOMETRY;
D O I
10.3150/21-BEJ1430
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Many statistical settings call for estimating a population parameter, most typically the population mean, based on a sample of matrices. The most natural estimate of the population mean is the arithmetic mean, but there are many other matrix means that may behave differently, especially in high dimensions. Here we consider the matrix har-monic mean as an alternative to the arithmetic matrix mean. We show that in certain high-dimensional regimes, the harmonic mean yields an improvement over the arithmetic mean in estimation error as measured by the opera-tor norm. Counter-intuitively, studying the asymptotic behavior of these two matrix means in a spiked covariance estimation problem, we find that this improvement in operator norm error does not imply better recovery of the leading eigenvector. We also show that a Rao-Blackwellized version of the harmonic mean is equivalent to a linear shrinkage estimator studied previously in the high-dimensional covariance estimation literature, while applying a similar Rao-Blackwellization to regularized sample covariance matrices yields a novel nonlinear shrinkage estima-tor. Simulations complement the theoretical results, illustrating the conditions under which the harmonic matrix mean yields an empirically better estimate.
引用
收藏
页码:2578 / 2605
页数:28
相关论文
共 61 条
[1]   Riemannian geometry of Grassmann manifolds with a view on algorithmic computation [J].
Absil, PA ;
Mahony, R ;
Sepulchre, R .
ACTA APPLICANDAE MATHEMATICAE, 2004, 80 (02) :199-220
[2]  
Anderson T. W., 2003, An Introduction to Multivariate Statistical Analysis
[3]  
[Anonymous], 1967, Mat. Sb. New Series, DOI DOI 10.1016/j.neuroimage.2008.09.062
[4]  
Bai Z, 2010, SPRINGER SER STAT, P1, DOI 10.1007/978-1-4419-0661-8
[5]   Phase transition of the largest eigenvalue for nonnull complex sample covariance matrices [J].
Baik, J ;
Ben Arous, G ;
Péché, S .
ANNALS OF PROBABILITY, 2005, 33 (05) :1643-1697
[6]   The eigenvalues and eigenvectors of finite, low rank perturbations of large random matrices [J].
Benaych-Georges, Florent ;
Nadakuditi, Raj Rao .
ADVANCES IN MATHEMATICS, 2011, 227 (01) :494-521
[7]   OPTIMAL DETECTION OF SPARSE PRINCIPAL COMPONENTS IN HIGH DIMENSION [J].
Berthet, Quentin ;
Rigollet, Philippe .
ANNALS OF STATISTICS, 2013, 41 (04) :1780-1815
[8]  
Bhatia R, 2007, PRINC SER APPL MATH, P1
[9]   Regularized estimation of large covariance matrices [J].
Bickel, Peter J. ;
Levina, Elizaveta .
ANNALS OF STATISTICS, 2008, 36 (01) :199-227
[10]   COVARIANCE REGULARIZATION BY THRESHOLDING [J].
Bickel, Peter J. ;
Levina, Elizaveta .
ANNALS OF STATISTICS, 2008, 36 (06) :2577-2604