Sampling without replacement from a high-dimensional finite population

被引:1
作者
Hu, Jiang [1 ,2 ]
Wang, Shaochen [3 ]
Zhang, Yangchun [4 ]
Zhou, Wang [5 ]
机构
[1] Northeast Normal Univ, KLASMOE, Changchun 130024, Peoples R China
[2] Northeast Normal Univ, Sch Math & Stat, Changchun 130024, Peoples R China
[3] South China Univ Technol, Sch Math, Guangzhou 510640, Peoples R China
[4] Shanghai Univ, Dept Met, Shanghai 200444, Peoples R China
[5] Natl Univ Singapore, Dept Stat & Data Sci, Singapore 117546, Singapore
关键词
Largest eigenvalue; Tracy-Widom law; sample covariance matrix; finite population model; parallel analysis; TRACY-WIDOM LIMIT; COVARIANCE MATRICES; LARGEST EIGENVALUE; PRINCIPAL COMPONENTS; UNIVERSALITY; FLUCTUATIONS;
D O I
10.3150/22-BEJ1580
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
It is well known that most of the existing theoretical results in statistics are based on the assumption that the sample is generated with replacement from an infinite population. However, in practice, available samples are almost always collected without replacement. If the population is a finite set of real numbers, whether we can still safely use the results from samples drawn without replacement becomes an important problem. In this paper, we focus on the eigenvalues of high-dimensional sample covariance matrices generated without replacement from finite populations. Specifically, we derive the Tracy-Widom laws for their largest eigenvalues and apply these results to parallel analysis. We provide new insight into the permutation methods proposed by Buja and Eyuboglu in [Multivar Behav Res. 27(4) (1992) 509-540]. Simulation and real data studies are conducted to demonstrate our results.
引用
收藏
页码:3198 / 3220
页数:23
相关论文
共 50 条
  • [41] Two sample tests for high-dimensional autocovariances
    Baek, Changryong
    Gates, Katheleen M.
    Leinwand, Benjamin
    Pipiras, Vladas
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2021, 153
  • [42] Interpolating Predictors in High-Dimensional Factor Regression
    Bunea, Florentina
    Strimas-Mackey, Seth
    Wegkamp, Marten
    JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [43] Sparse PCA for High-Dimensional Data With Outliers
    Hubert, Mia
    Reynkens, Tom
    Schmitt, Eric
    Verdonck, Tim
    TECHNOMETRICS, 2016, 58 (04) : 424 - 434
  • [44] Projected tests for high-dimensional covariance matrices
    Wu, Tung-Lung
    Li, Ping
    JOURNAL OF STATISTICAL PLANNING AND INFERENCE, 2020, 207 : 73 - 85
  • [45] Inference for high-dimensional differential correlation matrices
    Cai, T. Tony
    Zhang, Anru
    JOURNAL OF MULTIVARIATE ANALYSIS, 2016, 143 : 107 - 126
  • [46] HIGH-DIMENSIONAL FACTOR REGRESSION FOR HETEROGENEOUS SUBPOPULATIONS
    Wang, Peiyao
    Li, Quefeng
    Shen, Dinggan
    Liu, Yufeng
    STATISTICA SINICA, 2023, 33 (01) : 27 - 53
  • [47] Testing high-dimensional mean vector with applications
    Zhang, Jin-Ting
    Zhou, Bu
    Guo, Jia
    STATISTICAL PAPERS, 2022, 63 (04) : 1105 - 1137
  • [48] On statistical inference with high-dimensional sparse CCA
    Laha, Nilanjana
    Huey, Nathan
    Coull, Brent
    Mukherjee, Rajarshi
    INFORMATION AND INFERENCE-A JOURNAL OF THE IMA, 2023, 12 (04)
  • [49] Phase transitions of spectral initialization for high-dimensional non-convex estimation
    Lu, Yue M.
    Li, Gen
    INFORMATION AND INFERENCE-A JOURNAL OF THE IMA, 2020, 9 (03) : 507 - 541
  • [50] A UNIVERSAL TEST ON SPIKES IN A HIGH-DIMENSIONAL GENERALIZED SPIKED MODEL AND ITS APPLICATIONS
    Jiang, Dandan
    STATISTICA SINICA, 2023, 33 : 1749 - 1770