Sampling without replacement from a high-dimensional finite population

被引:1
作者
Hu, Jiang [1 ,2 ]
Wang, Shaochen [3 ]
Zhang, Yangchun [4 ]
Zhou, Wang [5 ]
机构
[1] Northeast Normal Univ, KLASMOE, Changchun 130024, Peoples R China
[2] Northeast Normal Univ, Sch Math & Stat, Changchun 130024, Peoples R China
[3] South China Univ Technol, Sch Math, Guangzhou 510640, Peoples R China
[4] Shanghai Univ, Dept Met, Shanghai 200444, Peoples R China
[5] Natl Univ Singapore, Dept Stat & Data Sci, Singapore 117546, Singapore
关键词
Largest eigenvalue; Tracy-Widom law; sample covariance matrix; finite population model; parallel analysis; TRACY-WIDOM LIMIT; COVARIANCE MATRICES; LARGEST EIGENVALUE; PRINCIPAL COMPONENTS; UNIVERSALITY; FLUCTUATIONS;
D O I
10.3150/22-BEJ1580
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
It is well known that most of the existing theoretical results in statistics are based on the assumption that the sample is generated with replacement from an infinite population. However, in practice, available samples are almost always collected without replacement. If the population is a finite set of real numbers, whether we can still safely use the results from samples drawn without replacement becomes an important problem. In this paper, we focus on the eigenvalues of high-dimensional sample covariance matrices generated without replacement from finite populations. Specifically, we derive the Tracy-Widom laws for their largest eigenvalues and apply these results to parallel analysis. We provide new insight into the permutation methods proposed by Buja and Eyuboglu in [Multivar Behav Res. 27(4) (1992) 509-540]. Simulation and real data studies are conducted to demonstrate our results.
引用
收藏
页码:3198 / 3220
页数:23
相关论文
共 50 条
  • [31] Wald Statistics in high-dimensional PCA
    Loffler, Matthias
    ESAIM-PROBABILITY AND STATISTICS, 2019, 23 : 662 - 671
  • [32] CONVERGENCE AND PREDICTION OF PRINCIPAL COMPONENT SCORES IN HIGH-DIMENSIONAL SETTINGS
    Lee, Seunggeun
    Zou, Fei
    Wright, Fred A.
    ANNALS OF STATISTICS, 2010, 38 (06) : 3605 - 3629
  • [33] RANDOM MATRIX-OPTIMIZED HIGH-DIMENSIONAL MVDR BEAMFORMING
    Yang, Liusha
    McKay, Matthew
    Couillet, Romain
    2018 IEEE STATISTICAL SIGNAL PROCESSING WORKSHOP (SSP), 2018, : 473 - 477
  • [34] Optimal recovery of precision matrix for Mahalanobis distance from high-dimensional noisy observations in manifold learning
    Gavish, Matan
    Su, Pei-Chun
    Talmon, Ronen
    Wu, Hau-Tieng
    INFORMATION AND INFERENCE-A JOURNAL OF THE IMA, 2022, 11 (04) : 1173 - 1202
  • [35] Phase II monitoring of changes in mean from high-dimensional data
    Lim, Johan
    Lee, Sungim
    APPLIED STOCHASTIC MODELS IN BUSINESS AND INDUSTRY, 2017, 33 (06) : 626 - 639
  • [36] High-dimensional limit theorems for SGD: Effective dynamics and critical scaling
    Arous, Gerard Ben
    Gheissari, Reza
    Jagannath, Aukosh
    COMMUNICATIONS ON PURE AND APPLIED MATHEMATICS, 2024, 77 (03) : 2030 - 2080
  • [37] Spectral analysis of high-dimensional sample covariance matrices with missing observations
    Jurczak, Kamil
    Rohde, Angelika
    BERNOULLI, 2017, 23 (4A) : 2466 - 2532
  • [38] Efficient Minimax Estimation of a Class of High-Dimensional Sparse Precision Matrices
    Chen, Xiaohui
    Kim, Young-Heon
    Wang, Z. Jane
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2012, 60 (06) : 2899 - 2912
  • [39] Sample canonical correlation coefficients of high-dimensional random vectors: Local law and Tracy-Widom limit
    Yang, Fan
    RANDOM MATRICES-THEORY AND APPLICATIONS, 2022, 11 (01)
  • [40] High-Dimensional Clustering via Random Projections
    Anderlucci, Laura
    Fortunato, Francesca
    Montanari, Angela
    JOURNAL OF CLASSIFICATION, 2022, 39 (01) : 191 - 216