Teaching Principal Components Using Correlations

被引:8
作者
Westfall, Peter H. [1 ]
Arias, Andrea L. [2 ,3 ]
Fulton, Lawrence V. [4 ]
机构
[1] Texas Tech Univ, Area Informat Syst & Quantitat Sci, Lubbock, TX 79409 USA
[2] Pontificia Univ Catolica Valparaiso, Sch Ind Engn, Valparaiso, Chile
[3] Texas Tech Univ, Dept Ind Engn, Lubbock, TX 79409 USA
[4] Texas Tech Univ, Area Hlth Org Management, Lubbock, TX 79409 USA
关键词
Factor analysis; heat map; optimality; rotation; variance explained; SIMILARITY;
D O I
10.1080/00273171.2017.1340824
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Introducing principal components (PCs) to students is difficult. First, the matrix algebra and mathematical maximization lemmas are daunting, especially for students in the social and behavioral sciences. Second, the standard motivation involving variance maximization subject to unit length constraint does not directly connect to the variance explained interpretation. Third, the unit length and uncorrelatedness constraints of the standard motivation do not allow re-scaling or oblique rotations, which are common in practice. Instead, we propose to motivate the subject in terms of optimizing (weighted) average proportions of variance explained in the original variables; this approach may be more intuitive, and hence easier to understand because it links directly to the familiar R-squared statistic. It also removes the need for unit length and uncorrelatedness constraints, provides a direct interpretation of variance explained, and provides a direct answer to the question of whether to use covariance-based or correlation-based PCs. Furthermore, the presentation can be made without matrix algebra or optimization proofs. Modern tools from data science, including heat maps and text mining, provide further help in the interpretation and application of PCs; examples are given. Together, these techniques may be used to revise currently used methods for teaching and learning PCs in the behavioral sciences.
引用
收藏
页码:648 / 660
页数:13
相关论文
共 38 条
[11]   The neural representation of face space dimensions [J].
Gao, Xiaoqing ;
Wilson, Hugh R. .
NEUROPSYCHOLOGIA, 2013, 51 (10) :1787-1793
[12]   Principal Component Analysis Coupled with Artificial Neural Networks-A Combined Technique Classifying Small Molecular Structures Using a Concatenated Spectral Database [J].
Gosav, Steluta ;
Praisler, Mirela ;
Birsa, Mihail Lucian .
INTERNATIONAL JOURNAL OF MOLECULAR SCIENCES, 2011, 12 (10) :6668-6684
[13]  
Härdle WK, 2019, APPLIED MULTIVARIATE STATISTICAL ANALYSIS, 5TH EDITION, P107, DOI 10.1007/978-3-030-26006-4_4
[14]   A Meta-View of Multivariate Statistical Inference Methods in European Psychology Journals [J].
Harlow, Lisa L. ;
Korendijk, Elly ;
Hamaker, Ellen L. ;
Hox, Joop ;
Duerr, Sunny R. .
MULTIVARIATE BEHAVIORAL RESEARCH, 2013, 48 (05) :749-774
[15]  
Hastie T., 2009, The Elements of Statistical Learning: Data Mining, Inference, and Prediction, P9
[16]   Analysis of a complex of statistical variables into principal components [J].
Hotelling, H .
JOURNAL OF EDUCATIONAL PSYCHOLOGY, 1933, 24 :417-441
[17]  
Johnson RA., 2007, Applied multivariate statistical analysis, sixth edition M
[18]  
Jolliffe I.T., 2002, Principal Component Analysis
[19]   Accurate and efficient classification based on common principal components analysis for multivariate time series [J].
Li, Hailin .
NEUROCOMPUTING, 2016, 171 :744-753
[20]   A review of multivariate analyses in imaging genetics [J].
Liu, Jingyu ;
Calhoun, Vince D. .
FRONTIERS IN NEUROINFORMATICS, 2014, 8