Automated learning of mixtures of factor analysis models with missing information

被引:0
作者
Wan-Lun Wang
Tsung-I Lin
机构
[1] Feng Chia University,Department of Statistics, Graduate Institute of Statistics and Actuarial Science
[2] National Chung Hsing University,Institute of Statistics
[3] China Medical University,Department of Public Health
来源
TEST | 2020年 / 29卷
关键词
Automated learning; Factor analysis; Maximum likelihood estimation; Missing values; Model selection; One-stage algorithm; 62H12; 62H25; 62H30;
D O I
暂无
中图分类号
学科分类号
摘要
The mixture of factor analyzers (MFA) model has emerged as a useful tool to perform dimensionality reduction and model-based clustering for heterogeneous data. In seeking the most appropriate number of factors (q) of a MFA model with the number of components (g) fixed a priori, a two-stage procedure is commonly implemented by firstly carrying out parameter estimation over a set of prespecified numbers of factors, and then selecting the best q according to certain penalized likelihood criteria. When the dimensionality of data grows higher, such a procedure can be computationally prohibitive. To overcome this obstacle, we develop an automated learning scheme, called the automated MFA (AMFA) algorithm, to effectively merge parameter estimation and selection of q into a one-stage algorithm. The proposed AMFA procedure that allows for much lower computational cost is also extended to accommodate missing values. Moreover, we explicitly derive the score vector and the empirical information matrix for calculating standard errors associated with the estimated parameters. The potential and applicability of the proposed method are demonstrated through a number of real datasets with genuine and synthetic missing values.
引用
收藏
页码:1098 / 1124
页数:26
相关论文
共 50 条
[41]   Learning causal structure from mixed data with missing values using Gaussian copula models [J].
Cui, Ruifei ;
Groot, Perry ;
Heskes, Tom .
STATISTICS AND COMPUTING, 2019, 29 (02) :311-333
[42]   Extending mixtures of factor models using the restricted multivariate skew-normal distribution [J].
Lin, Tsung-I ;
McLachlan, Geoffrey J. ;
Lee, Sharon X. .
JOURNAL OF MULTIVARIATE ANALYSIS, 2016, 143 :398-413
[43]   A Two-Tier Full-Information Item Factor Analysis Model with Applications [J].
Cai, Li .
PSYCHOMETRIKA, 2010, 75 (04) :581-612
[44]   Automated selection of nanoparticle models for small-angle X-ray scattering data analysis using machine learning [J].
Monge, Nicolas ;
Deschamps, Alexis ;
Amini, Massih-Reza .
ACTA CRYSTALLOGRAPHICA A-FOUNDATION AND ADVANCES, 2024, 80 :202-212
[45]   Local influence analysis of nonlinear structural equation models with nonignorable missing outcomes from reproductive dispersion models [J].
Fu, Ying-Zi ;
Tang, Nian-Sheng ;
Chen, Xing .
COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2009, 53 (10) :3671-3684
[46]   INFORMATION-THEORETIC ANALYSIS OF STOCHASTIC VOLATILITY MODELS [J].
Pfante, Oliver ;
Bertschinger, Nils .
ADVANCES IN COMPLEX SYSTEMS, 2019, 22 (01)
[47]   Can Factor Analysis be Applied to Spectra Taken in Binary Solvent Mixtures? [J].
Thurler, Thalyta S. ;
da Souza, Raquel G. S. ;
Maeder, Marcel ;
Norman, Sarah E. ;
Brinn, Ira M. .
JOURNAL OF THE BRAZILIAN CHEMICAL SOCIETY, 2009, 20 (09) :1617-1624
[48]   Bias correction of the Akaike information criterion in factor analysis [J].
Ogasawara, Haruhiko .
JOURNAL OF MULTIVARIATE ANALYSIS, 2016, 149 :144-159
[49]   Deep Learning with Hierarchical Convolutional Factor Analysis [J].
Chen, Bo ;
Polatkan, Gungor ;
Sapiro, Guillermo ;
Blei, David ;
Dunson, David ;
Carin, Lawrence .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (08) :1887-1901
[50]   Sparse Factor Analysis for Learning and Content Analytics [J].
Lan, Andrew S. ;
Waters, Andrew E. ;
Studer, Christoph ;
Baraniuk, Richard G. .
JOURNAL OF MACHINE LEARNING RESEARCH, 2014, 15 :1959-2008