Sparse factor regression via penalized maximum likelihood estimation

被引:6
作者
Hirose, Kei [1 ]
Imada, Miyuki [2 ]
机构
[1] Kyushu Univ, Inst Math Ind, Nishi Ku, 744 Motooka, Fukuoka 8190395, Japan
[2] NTT Network Innovat Labs, 3-9-11 Midori Cho, Musashino, Tokyo 1808585, Japan
关键词
Coordinate descent algorithm; Lasso; Multicollinearity; Penalized maximum likelihood estimation; Rotation technique; PREDICTION ERROR PROPERTY; VARIABLE SELECTION; IMPROPER SOLUTIONS; LASSO ESTIMATOR; REGULARIZATION; ALGORITHMS; ROTATION; MODELS;
D O I
10.1007/s00362-016-0781-8
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
In factor regression model, the maximum likelihood estimation suffers from three disadvantages: (i) the maximum likelihood estimates are unavailable when the number of variables exceeds the number of observations, (ii) the rotation technique based on maximum likelihood estimates produces an insufficiently sparse loading matrix, and (iii) multicollinearity can occur when the estimates of unique variances (specific variances) are small because the regression coefficients are sensitive to the inverse of unique variances. To handle these problems, we propose a penalized maximum likelihood procedure. Specifically, we impose a lasso-type penalty on the factor loadings to improve the sparseness of the solution. We also introduce a penalty on unique variances, which (given the factor scores) corresponds to the ridge penalty on the regression coefficient. Theoretical properties from a prediction viewpoint of our procedure are discussed. The effectiveness of the procedure is investigated through Monte Carlo simulations. The utility of our procedure is demonstrated on real data collected by an online questionnaire.
引用
收藏
页码:633 / 662
页数:30
相关论文
共 36 条
[1]   FACTOR-ANALYSIS AND AIC [J].
AKAIKE, H .
PSYCHOMETRIKA, 1987, 52 (03) :317-332
[2]   THE EFFECT OF SAMPLING ERROR ON CONVERGENCE, IMPROPER SOLUTIONS, AND GOODNESS-OF-FIT INDEXES FOR MAXIMUM-LIKELIHOOD CONFIRMATORY FACTOR-ANALYSIS [J].
ANDERSON, JC ;
GERBING, DW .
PSYCHOMETRIKA, 1984, 49 (02) :155-173
[3]  
Anderson T.W., 1956, P 3 BERKELEY S MATH, V5, P1
[4]   STATISTICAL ANALYSIS OF FACTOR MODELS OF HIGH DIMENSION [J].
Bai, Jushan ;
Li, Kunpeng .
ANNALS OF STATISTICS, 2012, 40 (01) :436-465
[5]   SIMULTANEOUS ANALYSIS OF LASSO AND DANTZIG SELECTOR [J].
Bickel, Peter J. ;
Ritov, Ya'acov ;
Tsybakov, Alexandre B. .
ANNALS OF STATISTICS, 2009, 37 (04) :1705-1732
[6]  
Bishop Christopher M, 2016, Pattern recognition and machine learning
[7]   COORDINATE DESCENT ALGORITHMS FOR NONCONVEX PENALIZED REGRESSION, WITH APPLICATIONS TO BIOLOGICAL FEATURE SELECTION [J].
Breheny, Patrick ;
Huang, Jian .
ANNALS OF APPLIED STATISTICS, 2011, 5 (01) :232-253
[8]   High-Dimensional Sparse Factor Modeling: Applications in Gene Expression Genomics [J].
Carvalho, Carlos M. ;
Chang, Jeffrey ;
Lucas, Joseph E. ;
Nevins, Joseph R. ;
Wang, Quanli ;
West, Mike .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2008, 103 (484) :1438-1456
[9]  
Choi J, 2010, STAT INTERFACE, V3, P429
[10]   Sparse partial least squares regression for simultaneous dimension reduction and variable selection [J].
Chun, Hyonho ;
Keles, Suenduez .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2010, 72 :3-25