Mixed Regression via Approximate Message Passing

被引:0
作者
Tan, Nelvin [1 ]
Venkataramanan, Ramji [1 ]
机构
[1] Univ Cambridge, Dept Engn, Cambridge CB2 1PZ, England
关键词
Approximate Message Passing; Mixed Linear Regression; Max-Affine Re-; gression; Mixture-of-Experts; Expectation-Maximization; MULTIVARIATE CONVEX REGRESSION; EM ALGORITHM; MIXTURE; GUARANTEES;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We study the problem of regression in a generalized linear model (GLM) with multiple signals and latent variables. This model, which we call a matrix GLM, covers many widely studied problems in statistical learning, including mixed linear regression, max-affine regression, and mixture-of-experts. The goal in all these problems is to estimate the signals, and possibly some of the latent variables, from the observations. We propose a novel approximate message passing (AMP) algorithm for estimation in a matrix GLM and rigorously characterize its performance in the high-dimensional limit. This characterization is in terms of a state evolution recursion, which allows us to precisely compute performance measures such as the asymptotic mean-squared error. The state evolution characterization can be used to tailor the AMP algorithm to take advantage of any structural information known about the signals. Using state evolution, we derive an optimal choice of AMP 'denoising' functions that minimizes the estimation error in each iteration. The theoretical results are validated by numerical simulations for mixed linear regression, max-affine regression, and mixture-of-experts. For max-affine regression, we propose an algorithm that combines AMP with expectation-maximization to estimate the intercepts of the model along with the signals. The numerical results show that AMP significantly outperforms other estimators for mixed linear regression and max-affine regression in most parameter regimes.
引用
收藏
页数:44
相关论文
共 95 条
[1]  
[Anonymous], 2018, Advances in Neural Information Processing Systems (NeurIPS)
[2]  
Arpino G, 2023, Arxiv, DOI arXiv:2303.02118
[3]   STATISTICAL GUARANTEES FOR THE EM ALGORITHM: FROM POPULATION TO SAMPLE-BASED ANALYSIS [J].
Balakrishnan, Sivaraman ;
Wainwrightt, Martin J. ;
Yu, Bin .
ANNALS OF STATISTICS, 2017, 45 (01) :77-120
[4]  
Balazs G., 2016, Convex regression: Theory, practice, and applications
[5]  
Balázs G, 2015, JMLR WORKSH CONF PRO, V38, P56
[6]   Optimal errors and phase transitions in high-dimensional generalized linear models [J].
Barbier, Jean ;
Krzakala, Florent ;
Macris, Nicolas ;
Miolane, Leo ;
Zdeborova, Lenka .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2019, 116 (12) :5451-5460
[7]  
Barik Adarsh, 2022, INT C MACHINE LEARNI, V162, P1627
[8]   The LASSO Risk for Gaussian Matrices [J].
Bayati, Mohsen ;
Montanari, Andrea .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2012, 58 (04) :1997-2017
[9]   The Dynamics of Message Passing on Dense Graphs, with Applications to Compressed Sensing [J].
Bayati, Mohsen ;
Montanari, Andrea .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2011, 57 (02) :764-785
[10]   Phase Retrieval via Wirtinger Flow: Theory and Algorithms [J].
Candes, Emmanuel J. ;
Li, Xiaodong ;
Soltanolkotabi, Mahdi .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2015, 61 (04) :1985-2007