Robust sparse regression by modeling noise as a mixture of gaussians

被引:4
|
作者
Xu, Shuang [1 ]
Zhang, Chun-Xia [1 ]
机构
[1] Xi An Jiao Tong Univ, Sch Math & Stat, Xian 710049, Shaanxi, Peoples R China
基金
中国国家自然科学基金;
关键词
Robust regression; penalized regression; variable selection; mixture of Gaussians; lasso; VARIABLE SELECTION; REGULARIZATION; SHRINKAGE; ALGORITHM;
D O I
10.1080/02664763.2019.1566448
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Regression analysis has been proven to be a quite effective tool in a large variety of fields. In many regression models, it is often assumed that noise is with a specific distribution. Although the theoretical analysis can be greatly facilitated, the model-fitting performance may be poor since the supposed noise distribution may deviate from real noise to a large extent. Meanwhile, the model is also expected to be robust in consideration of the complexity of real-world data. Without any assumption about noise, we propose in this paper a novel sparse regression method called MoG-Lasso to directly model noise in linear regression models via a mixture of Gaussian distributions (MoG). Meanwhile, the penalty is included as a part of the loss function of MoG-Lasso to enhance its ability to identify a sparse model. As for the parameters in MoG-Lasso, we present an efficient algorithm to estimate them via the EM (expectation maximization) and ADMM (alternating direction method of multipliers) algorithms. With some simulated and real data contaminated by complex noise, the experiments show that the novel model MoG-Lasso performs better than several other popular methods in both 'p>n' and 'p<n' situations, including Lasso, LAD-Lasso and Huber-Lasso.
引用
收藏
页码:1738 / 1755
页数:18
相关论文
共 50 条
  • [21] Linearly-Involved Moreau-Enhanced-Over-Subspace Model: Debiased Sparse Modeling and Stable Outlier-Robust Regression
    Yukawa, Masahiro
    Kaneko, Hiroyuki
    Suzuki, Kyohei
    Yamada, Isao
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2023, 71 : 1232 - 1247
  • [22] Robust subspace clustering via penalized mixture of Gaussians
    Yao, Jing
    Cao, Xiangyong
    Zhao, Qian
    Meng, Deyu
    Xu, Zongben
    NEUROCOMPUTING, 2018, 278 : 4 - 11
  • [23] Robust and Sparse Regression via γ-Divergence
    Kawashima, Takayuki
    Fujisawa, Hironori
    ENTROPY, 2017, 19 (11):
  • [24] Ultrahigh-Dimensional Robust and Efficient Sparse Regression Using Non-Concave Penalized Density Power Divergence
    Ghosh, Abhik
    Majumdar, Subhabrata
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2020, 66 (12) : 7812 - 7827
  • [25] Robust regression with compositional covariates
    Mishra, Aditya
    Muller, Christian L.
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2022, 165
  • [26] Robust sparse regression and tuning parameter selection via the efficient bootstrap information criteria
    Park, Heewon
    Sakaori, Fumitake
    Konishi, Sadanori
    JOURNAL OF STATISTICAL COMPUTATION AND SIMULATION, 2014, 84 (07) : 1596 - 1607
  • [27] Correlated variables in regression: Clustering and sparse estimation
    Buehlmann, Peter
    Ruetimann, Philipp
    van de Geer, Sara
    Zhang, Cun-Hui
    JOURNAL OF STATISTICAL PLANNING AND INFERENCE, 2013, 143 (11) : 1835 - 1858
  • [28] ON GROUPING EFFECT OF SPARSE STABLE OUTLIER-ROBUST REGRESSION
    Suzuki, Kyohei
    Yukawa, Masahiro
    2022 IEEE 32ND INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2022,
  • [29] Sparse regression for large data sets with outliers
    Bottmer, Lea
    Croux, Christophe
    Wilms, Ines
    EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2022, 297 (02) : 782 - 794
  • [30] Sparse estimation in semiparametric finite mixture of varying coefficient regression models
    Khalili, Abbas
    Shokoohi, Farhad
    Asgharian, Masoud
    Lin, Shili
    BIOMETRICS, 2023, 79 (04) : 3445 - 3457