Robust PACm : Training Ensemble Models Under Misspecification and Outliers

被引:2
|
作者
Zecchin, Matteo [1 ]
Park, Sangwoo [1 ]
Simeone, Osvaldo [1 ]
Kountouris, Marios [2 ]
Gesbert, David [2 ]
机构
[1] Kings Coll London, Dept Engn, Kings Commun Learning & Informat Proc KCLIP Lab, London, England
[2] EURECOM, Commun Syst Dept, F-06410 Sophia Antipolis, France
基金
欧洲研究理事会; 英国工程与自然科学研究理事会;
关键词
Bayesian learning; ensemble models; machine learning; misspecification; outliers; robustness; BAYESIAN-INFERENCE;
D O I
10.1109/TNNLS.2023.3295168
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Standard Bayesian learning is known to have suboptimal generalization capabilities under misspecification and in the presence of outliers. Probably approximately correct (PAC)-Bayes theory demonstrates that the free energy criterion minimized by Bayesian learning is a bound on the generalization error for Gibbs predictors (i.e., for single models drawn at random from the posterior) under the assumption of sampling distributions uncontaminated by outliers. This viewpoint provides a justification for the limitations of Bayesian learning when the model is misspecified, requiring ensembling, and when data are affected by outliers. In recent work, PAC-Bayes bounds-referred to as PAC(m)-were derived to introduce free energy metrics that account for the performance of ensemble predictors, obtaining enhanced performance under misspecification. This work presents a novel robust free energy criterion that combines the generalized logarithm score function with PAC(m) ensemble bounds. The proposed free energy training criterion produces predictive distributions that are able to concurrently counteract the detrimental effects of misspecification-with respect to both likelihood and prior distribution-and outliers.
引用
收藏
页码:1 / 15
页数:15
相关论文
共 50 条
  • [1] Asymptotics of Bayesian Inference for a Class of Probabilistic Models under Misspecification
    Miya, Nozomi
    Suko, Tota
    Yasuda, Goki
    Matsushima, Toshiyasu
    IEICE TRANSACTIONS ON FUNDAMENTALS OF ELECTRONICS COMMUNICATIONS AND COMPUTER SCIENCES, 2014, E97A (12) : 2352 - 2360
  • [2] Learning under Model Misspecification: Applications to Variational and Ensemble methods
    Masegosa, Andres R.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [3] Robust estimation of models for longitudinal data with dropouts and outliers
    Zhang, Yuexia
    Qin, Guoyou
    Zhu, Zhongyi
    Fu, Bo
    JOURNAL OF APPLIED STATISTICS, 2022, 49 (04) : 902 - 925
  • [4] Bias in dynamic panel models under time series misspecification
    Lee, Yoonseok
    JOURNAL OF ECONOMETRICS, 2012, 169 (01) : 54 - 60
  • [5] Novel hybrid and weighted ensemble models to predict river discharge series with outliers
    Shabbir, Maha
    Chand, Sohail
    Iqbal, Farhat
    KUWAIT JOURNAL OF SCIENCE, 2024, 51 (02)
  • [6] Robust estimation methods for addressing multicollinearity and outliers in beta regression models
    Olalekan T. Olaluwoye
    Adewale F. Lukman
    Masad A. Alrasheedi
    Wycliffe N. Nzomo
    Rasha A. Farghali
    Scientific Reports, 15 (1)
  • [7] Tests for unbalanced error-components models under local misspecification
    Sosa-Escudero, Walter
    Bera, Anil K.
    STATA JOURNAL, 2008, 8 (01) : 68 - 78
  • [8] Robust estimation in long-memory processes under additive outliers
    Molinares, Fabio Fajardo
    Reisen, Valderio Anselmo
    Cribari-Neto, Francisco
    JOURNAL OF STATISTICAL PLANNING AND INFERENCE, 2009, 139 (08) : 2511 - 2525
  • [9] Robust time series clustering of GARCH (1,1) models with outliers
    Lestari, Vemmie Nastiti
    Abdurakhman, Dedi
    Rosadi, Dedi
    STATISTICS, 2025, 59 (01) : 152 - 166
  • [10] Robust quasi-likelihood inference in generalized linear mixed models with outliers
    Sutradhar, Brajendra C.
    Bari, Wasimul
    JOURNAL OF STATISTICAL COMPUTATION AND SIMULATION, 2011, 81 (02) : 233 - 258