Robust PACm : Training Ensemble Models Under Misspecification and Outliers

被引:2
作者
Zecchin, Matteo [1 ]
Park, Sangwoo [1 ]
Simeone, Osvaldo [1 ]
Kountouris, Marios [2 ]
Gesbert, David [2 ]
机构
[1] Kings Coll London, Dept Engn, Kings Commun Learning & Informat Proc KCLIP Lab, London, England
[2] EURECOM, Commun Syst Dept, F-06410 Sophia Antipolis, France
基金
欧洲研究理事会; 英国工程与自然科学研究理事会;
关键词
Bayesian learning; ensemble models; machine learning; misspecification; outliers; robustness; BAYESIAN-INFERENCE;
D O I
10.1109/TNNLS.2023.3295168
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Standard Bayesian learning is known to have suboptimal generalization capabilities under misspecification and in the presence of outliers. Probably approximately correct (PAC)-Bayes theory demonstrates that the free energy criterion minimized by Bayesian learning is a bound on the generalization error for Gibbs predictors (i.e., for single models drawn at random from the posterior) under the assumption of sampling distributions uncontaminated by outliers. This viewpoint provides a justification for the limitations of Bayesian learning when the model is misspecified, requiring ensembling, and when data are affected by outliers. In recent work, PAC-Bayes bounds-referred to as PAC(m)-were derived to introduce free energy metrics that account for the performance of ensemble predictors, obtaining enhanced performance under misspecification. This work presents a novel robust free energy criterion that combines the generalized logarithm score function with PAC(m) ensemble bounds. The proposed free energy training criterion produces predictive distributions that are able to concurrently counteract the detrimental effects of misspecification-with respect to both likelihood and prior distribution-and outliers.
引用
收藏
页码:1 / 15
页数:15
相关论文
共 50 条
[1]   Asymptotics of Bayesian Inference for a Class of Probabilistic Models under Misspecification [J].
Miya, Nozomi ;
Suko, Tota ;
Yasuda, Goki ;
Matsushima, Toshiyasu .
IEICE TRANSACTIONS ON FUNDAMENTALS OF ELECTRONICS COMMUNICATIONS AND COMPUTER SCIENCES, 2014, E97A (12) :2352-2360
[2]   Learning under Model Misspecification: Applications to Variational and Ensemble methods [J].
Masegosa, Andres R. .
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
[3]   Robust estimation of models for longitudinal data with dropouts and outliers [J].
Zhang, Yuexia ;
Qin, Guoyou ;
Zhu, Zhongyi ;
Fu, Bo .
JOURNAL OF APPLIED STATISTICS, 2022, 49 (04) :902-925
[4]   Bias in dynamic panel models under time series misspecification [J].
Lee, Yoonseok .
JOURNAL OF ECONOMETRICS, 2012, 169 (01) :54-60
[5]   Novel hybrid and weighted ensemble models to predict river discharge series with outliers [J].
Shabbir, Maha ;
Chand, Sohail ;
Iqbal, Farhat .
KUWAIT JOURNAL OF SCIENCE, 2024, 51 (02)
[6]   Robust estimation methods for addressing multicollinearity and outliers in beta regression models [J].
Olalekan T. Olaluwoye ;
Adewale F. Lukman ;
Masad A. Alrasheedi ;
Wycliffe N. Nzomo ;
Rasha A. Farghali .
Scientific Reports, 15 (1)
[7]   Tests for unbalanced error-components models under local misspecification [J].
Sosa-Escudero, Walter ;
Bera, Anil K. .
STATA JOURNAL, 2008, 8 (01) :68-78
[8]   Robust estimation in long-memory processes under additive outliers [J].
Molinares, Fabio Fajardo ;
Reisen, Valderio Anselmo ;
Cribari-Neto, Francisco .
JOURNAL OF STATISTICAL PLANNING AND INFERENCE, 2009, 139 (08) :2511-2525
[9]   Robust inferences in longitudinal models for binary and count panel data in the presence of outliers [J].
Bari W. ;
Sutradhar B.C. .
Sankhya B, 2010, 72 (1) :11-37
[10]   Robust time series clustering of GARCH (1,1) models with outliers [J].
Lestari, Vemmie Nastiti ;
Abdurakhman, Dedi ;
Rosadi, Dedi .
STATISTICS, 2025, 59 (01) :152-166