Ensemble of Averages: Improving Model Selection and Boosting Performance in Domain Generalization

被引:0
作者
Arpit, Devansh [1 ]
Wang, Huan [1 ]
Zhou, Yingbo [1 ]
Xiong, Caiming [1 ]
机构
[1] Salesforce Res, Palo Alto, CA 94301 USA
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022 | 2022年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In Domain Generalization (DG) settings, models trained independently on a given set of training domains have notoriously chaotic performance on distribution shifted test domains, and stochasticity in optimization (e.g. seed) plays a big role. This makes deep learning models unreliable in real world settings. We first show that this chaotic behavior exists even along the training optimization trajectory of a single model, and propose a simple model averaging protocol that both significantly boosts domain generalization and diminishes the impact of stochasticity by improving the rank correlation between the in-domain validation accuracy and out-domain test accuracy, which is crucial for reliable early stopping. Taking advantage of our observation, we show that instead of ensembling unaveraged models (that is typical in practice), ensembling moving average models (EoA) from independent runs further boosts performance. We theoretically explain the boost in performance of ensembling and model averaging by adapting the well known Bias-Variance trade-off to the domain generalization setting. On the DomainBed benchmark, when using a pre-trained ResNet-50, this ensemble of averages achieves an average of 68.0%, beating vanilla ERM (w/o averaging/ensembling) by similar to 4%, and when using a pre-trained RegNetY-16GF, achieves an average of 76.6%, beating vanilla ERM by 6%. Our code is available at https://github.com/salesforce/ensemble-of- averages.
引用
收藏
页数:13
相关论文
共 52 条
[1]  
Albuquerque I., 2020, ARXIV200313525
[2]  
[Anonymous], 2019, Billion-scale semisupervised learning for image classification
[3]  
[Anonymous], 1998, N Y
[4]  
Arjovsky Martin, 2019, arXiv preprint arXiv:1907.02893
[5]   Recognition in Terra Incognita [J].
Beery, Sara ;
Van Horn, Grant ;
Perona, Pietro .
COMPUTER VISION - ECCV 2018, PT XVI, 2018, 11220 :472-489
[6]  
Ben-David S., 2007, Advances in Neural Information Processing Systems, P137, DOI DOI 10.7551/MITPRESS/7503.003.0022
[7]  
Beutel Alex, 2020, ARXIV201103395, P1
[8]  
Blanchard Gilles, 2011, P 24 INT C NEUR INF
[9]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[10]  
Bucci Silvia, 2021, IEEE Transactions on Pattern Analysis and Machine Intelligence