Transfer Knowledge from Head to Tail: Uncertainty Calibration under Long-tailed Distribution

被引:22
作者
Chen, Jiahao [1 ,2 ]
Su, Bing [1 ,2 ]
机构
[1] Renmin Univ China, Gaoling Sch Artificial Intelligence, Beijing, Peoples R China
[2] Beijing Key Lab Big Data Management & Anal Method, Beijing, Peoples R China
来源
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2023年
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52729.2023.01913
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
How to estimate the uncertainty of a given model is a crucial problem. Current calibration techniques treat different classes equally and thus implicitly assume that the distribution of training data is balanced, but ignore the fact that real-world data often follows a long-tailed distribution. In this paper, we explore the problem of calibrating the model trained from a long-tailed distribution. Due to the difference between the imbalanced training distribution and balanced test distribution, existing calibration methods such as temperature scaling can not generalize well to this problem. Specific calibration methods for domain adaptation are also not applicable because they rely on unlabeled target domain instances which are not available. Models trained from a long-tailed distribution tend to be more overconfident to head classes. To this end, we propose a novel knowledge-transferring-based calibration method by estimating the importance weights for samples of tail classes to realize long-tailed calibration. Our method models the distribution of each class as a Gaussian distribution and views the source statistics of head classes as a prior to calibrate the target distributions of tail classes. We adaptively transfer knowledge from head classes to get the target probability density of tail classes. The importance weight is estimated by the ratio of the target probability density over the source probability density. Extensive experiments on CIFAR-10LT, MNIST-LT, CIFAR-100-LT, and ImageNet-LT datasets demonstrate the effectiveness of our method.
引用
收藏
页码:19978 / 19987
页数:10
相关论文
共 45 条
[1]  
Amodei Dario, 2016, Concrete problems in ai safety
[2]  
[Anonymous], 1961, Proceedings of the fourth Berkeley Symposium on mathematical statistics and probability, DOI DOI 10.1021/JP106846B
[3]  
[Anonymous], 2018, Do cifar-10 classifiers generalize to cifar-107
[4]  
[Anonymous], 2017, INT C MACH LEARN, DOI DOI 10.1109/DSC.2017.89
[5]  
Bickel Steffen, 2006, Advances in Neural Information Processing Systems, V19
[6]  
Cao KD, 2019, ADV NEUR IN, V32
[7]  
Cheng Jiacheng, 2022, P IEEE CVF C COMP VI, P13709
[8]  
CORTES C, 1995, MACH LEARN, V20, P273, DOI 10.1023/A:1022627411411
[9]   Class-Balanced Loss Based on Effective Number of Samples [J].
Cui, Yin ;
Jia, Menglin ;
Lin, Tsung-Yi ;
Song, Yang ;
Belongie, Serge .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :9260-9269
[10]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848