Balanced Product of Calibrated Experts for Long-Tailed Recognition

被引:17
作者
Aimar, Emanuel Sanchez [1 ]
Jonnarth, Arvi [1 ,3 ]
Felsberg, Michael [1 ,4 ]
Kuhlmann, Marco [2 ]
机构
[1] Linkoping Univ, Dept Elect Engn, Linkoping, Sweden
[2] Linkoping Univ, Dept Comp & Informat Sci, Linkoping, Sweden
[3] Husqvarna Grp, Huskvarna, Sweden
[4] Univ KwaZulu Natal, Durban, South Africa
来源
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2023年
基金
瑞典研究理事会;
关键词
MIXTURES;
D O I
10.1109/CVPR52729.2023.01912
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Many real-world recognition problems are characterized by long-tailed label distributions. These distributions make representation learning highly challenging due to limited generalization over the tail classes. If the test distribution differs from the training distribution, e.g. uniform versus long-tailed, the problem of the distribution shift needs to be addressed. A recent line of work proposes learning multiple diverse experts to tackle this issue. Ensemble diversity is encouraged by various techniques, e.g. by specializing different experts in the head and the tail classes. In this work, we take an analytical approach and extend the notion of logit adjustment to ensembles to form a Balanced Product of Experts (BalPoE). BalPoE combines a family of experts with different test-time target distributions, generalizing several previous approaches. We show how to properly define these distributions and combine the experts in order to achieve unbiased predictions, by proving that the ensemble is Fisher-consistent for minimizing the balanced error. Our theoretical analysis shows that our balanced ensemble requires calibrated experts, which we achieve in practice using mixup. We conduct extensive experiments and our method obtains new state-of-the-art results on three long-tailed datasets: CIFAR-100-LT, ImageNet-LT, and iNaturalist-2018. Our code is available at https://github.com/emasa/BalPoE-CalibratedLT.
引用
收藏
页码:19967 / 19977
页数:11
相关论文
共 72 条
  • [1] Applying support vector machines to imbalanced datasets
    Akbani, R
    Kwek, S
    Japkowicz, N
    [J]. MACHINE LEARNING: ECML 2004, PROCEEDINGS, 2004, 3201 : 39 - 50
  • [2] [Anonymous], 2017, INT C MACH LEARN, DOI DOI 10.1109/DSC.2017.89
  • [3] Brodersen Kay H., 2010, Proceedings of the 2010 20th International Conference on Pattern Recognition (ICPR 2010), P3121, DOI 10.1109/ICPR.2010.764
  • [4] Reliability, sufficiency, and the decomposition of proper scores
    Broecker, Jochen
    [J]. QUARTERLY JOURNAL OF THE ROYAL METEOROLOGICAL SOCIETY, 2009, 135 (643) : 1512 - 1519
  • [5] A systematic study of the class imbalance problem in convolutional neural networks
    Buda, Mateusz
    Maki, Atsuto
    Mazurowski, Maciej A.
    [J]. NEURAL NETWORKS, 2018, 106 : 249 - 259
  • [6] ACE: Ally Complementary Experts for Solving Long-Tailed Recognition in One-Shot
    Cai, Jiarui
    Wang, Yizhou
    Hwang, Jenq-Neng
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 112 - 121
  • [7] Cao KD, 2019, ADV NEUR IN, V32
  • [8] Chan P. K., 1998, Proceedings Fourth International Conference on Knowledge Discovery and Data Mining, P164
  • [9] Chan Philip K, 1998, WORKSH NOT KDD 98 WO
  • [10] SMOTE: Synthetic minority over-sampling technique
    Chawla, Nitesh V.
    Bowyer, Kevin W.
    Hall, Lawrence O.
    Kegelmeyer, W. Philip
    [J]. 2002, American Association for Artificial Intelligence (16)