Reviving undersampling for long-tailed learning

被引:0
作者
Yu, Hao
Du, Yingxiao
Wu, Jianxin [1 ]
机构
[1] Nanjing Univ, Natl Key Lab Novel Software Technol, Nanjing, Peoples R China
基金
中国国家自然科学基金;
关键词
Image classification; Long-tailed learning; Undersampling;
D O I
10.1016/j.patcog.2024.111200
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The training datasets used in long-tailed recognition are extremely unbalanced, resulting insignificant variation in per-class accuracy across categories. Prior works mostly used average accuracy to evaluate their algorithms, which easily ignores those worst-performing categories. In this paper, we aim to enhance the accuracy of the worst-performing categories and utilize the harmonic mean and geometric mean to assess the model's performance. We revive the balanced undersampling idea to achieve this goal. In few-shot learning, balanced subsets are few-shot and will surely under-fit, hence it is not used in modern long-tailed learning. But, we find that it produces a more equitable distribution of accuracy across categories with much higher harmonic and geometric mean accuracy, but with lower average accuracy. Moreover, we devise a straightforward model ensemble strategy, which does not result in any additional overhead and achieves improved harmonic and geometric mean while keeping the average accuracy almost intact when compared to state-of-the-art longtailed learning methods. We validate the effectiveness of our approach on widely utilized benchmark datasets for long-tailed learning. Our code is at https://github.com/yuhao318/BTM/.
引用
收藏
页数:9
相关论文
共 37 条
  • [1] DBN-Mix: Training dual branch network using bilateral mixup augmentation for long-tailed visual recognition
    Baik, Jae Soon
    Yoon, In Young
    Choi, Jun Won
    [J]. PATTERN RECOGNITION, 2024, 147
  • [2] SMOTE: Synthetic minority over-sampling technique
    Chawla, Nitesh V.
    Bowyer, Kevin W.
    Hall, Lawrence O.
    Kegelmeyer, W. Philip
    [J]. 2002, American Association for Artificial Intelligence (16)
  • [3] Transfer Knowledge from Head to Tail: Uncertainty Calibration under Long-tailed Distribution
    Chen, Jiahao
    Su, Bing
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 19978 - 19987
  • [4] Generalized Parametric Contrastive Learning
    Cui, Jiequan
    Zhong, Zhisheng
    Tian, Zhuotao
    Liu, Shu
    Yu, Bei
    Jia, Jiaya
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (12) : 7463 - 7474
  • [5] Parametric Contrastive Learning
    Cui, Jiequan
    Zhong, Zhisheng
    Liu, Shu
    Yu, Bei
    Jia, Jiaya
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 695 - 704
  • [6] Large Scale Fine-Grained Categorization and Domain-Specific Transfer Learning
    Cui, Yin
    Song, Yang
    Sun, Chen
    Howard, Andrew
    Belongie, Serge
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 4109 - 4118
  • [7] Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
  • [8] Dong B., 2023, INT C LEARN REPR
  • [9] No One Left Behind: Improving the Worst Categories in Long-Tailed Learning
    Du, Yingxiao
    Wu, Jianxin
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 15804 - 15813
  • [10] Gao Junqi, 2023, Advances in Neural Information Processing Systems