Global and Local Mixture Consistency Cumulative Learning for Long-tailed Visual Recognitions

被引:29
作者
Du, Fei [1 ,2 ,3 ]
Yang, Peng [1 ,3 ]
Jia, Qi [1 ,3 ]
Nan, Fengtao [1 ,2 ,3 ]
Chen, Xiaoting [1 ,3 ]
Yang, Yun [1 ,3 ]
机构
[1] Yunnan Univ, Natl Pilot Sch Software, Kunming, Peoples R China
[2] Yunnan Univ, Sch Informat Sci & Engn, Kunming, Peoples R China
[3] Yunnan Key Lab Software Engn, Kunming, Peoples R China
来源
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2023年
关键词
D O I
10.1109/CVPR52729.2023.01518
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, our goal is to design a simple learning paradigm for long-tail visual recognition, which not only improves the robustness of the feature extractor but also alleviates the bias of the classifier towards head classes while reducing the training skills and overhead. We propose an efficient one-stage training strategy for long-tailed visual recognition called Global and Local Mixture Consistency cumulative learning (GLMC). Our core ideas are twofold: (1) a global and local mixture consistency loss improves the robustness of the feature extractor. Specifically, we generate two augmented batches by the global MixUp and local CutMix from the same batch data, respectively, and then use cosine similarity to minimize the difference. (2) A cumulative head-tail soft label reweighted loss mitigates the head class bias problem. We use empirical class frequencies to reweight the mixed label of the head-tail class for long-tailed data and then balance the conventional loss and the rebalanced loss with a coefficient accumulated by epochs. Our approach achieves state-of-the-art accuracy on CIFAR10-LT, CIFAR100-LT, and ImageNet-LT datasets. Additional experiments on balanced ImageNet and CIFAR demonstrate that GLMC can significantly improve the generalization of backbones. Code is made publicly available at https://github.com/ynu-yangpeng/GLMC.
引用
收藏
页码:15814 / 15823
页数:10
相关论文
共 47 条
[1]   Long-Tailed Recognition via Weight Balancing [J].
Alshammari, Shaden ;
Wang, Yu-Xiong ;
Ramanan, Deva ;
Kong, Shu .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :6887-6897
[2]  
Boski M, 2017, 2017 10TH INTERNATIONAL WORKSHOP ON MULTIDIMENSIONAL (ND) SYSTEMS (NDS)
[3]  
Byrd Jonathon, 2019, INT C MACH LEARN, P872
[4]  
Cao KD, 2019, ADV NEUR IN, V32
[5]  
Cao Y, 2016, AAAI CONF ARTIF INTE, P3457
[6]   SMOTE: Synthetic minority over-sampling technique [J].
Chawla, Nitesh V. ;
Bowyer, Kevin W. ;
Hall, Lawrence O. ;
Kegelmeyer, W. Philip .
2002, American Association for Artificial Intelligence (16)
[7]   Exploring Simple Siamese Representation Learning [J].
Chen, Xinlei ;
He, Kaiming .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :15745-15753
[8]   Parametric Contrastive Learning [J].
Cui, Jiequan ;
Zhong, Zhisheng ;
Liu, Shu ;
Yu, Bei ;
Jia, Jiaya .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :695-704
[9]   Class-Balanced Loss Based on Effective Number of Samples [J].
Cui, Yin ;
Jia, Menglin ;
Lin, Tsung-Yi ;
Song, Yang ;
Belongie, Serge .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :9260-9269
[10]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848