GRADIENT-AWARE LOGIT ADJUSTMENT LOSS FOR LONG-TAILED CLASSIFIER

被引:1
作者
Zhang, Fan [1 ,6 ]
Qin, Wei [2 ]
Ren, Weijieying [3 ]
Wang, Lei [4 ]
Chen, Zetong [5 ]
Hong, Richang [2 ]
机构
[1] Anhui Univ, AHU IAI AI Joint Lab, Hefei, Peoples R China
[2] Hefei Univ Technol, Hefei, Peoples R China
[3] Penn State Univ, University Pk, PA 16802 USA
[4] Singapore Managerment Univ, Singapore, Singapore
[5] Univ Sci & Technol China, Hefei, Peoples R China
[6] Hefei Comprehens Natl Sci Ctr, Inst Artificial Intelligence, Hefei, Peoples R China
来源
2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024 | 2024年
关键词
Long-tailed distribution; imbalanced gradient; post hoc methods;
D O I
10.1109/ICASSP48485.2024.10445992
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
In the real-world setting, data often follows a long-tailed distribution, where head classes contain significantly more training samples than tail classes. Consequently, models trained on such data tend to be biased toward head classes. The medium of this bias is imbalanced gradients, which include not only the ratio of scale between positive and negative gradients but also imbalanced gradients from different negative classes. Therefore, we propose the Gradient-Aware Logit Adjustment (GALA) loss, which adjusts the logits based on accumulated gradients to balance the optimization process. Additionally, We find that most of the solutions to long-tailed problems are still biased towards head classes in the end, and we propose a simple and post hoc prediction re-balancing strategy to further mitigate the basis toward head class. Extensive experiments are conducted on multiple popular longtailed recognition benchmark datasets to evaluate the effectiveness of these two designs. Our approach achieves top1 accuracy of 48.5%, 41.4%, and 73.3% on CIFAR100-LT, Places-LT, and iNaturalist, outperforming the state-of-the-art method GCL by a significant margin of 3.62%, 0.76% and 1.2%, respectively. Code is available at https://github.com/ltproject-repository/lt-project.
引用
收藏
页码:3190 / 3194
页数:5
相关论文
共 21 条
[1]   Long-Tailed Recognition via Weight Balancing [J].
Alshammari, Shaden ;
Wang, Yu-Xiong ;
Ramanan, Deva ;
Kong, Shu .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :6887-6897
[2]  
[Anonymous], 2021, IEEE T MULTIMEDIA, DOI DOI 10.1111/1541-4337.12808
[3]  
Cao K., 2019, Learning Imbalanced Datasets with Label-Distribution-Aware Margin Loss
[4]   Distilling Virtual Examples for Long-tailed Recognition [J].
He, Yin-Yin ;
Wu, Jianxin ;
Wei, Xiu-Shen .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :235-244
[5]  
Hinton G., 2015, ARXIV150302531
[6]  
Iscen Ahmet, 2021, BMVC
[7]  
Kang B., 2020, P INT C LEARN REPR
[8]   Long-tailed Visual Recognition via Gaussian Clouded Logit Adjustment [J].
Li, Mengke ;
Cheung, Yiu-Ming ;
Lu, Yang .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :6919-6928
[9]   Large-Scale Long-Tailed Recognition in an Open World [J].
Liu, Ziwei ;
Miao, Zhongqi ;
Zhan, Xiaohang ;
Wang, Jiayun ;
Gong, Boqing ;
Yu, Stella X. .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :2532-2541
[10]  
Menon Aditya Krishna, 2021, INT C LEARN REPR