Separating Noisy Samples From Tail Classes for Long-Tailed Image Classification With Label Noise

被引:14
作者
Fang, Chaowei [1 ]
Cheng, Lechao [2 ]
Mao, Yining [2 ]
Zhang, Dingwen [3 ]
Fang, Yixiang [4 ]
Li, Guanbin [5 ]
Qi, Huiyan [6 ]
Jiao, Licheng [1 ]
机构
[1] Xidian Univ, Sch Artificial Intelligence, Key Lab Intelligent Percept & Image Understanding, Minist Educ, Xian 710071, Peoples R China
[2] Zhejiang Lab, Hangzhou 310012, Peoples R China
[3] Northwestern Polytech Univ, Sch Automat, Brain & Artificial Intelligence Lab, Xian 710072, Peoples R China
[4] Chinese Univ Hong Kong, Sch Data Sci, Shenzhen 518172, Peoples R China
[5] Sun Yat Sen Univ, Sch Comp Sci & Engn, Res Inst, Guangzhou 510275, Peoples R China
[6] Fudan Univ, Sch Comp Sci, Shanghai 200433, Peoples R China
基金
中国国家自然科学基金;
关键词
Noise measurement; Tail; Training; Image classification; Task analysis; Biological system modeling; Predictive models; long-tailed; noisy labels; DEEP;
D O I
10.1109/TNNLS.2023.3291695
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most existing methods that cope with noisy labels usually assume that the classwise data distributions are well balanced. They are difficult to deal with the practical scenarios where training samples have imbalanced distributions, since they are not able to differentiate noisy samples from tail classes' clean samples. This article makes an early effort to tackle the image classification task in which the provided labels are noisy and have a long-tailed distribution. To deal with this problem, we propose a new learning paradigm which can screen out noisy samples by matching between inferences on weak and strong data augmentations. A leave-noise-out regularization (LNOR) is further introduced to eliminate the effect of the recognized noisy samples. Besides, we propose a prediction penalty based on the online classwise confidence levels to avoid the bias toward easy classes which are dominated by head classes. Extensive experiments on five datasets including CIFAR-10, CIFAR-100, MNIST, FashionMNIST, and Clothing1M demonstrate that the proposed method outperforms the existing algorithms for learning with long-tailed distribution and label noise.
引用
收藏
页码:16036 / 16048
页数:13
相关论文
共 59 条
[1]  
Arazo E, 2019, PR MACH LEARN RES, V97
[2]  
Arpit D, 2017, PR MACH LEARN RES, V70
[3]  
Berthelot D, 2019, ADV NEUR IN, V32
[4]   A systematic study of the class imbalance problem in convolutional neural networks [J].
Buda, Mateusz ;
Maki, Atsuto ;
Mazurowski, Maciej A. .
NEURAL NETWORKS, 2018, 106 :249-259
[5]  
Cao K., 2019, ARXIV
[6]   Domain-Specific Batch Normalization for Unsupervised Domain Adaptation [J].
Chang, Woong-Gi ;
You, Tackgeun ;
Seo, Seonguk ;
Kwak, Suha ;
Han, Bohyung .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :7346-7354
[7]   SMOTE: Synthetic minority over-sampling technique [J].
Chawla, Nitesh V. ;
Bowyer, Kevin W. ;
Hall, Lawrence O. ;
Kegelmeyer, W. Philip .
2002, American Association for Artificial Intelligence (16)
[8]   AutoAugment: Learning Augmentation Strategies from Data [J].
Cubuk, Ekin D. ;
Zoph, Barret ;
Mane, Dandelion ;
Vasudevan, Vijay ;
Le, Quoc V. .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :113-123
[9]   Large Scale Fine-Grained Categorization and Domain-Specific Transfer Learning [J].
Cui, Yin ;
Song, Yang ;
Sun, Chen ;
Howard, Andrew ;
Belongie, Serge .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :4109-4118
[10]   DeepSMOTE: Fusing Deep Learning and SMOTE for Imbalanced Data [J].
Dablain, Damien ;
Krawczyk, Bartosz ;
Chawla, Nitesh, V .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (09) :6390-6404