Learning With Noisy Labels via Self-Reweighting From Class Centroids

被引:31
作者
Ma, Fan [1 ,2 ]
Wu, Yu [1 ,2 ]
Yu, Xin [1 ,2 ]
Yang, Yi [1 ,2 ]
机构
[1] Univ Technol Sydney, ReLER Lab, Sydney, NSW 2007, Australia
[2] Univ Technol Sydney, Australian Artificial Intelligence Inst, Sydney, NSW 2007, Australia
基金
澳大利亚研究理事会;
关键词
Noise measurement; Training; Data models; Optimization; Task analysis; Neural networks; Interpolation; Centroids; noisy labels; self-reweighting;
D O I
10.1109/TNNLS.2021.3073248
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Although deep neural networks have been proved effective in many applications, they are data hungry, and training deep models often requires laboriously labeled data. However, when labeled data contain erroneous labels, they often lead to model performance degradation. A common solution is to assign each sample with a dynamic weight during optimization, and the weight is adjusted in accordance with the loss. However, those weights are usually unreliable since they are measured by the losses of corrupted labels. Thus, this scheme might impede the discriminative ability of neural networks trained on noisy data. To address this issue, we propose a novel reweighting method, dubbed self-reweighting from class centroids (SRCC), by assigning sample weights based on the similarities between the samples and our online learned class centroids. Since we exploit statistical class centers in the image feature space to reweight data samples in learning, our method is robust to noise caused by corrupted labels. In addition, even after reweighting the noisy data, the decision boundaries might still suffer distortions. Thus, we leverage mixed inputs that are generated by linearly interpolating two random images and their labels to further regularize the boundaries. We employ the learned class centroids to evaluate the confidence of our generated mixed data via measuring feature similarities. During the network optimization, the class centroids are updated as more discriminative feature representations of original images are learned. In doing so, SRCC will generate more robust weighting coefficients for noisy and mixed data and facilitates our feature representation learning in return. Extensive experiments on both the synthetic and real image recognition tasks demonstrate that our method SRCC outperforms the state of the art on learning with noisy data.
引用
收藏
页码:6275 / 6285
页数:11
相关论文
共 52 条
[21]  
Kumar M.P., 2010, ADV NEURAL INFORM PR
[22]  
Li DX, 2020, IEEE WINT CONF APPL, P1448, DOI [10.1109/wacv45572.2020.9093512, 10.1109/WACV45572.2020.9093512]
[23]  
Liang J., 2016, IJCAI, P1746
[24]   Classification with Noisy Labels by Importance Reweighting [J].
Liu, Tongliang ;
Tao, Dacheng .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2016, 38 (03) :447-461
[25]  
Lyu Y., P INT C LEARN REPR, P1
[26]  
Ma F., 2020, J. Mach. Learn. Res., V21, P1
[27]  
Ma F, 2017, PR MACH LEARN RES, V70
[28]   SF-Net: Single-Frame Supervision for Temporal Action Localization [J].
Ma, Fan ;
Zhu, Linchao ;
Yang, Yi ;
Zha, Shengxin ;
Kundu, Gourab ;
Feiszli, Matt ;
Shou, Zheng .
COMPUTER VISION - ECCV 2020, PT IV, 2020, 12349 :420-437
[29]   Noise Tolerance Under Risk Minimization [J].
Manwani, Naresh ;
Sastry, P. S. .
IEEE TRANSACTIONS ON CYBERNETICS, 2013, 43 (03) :1146-1151
[30]   A theoretical understanding of self-paced learning [J].
Meng, Deyu ;
Zhao, Qian ;
Jiang, Lu .
INFORMATION SCIENCES, 2017, 414 :319-328