Self-paced resistance learning against overfitting on noisy labels

被引:23
作者
Shi, Xiaoshuan [1 ]
Guo, Zhenhua [2 ]
Li, Kang [3 ,4 ,5 ]
Liang, Yun [6 ]
Zhu, Xiaofeng [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Comp Sci & Technol, Chengdu, Sichuan, Peoples R China
[2] Smart Transportat, Suzhou, Peoples R China
[3] Sichuan Univ, West China Hosp, West China Biomed Big Data Ctr, Chengdu, Sichuan, Peoples R China
[4] Sichuan Univ, MedX Ctr Informat, Chengdu, Sichuan, Peoples R China
[5] Shanghai Artificial Intelligence Lab, Shanghai, Peoples R China
[6] Univ Florida, J Crayton Pruitt Family Dept Biomed Engn, Gainesville, FL 32611 USA
关键词
Convolutional neural networks; Self-paced resistance; Model overfitting; Noisy labels; CLASSIFICATION;
D O I
10.1016/j.patcog.2022.109080
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Noisy labels composed of correct and corrupted ones are pervasive in practice. They might significantly deteriorate the performance of convolutional neural networks (CNNs), because CNNs are easily overfit-ted on corrupted labels. To address this issue, inspired by an observation, deep neural networks might first memorize the probably correct-label data and then corrupt-label samples, we propose a novel yet simple self-paced resistance framework to resist corrupted labels, without using any clean valida-tion data. The proposed framework first utilizes the memorization effect of CNNs to learn a curricu-lum, which contains confident samples and provides meaningful supervision for other training samples. Then it adopts selected confident samples and a proposed resistance loss to update model parameters; the resistance loss tends to smooth model parameters' update or attain equivalent prediction over each class, thereby resisting model overfitting on corrupted labels. Finally, we unify these two modules into a single loss function and optimize it in an alternative learning. Extensive experiments demonstrate the significantly superior performance of the proposed framework over recent state-of-the-art methods on noisy-label data. Source codes of the proposed method are available on https://github.com/xsshi2015/ Self- paced- Resistance-Learning .(c) 2022 Elsevier Ltd. All rights reserved.
引用
收藏
页数:12
相关论文
共 68 条
[1]  
Arazo E, 2019, PR MACH LEARN RES, V97
[2]  
Arpit D, 2017, PR MACH LEARN RES, V70
[3]  
Bengio Y, 2009, P 26 ANN INT C MACH, P41, DOI [10.1145/1553374.1553380, DOI 10.1145/1553374.1553380]
[4]  
Berthelot D, 2019, ADV NEUR IN, V32
[5]  
Bossard L, 2014, LECT NOTES COMPUT SC, V8694, P446, DOI 10.1007/978-3-319-10599-4_29
[6]  
Chapelle O., 2009, SemiSupervised Learning, V20, P542
[7]  
Feng YF, 2019, AAAI CONF ARTIF INTE, P3558
[8]   GVCNN: Group-View Convolutional Neural Networks for 3D Shape Recognition [J].
Feng, Yifan ;
Zhang, Zizhao ;
Zhao, Xibin ;
Ji, Rongrong ;
Gao, Yue .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :264-272
[9]   Classification in the Presence of Label Noise: a Survey [J].
Frenay, Benoit ;
Verleysen, Michel .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2014, 25 (05) :845-869
[10]  
Ghosh A, 2017, AAAI CONF ARTIF INTE, P1919