Symmetric Cross Entropy for Robust Learning with Noisy Labels

被引:743
作者
Wang, Yisen [1 ]
Ma, Xingjun [2 ]
Chen, Zaiyi [3 ]
Luo, Yuan [1 ]
Yi, Jinfeng [4 ]
Bailey, James [2 ]
机构
[1] Shanghai Jiao Tong Univ, Shanghai, Peoples R China
[2] Univ Melbourne, Melbourne, Vic, Australia
[3] Cainiao AI, Hangzhou, Peoples R China
[4] JD AI, Nanjing, Peoples R China
来源
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019) | 2019年
关键词
D O I
10.1109/ICCV.2019.00041
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Training accurate deep neural networks (DNNs) in the presence of noisy labels is an important and challenging task. Though a number of approaches have been proposed for learning with noisy labels, many open issues remain. In this paper, we show that DNN learning with Cross Entropy (CE) exhibits overfitting to noisy labels on some classes ("easy" classes), but more surprisingly, it also suffers from significant under learning on some other classes ("hard" classes). Intuitively, CE requires an extra term to facilitate learning of hard classes, and more importantly, this term should be noise tolerant, so as to avoid overfitting to noisy labels. Inspired by the symmetric KL-divergence, we propose the approach of Symmetric cross entropy Learning (SL), boosting CE symmetrically with a noise robust counterpart Reverse Cross Entropy (RCE). Our proposed SL approach simultaneously addresses both the under learning and overfitting problem of CE in the presence of noisy labels. We provide a theoretical analysis of SL and also empirically show, on a range of benchmark and real-world datasets, that SL outperforms state-of-the-art methods. We also show that SL can be easily incorporated into existing methods in order to further enhance their performance.
引用
收藏
页码:322 / 330
页数:9
相关论文
共 29 条
[11]   Learning from Noisy Labels with Distillation [J].
Li, Yuncheng ;
Yang, Jianchao ;
Song, Yale ;
Cao, Liangliang ;
Luo, Jiebo ;
Li, Li-Jia .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :1928-1936
[12]  
Ma Xingjun, 2018, P INT C MACH LEARN
[13]   Femtosecond-Laser Direct Writing of Metallic Micro/Nanostructures: From Fabrication Strategies to Future Applications [J].
Ma, Zhuo-Chen ;
Zhang, Yong-Lai ;
Han, Bing ;
Chen, Qi-Dai ;
Sun, Hong-Bo .
SMALL METHODS, 2018, 2 (07) :1-20
[14]  
Malach E., 2017, NEURIPS
[15]   Making Deep Neural Networks Robust to Label Noise: a Loss Correction Approach [J].
Patrini, Giorgio ;
Rozza, Alessandro ;
Menon, Aditya Krishna ;
Nock, Richard ;
Qu, Lizhen .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :2233-2241
[16]  
Pereyra G., 2019, P INT C LEARN REPR I
[17]  
Reed Scott, 2014, Training deep neural networks on noisy labels with bootstrapping
[18]  
Shwartz-Ziv R., 2017, ARXIV170300810
[19]  
Sukhbaatar Sainbayar, 2014, Training convolutional networks with noisy labels
[20]   Rethinking the Inception Architecture for Computer Vision [J].
Szegedy, Christian ;
Vanhoucke, Vincent ;
Ioffe, Sergey ;
Shlens, Jon ;
Wojna, Zbigniew .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :2818-2826