Learning from Noisy Labels with Distillation

被引:355
作者
Li, Yuncheng [1 ]
Yang, Jianchao [1 ]
Song, Yale [2 ]
Cao, Liangliang [3 ]
Luo, Jiebo [4 ]
Li, Li-Jia [5 ]
机构
[1] Snap Inc, Venice, CA 90291 USA
[2] Yahoo Res, New York, NY USA
[3] Hellovera Ai, New York, NY USA
[4] Univ Rochester, Rochester, NY 14627 USA
[5] Google Inc, Mountain View, CA USA
来源
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV) | 2017年
关键词
D O I
10.1109/ICCV.2017.211
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The ability of learning from noisy labels is very useful in many visual recognition tasks, as a vast amount of data with noisy labels are relatively easy to obtain. Traditionally, label noise has been treated as statistical outliers, and techniques such as importance re-weighting and bootstrapping have been proposed to alleviate the problem. According to our observation, the real-world noisy labels exhibit multi-mode characteristics as the true labels, rather than behaving like independent random outliers. In this work, we propose a unified distillation framework to use "side" information, including a small clean dataset and label relations in knowledge graph, to "hedge the risk" of learning from noisy labels. Unlike the traditional approaches evaluated based on simulated label noises, we propose a suite of new benchmark datasets, in Sports, Species and Artifacts domains, to evaluate the task of learning from noisy labels in the practical setting. The empirical study demonstrates the effectiveness of our proposed method in all the domains.
引用
收藏
页码:1928 / 1936
页数:9
相关论文
共 24 条
[1]  
[Anonymous], 1998, STAT LEARNING THEORY
[2]  
[Anonymous], ICML
[3]  
[Anonymous], I SEMANTICS
[4]  
[Anonymous], 2010, Advances in neural information processing systems
[5]  
[Anonymous], 2015, COMPUTER SCI
[6]  
[Anonymous], 2015, ICCV
[7]  
[Anonymous], 2000, PROC INT C MACH LEAR
[8]  
[Anonymous], 2015, ARXIV151103643
[9]  
[Anonymous], 2015, ARXIV151106789
[10]  
[Anonymous], 2014, arXiv preprint arXiv:1406.2080 2