Sample Prior Guided Robust Model Learning to Suppress Noisy Labels

被引:3
作者
Chen, Wenkai [1 ]
Zhu, Chuang [1 ]
Li, Mengting [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Sch Artificial Intelligence, Beijing, Peoples R China
来源
MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT II | 2023年 / 14170卷
关键词
Noisy label; Hard sample; Semi-supervised learning; Pseudo-label;
D O I
10.1007/978-3-031-43415-0_1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Imperfect labels are ubiquitous in real-world datasets and seriously harm the model performance. Several recent effective methods for handling noisy labels have two key steps: 1) dividing samples into cleanly labeled and wrongly labeled sets by training loss, 2) using semi-supervised methods to generate pseudo-labels for samples in the wrongly labeled set. However, current methods always hurt the informative hard samples due to the similar loss distribution between the hard samples and the noisy ones. In this paper, we proposed PGDF (Prior Guided Denoising Framework), a novel framework to learn a deep model to suppress noisy label by using the training history to generate the sample prior knowledge, which is integrated into both sample dividing step and semi-supervised step. Our framework can save more informative hard clean samples into the cleanly labeled set. Besides, our framework also promotes the quality of pseudo-labels during the semi-supervised step by suppressing the noise in the current pseudo-labels generating scheme. To further enhance the hard samples, we reweight the samples in the cleanly labeled set during training. We evaluated our method using synthetic datasets based on CIFAR-10 and CIFAR-100, as well as on the real-world datasets WebVision and Clothing1M. The results demonstrate substantial improvements over state-of-the-art methods. The code is available at https://github.com/bupt-ai-cz/PGDF.
引用
收藏
页码:3 / 19
页数:17
相关论文
共 46 条
  • [1] Arazo E, 2019, PR MACH LEARN RES, V97
  • [2] Me-Momentum: Extracting Hard Confident Examples from Noisily Labeled Data
    Bai, Yingbin
    Liu, Tongliang
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 9292 - 9301
  • [3] Berthelot D, 2019, ADV NEUR IN, V32
  • [4] Chen PF, 2019, Arxiv, DOI arXiv:1905.05040
  • [5] Chen PF, 2021, AAAI CONF ARTIF INTE, V35, P11442
  • [6] Cordeiro FR, 2021, Arxiv, DOI arXiv:2103.04173
  • [7] Ghosh A, 2017, AAAI CONF ARTIF INTE, P1919
  • [8] Goldberger J., 2017, 5 INT C LEARNING REP
  • [9] Co-teaching: Robust Training of Deep Neural Networks with Extremely Noisy Labels
    Han, Bo
    Yao, Quanming
    Yu, Xingrui
    Niu, Gang
    Xu, Miao
    Hu, Weihua
    Tsang, Ivor W.
    Sugiyama, Masashi
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [10] Deep Self-Learning From Noisy Labels
    Han, Jiangfan
    Luo, Ping
    Wang, Xiaogang
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 5137 - 5146