A Progressive Deep Neural Network Training Method for Image Classification with Noisy Labels

被引:0
作者
Yan, Xuguo [1 ,2 ,3 ]
Xia, Xuhui [1 ,2 ,3 ]
Wang, Lei [1 ,2 ,3 ]
Zhang, Zelin [1 ,2 ,3 ]
机构
[1] Wuhan Univ Sci & Technol, Key Lab Met Equipment & Control Technol, Minist Educ, Wuhan 430081, Peoples R China
[2] Wuhan Univ Sci & Technol, Hubei Key Lab Mech Transmiss & Mfg Engn, Wuhan 430081, Peoples R China
[3] Wuhan Univ Sci & Technol, Precis Mfg Inst, Wuhan 430081, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2022年 / 12卷 / 24期
基金
中国国家自然科学基金;
关键词
DNNs; curriculum learning; progressive learning; noisy labels; image classification;
D O I
10.3390/app122412754
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Deep neural networks (DNNs) require large amounts of labeled data for model training. However, label noise is a common problem in datasets due to the difficulty of classification and high cost of labeling processes. Introducing the concepts of curriculum learning and progressive learning, this paper presents a novel solution that is able to handle massive noisy labels and improve model generalization ability. It proposes a new network model training strategy that considers mislabeled samples directly in the network training process. The new learning curriculum is designed to measures the complexity of the data with their distribution density in a feature space. The sample data in each category are then divided into easy-to-classify (clean samples), relatively easy-to-classify, and hard-to-classify (noisy samples) subsets according to the smallest intra-class local density with each cluster. On this basis, DNNs are trained progressively in three stages, from easy to hard, i.e., from clean to noisy samples. The experimental results demonstrate that the accuracy of image classification can be improved through data augmentation, and the classification accuracy of the proposed method is clearly higher than that of standard Inception_v2 for the NEU dataset after data augmentation, when the proportion of noisy labels in the training set does not exceed 60%. With 50% noisy labels in the training set, the classification accuracy of the proposed method outperformed recent state-of-the-art label noise learning methods, CleanNet and MentorNet. The proposed method also performed well in practical applications, where the number of noisy labels was uncertain and unevenly distributed. In this case, the proposed method not only can alleviate the adverse effects of noisy labels, but it can also improve the generalization ability of standard deep networks and their overall capability.
引用
收藏
页数:18
相关论文
共 31 条
  • [1] Ensemble diversified learning for image classification with noisy labels
    Ahmed, Ahmed
    Yousif, Hayder
    He, Zhihai
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (14) : 20759 - 20772
  • [2] Image classification with deep learning in the presence of noisy labels: A survey
    Algan, Gorkem
    Ulusoy, Ilkay
    [J]. KNOWLEDGE-BASED SYSTEMS, 2021, 215
  • [3] Towards Explainable Deep Neural Networks for the Automatic Detection of Diabetic Retinopathy
    Alghamdi, Hanan Saleh
    [J]. APPLIED SCIENCES-BASEL, 2022, 12 (19):
  • [4] Dani S., 2021, J EMERG TECHNOL INNO, V8, P609
  • [5] Label Noise Cleaning with an Adaptive Ensemble Method Based on Noise Detection Metric
    Feng, Wei
    Quan, Yinghui
    Dauphin, Gabriel
    [J]. SENSORS, 2020, 20 (23) : 1 - 16
  • [6] Ghahreman N., 2011, COMPUT INF SCI, V4, P48
  • [7] Ghosh A, 2017, AAAI CONF ARTIF INTE, P1919
  • [8] Grandvalet Y., 2005, CAP
  • [9] Application and Research of Distance and Density on Improved K-means
    Gu, Hongbo
    [J]. 2018 INTERNATIONAL CONFERENCE ON COMPUTER INFORMATION SCIENCE AND APPLICATION TECHNOLOGY, 2019, 1168
  • [10] Robust Semi-Supervised Classification for Noisy Labels Based on Self-Paced Learning
    Gu, Nannan
    Fan, Mingyu
    Meng, Deyu
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2016, 23 (12) : 1806 - 1810