Adaptive Weighted Losses With Distribution Approximation for Efficient Consistency-Based Semi-Supervised Learning

被引:16
作者
Li, Di [1 ]
Liu, Yang [2 ]
Song, Liang [2 ]
机构
[1] Henan Univ Sci & Technol, Coll Informat Engn, Luoyang 471000, Peoples R China
[2] Fudan Univ, Acad Engn & Technol, Shanghai 200433, Peoples R China
关键词
Semi-supervised learning; consistency; adaptive weighted losses; distribution approximation; RECOGNITION;
D O I
10.1109/TCSVT.2022.3186041
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Recent semi-supervised learning (SSL) algorithms such as FixMatch achieve state-of-the-art performance by exploiting consistency regularization and entropy minimization techniques. However, many consistency-based SSL algorithms extract pseudo-labels from unlabeled data through a fixed threshold and ignore the different learning progress of each category, which makes the easy-to-learn categories have more examples contributing to the loss, resulting in a class-imbalance problem and affecting training efficiency. In order to improve the training reliability, we propose adaptive weighted losses (AWL). Through the evaluation of the class-wise learning progress, the loss contribution of the pseudo-labeled data of each category is continuously and dynamically adjusted during the learning process, and the pseudo-label discrimination ability of the model can be steadily improved. Moreover, to improve the training efficiency, we propose a bidirectional distribution approximation (DA) method, which introduces the consistency information of the predictions under the threshold into the loss calculation, and significantly improves the model convergence speed. Through the combination of AWL and DA, our method surpasses the performance of other algorithms on multiple benchmarks with a faster convergence efficiency, especially in the case of labeled data extremely limited. For example, AWL&DA achieves 95.29% test accuracy on the CIFAR-10-40-labels experiment and 92.56% accuracy on a faster experiment setting with only 2(18) iterations.
引用
收藏
页码:7832 / 7842
页数:11
相关论文
共 48 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]  
[Anonymous], 2017, INT C LEARN REPR ICL
[3]  
Athiwaratkun B., 2019, INT C LEARN REPR ICL
[4]  
Bachman P, 2014, ADV NEUR IN, V27
[5]  
Berthelot D., 2020, ICLR, V1, P3
[6]  
Berthelot D, 2019, ADV NEUR IN, V32
[7]   Attention-based label consistency for semi-supervised deep learning based image classification [J].
Chen, Jiaming ;
Yang, Meng ;
Ling, Jie .
NEUROCOMPUTING, 2021, 453 :731-741
[8]   Semi-Supervised Learning Based on Hybrid Neural Network for the Signal Integrity Analysis [J].
Chen, Siyu ;
Chen, Jienan ;
Zhang, Tingrui ;
Wei, Shuwu .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2020, 67 (10) :1934-1938
[9]   Deep Co-Space: Sample Mining Across Feature Transformation for Semi-Supervised Learning [J].
Chen, Ziliang ;
Wang, Keze ;
Wang, Xiao ;
Peng, Pai ;
Izquierdo, Ebroul ;
Lin, Liang .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2018, 28 (10) :2667-2678
[10]   Randaugment: Practical automated data augmentation with a reduced search space [J].
Cubuk, Ekin D. ;
Zoph, Barret ;
Shlens, Jonathon ;
Le, Quoc, V .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, :3008-3017