Semi-supervised learning method based on distance metric loss framework

被引:0
作者
Liu B.-T. [1 ,2 ]
Ye Z.-T. [2 ]
Qin H.-L. [3 ]
Wang K. [1 ,4 ]
Zheng Q.-H. [1 ]
Wang Z.-Q. [1 ,2 ]
机构
[1] College of Information Science and Technology, Zhejiang Shuren University, Hangzhou
[2] College of Computer Science and Artificial Intelligence, Changzhou University, Changzhou
[3] Zhejiang Lvcheng Future Digital Intelligence Technology Limited Company, Hangzhou
[4] State Key Laboratory of Industrial Control Technology, Zhejiang University, Hangzhou
来源
Zhejiang Daxue Xuebao (Gongxue Ban)/Journal of Zhejiang University (Engineering Science) | 2023年 / 57卷 / 04期
关键词
classification; loss framework; loss function; metric learning; semi-supervised learning;
D O I
10.3785/j.issn.1008-973X.2023.04.012
中图分类号
学科分类号
摘要
A semi-supervised learning method based on the distance metric loss framework was proposed in order to solve the problems of different types of loss functions and inconsistent loss scales in the training process of semi-supervised learning methods, which make it difficult to adjust the loss weights, inconsistent optimization directions and insufficient generalization ability. A unify loss framework function was proposed from the perspective of distance metric loss, and the adjustment of loss weights between different loss functions in semi-supervised tasks was achieved. Adaptive similarity weights were introduced for the target region problem of embedding vectors in the loss framework in order to avoid the conflict of optimization directions of traditional metric learning loss functions and improve the generalization performance of the model. CNN13 and ResNet18 networks were used to construct semi-supervised learning models on CIFAR-10, CIFAR-100, SVHN, STL-10 standard image dataset and medical pneumonia dataset Pneumonia Chest X-ray, respectively, for comparison with commonly used semi-supervised methods in order to validate the effectiveness of the method. Results show that the method has the optimal classification accuracy under the condition of the same number of labels. © 2023 Zhejiang University. All rights reserved.
引用
收藏
页码:744 / 752
页数:8
相关论文
共 25 条
[1]  
KORNBLITH S, SHLENS J, LE Q V., Do better imagenet models transfer better? [C], Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2661-2671, (2019)
[2]  
YANG S, LUO P, LOY C C, Et al., WIDER FACE: a face detection benchmark [C], Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5525-5533, (2016)
[3]  
XU Jia-hui, WANG Jing-chang, CHEN Ling, Et al., Surface water quality prediction model based on graph neural network [J], Journal of Zhejiang University: Engineering Science, 55, 4, pp. 601-607, (2021)
[4]  
LEE D H., Pseudo-label: the simple and efficient semi-supervised learning method for deep neural networks, ICML 2013 Workshop on Challenges in Representation Learning, (2013)
[5]  
TARVAINEN A, VALPOLA H., Mean teachers are better role models: weight-averaged consistency targets improve semi-supervised deep learning results [J], Advances in Neural Information Processing Systems, 30, pp. 1195-1204, (2017)
[6]  
XIE Q, DAI Z, HOVV E, Et al., Unsupervised data augmentation for consistency training [J], Advances in Neural Information Processing Systems, 33, 2, pp. 6256-6268, (2020)
[7]  
MIYATO T, MAEDA S, KOYAMA M, Et al., Virtual adversarial training: a regularization method for supervised and semi-supervised learning [J], IEEE Transactions on Pattern Analysis and Machine Intelligence, 41, 8, pp. 1979-1993, (2018)
[8]  
WANG F, CHENG J, LIU W, Et al., Additive margin softmax for face verification [J], IEEE Signal Processing Letters, 25, 7, pp. 926-930, (2018)
[9]  
LAINE S, AILA T., Temporal ensembling for semi-supervised learning [C], International Conference on Learning Representations, pp. 1-13, (2017)
[10]  
SAJJADI M, JAVANMARDI M, TASDIZEN T., Regularization with stochastic transformations and perturbations for deep semi-supervised learning [J], Advances in Neural Information Processing Systems, 29, 7, pp. 1163-1171, (2016)