Semi-supervised Learning Using Variational Autoencoder - A Cluster Based Approach

被引:0
作者
Vengalil, Sunil Kumar [1 ]
Sinha, Neelam [1 ]
机构
[1] Int Inst Informat Technol Bangalore, Bangalore, Karnataka, India
来源
PATTERN RECOGNITION AND MACHINE INTELLIGENCE, PREMI 2021 | 2024年 / 13102卷
关键词
Semi-supervised Learning; Variational Autoencoder; Active Learning; Clustering;
D O I
10.1007/978-3-031-12700-7_54
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The successful application of deep neural networks for solving complex tasks like image classification, object detection and segmentation depends critically on the availability of large number of labelled training samples. To achieve good generalization for a reasonably complex model with about 60 million parameters, as in AlexNet, one needs about one million labelled training samples. In almost all practical applications, like natural image classification and segmentation, plenty of unlabelled samples are available but labelling these samples is a tedious manual task. We introduce a novel mechanism to automatically label all the samples in an unlabelled dataset. Starting with completely unlabelled dataset, an iterative algorithm incrementally assigns labels along with a confidence to all training samples. During each iteration, 10-30 new representative samples are generated in a latent space learned using a variational autoencoder and labels for these samples are obtained from a human expert. The proposed idea is demonstrated on MNIST dataset without using the labels provided in the dataset. At regular intervals of training, the low dimensional latent vectors are clustered and only cluster centers are annotated. The manual labels of cluster centers are propagated to other samples in the cluster based on the distance and a confidence function. The loss function in successive training is modified to incorporate the manual information provided. We run multiple experiments with different choices of clustering algorithm, confidence function and distance metric and compare the results. With GMM clustering, best classification accuracy of 93.9% was obtained on MNIST test images after 5 iterations.
引用
收藏
页码:529 / 536
页数:8
相关论文
共 21 条
[1]  
[Anonymous], 2017, ICLR
[2]  
Belkin M, 2006, J MACH LEARN RES, V7, P2399
[3]   The power of ensembles for active learning in image classification [J].
Beluch, William H. ;
Genewein, Tim ;
Nuernberger, Andreas ;
Koehler, Jan M. .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :9368-9377
[4]  
Bennett KP, 1999, ADV NEUR IN, V11, P368
[5]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[6]   A Novel Manifold Regularized Online Semi-supervised Learning Model [J].
Ding, Shuguang ;
Xi, Xuanyang ;
Liu, Zhiyong ;
Qiao, Hong ;
Zhang, Bo .
COGNITIVE COMPUTATION, 2018, 10 (01) :49-61
[7]  
Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672
[8]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[9]  
Joachims T, 1999, MACHINE LEARNING, PROCEEDINGS, P200
[10]  
King DB, 2015, ACS SYM SER, V1214, P1, DOI 10.1021/bk-2015-1214.ch001