On semi-supervised learning

被引:2
|
作者
Cholaquidis, A. [1 ]
Fraiman, R. [1 ]
Sued, M. [2 ]
机构
[1] Univ Republica, Fac Ciencias, Montevideo, Uruguay
[2] INst Calculo, Fac Ciencias Exactas & Nat, Buenos Aires, DF, Argentina
关键词
Semi-supervised learning; Small training sample; Consistency; PATTERN-RECOGNITION; ERROR;
D O I
10.1007/s11749-019-00690-2
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Major efforts have been made, mostly in the machine learning literature, to construct good predictors combining unlabelled and labelled data. These methods are known as semi-supervised. They deal with the problem of how to take advantage, if possible, of a huge amount of unlabelled data to perform classification in situations where there are few labelled data. This is not always feasible: it depends on the possibility to infer the labels from the unlabelled data distribution. Nevertheless, several algorithms have been proposed recently. In this work, we present a new method that, under almost necessary conditions, attains asymptotically the performance of the best theoretical rule when the size of the unlabelled sample goes to infinity, even if the size of the labelled sample remains fixed. Its performance and computational time are assessed through simulations and in the well- known "Isolet" real data of phonemes, where a strong dependence on the choice of the initial training sample is shown. The main focus of this work is to elucidate when and why semi-supervised learning works in the asymptotic regime described above. The set of necessary assumptions, although reasonable, show that semi-parametric methods only attain consistency for very well-conditioned problems.
引用
收藏
页码:914 / 937
页数:24
相关论文
共 50 条
  • [31] Feature ranking for semi-supervised learning
    Matej Petković
    Sašo Džeroski
    Dragi Kocev
    Machine Learning, 2023, 112 : 4379 - 4408
  • [32] Multiview Semi-Supervised Learning with Consensus
    Li, Guangxia
    Chang, Kuiyu
    Hoi, Steven C. H.
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2012, 24 (11) : 2040 - 2051
  • [33] Neighbor Matching for Semi-supervised Learning
    Wang, Renzhen
    Wu, Yichen
    Chen, Huai
    Wang, Lisheng
    Meng, Deyu
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2021, PT II, 2021, 12902 : 439 - 449
  • [34] A topological approach for semi-supervised learning
    Ines, A.
    Dominguez, C.
    Heras, J.
    Mata, G.
    Rubio, J.
    JOURNAL OF COMPUTATIONAL SCIENCE, 2024, 82
  • [35] Semi-supervised learning with regularized Laplacian
    Avrachenkov, K.
    Chebotarev, P.
    Mishenin, A.
    OPTIMIZATION METHODS & SOFTWARE, 2017, 32 (02) : 222 - 236
  • [36] Semi-Supervised Learning of Speech Sounds
    Jansen, Aren
    Niyogi, Partha
    INTERSPEECH 2007: 8TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION, VOLS 1-4, 2007, : 2264 - 2267
  • [37] Image categorization with semi-supervised learning
    Yu, Zhenghua
    2006 IEEE International Conference on Image Processing, ICIP 2006, Proceedings, 2006, : 3173 - 3176
  • [38] FMixCutMatch for semi-supervised deep learning
    Wei, Xiang
    Wei, Xiaotao
    Kong, Xiangyuan
    Lu, Siyang
    Xing, Weiwei
    Lu, Wei
    NEURAL NETWORKS, 2021, 133 : 166 - 176
  • [39] AN OPERATOR METHOD FOR SEMI-SUPERVISED LEARNING
    Lu, Wei-Jun
    Bai, Yan
    Tang, Yi
    Tao, Yan-Fang
    PROCEEDINGS OF 2009 INTERNATIONAL CONFERENCE ON WAVELET ANALYSIS AND PATTERN RECOGNITION, 2009, : 123 - +
  • [40] Semi-supervised Deep Learning with Memory
    Chen, Yanbei
    Zhu, Xiatian
    Gong, Shaogang
    COMPUTER VISION - ECCV 2018, PT I, 2018, 11205 : 275 - 291