Self-training: A survey

被引:1
|
作者
Amini, Massih-Reza [1 ]
Feofanov, Vasilii [1 ]
Pauletto, Loic [1 ]
Hadjadj, Lies [1 ]
Devijver, Emilie [1 ]
Maximov, Yury [2 ]
机构
[1] Univ Grenoble Alpes, CNRS, Lab Informat Grenoble, Grenoble, France
[2] Los Alamos Natl Lab, Theoret Div, Los Alamos, NM USA
关键词
Semi-supervised learning; Self-training;
D O I
10.1016/j.neucom.2024.128904
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Self-training methods have gained significant attention in recent years due to their effectiveness in leveraging small labeled datasets and large unlabeled observations for prediction tasks. These models identify decision boundaries in low-density regions without additional assumptions about data distribution, using the confidence scores of a learned classifier. The core principle of self-training involves iteratively assigning pseudo-labels to unlabeled samples with confidence scores above a certain threshold, enriching the labeled dataset and retraining the classifier. This paper presents self-training methods for binary and multi-class classification, along with variants and related approaches such as consistency-based methods and transductive learning. We also briefly describe self-supervised learning and reinforced self-training. Furthermore, we highlight popular applications of self-training and discuss the importance of dynamic thresholding and reducing pseudo-label noise for performance improvement. To the best of our knowledge, this is the first thorough and complete survey on self-training.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Self-training ABS
    Akhmetshin, A.M.
    Avtomobil'naya Promyshlennost, 2001, (06): : 34 - 36
  • [2] SETRED: Self-training with editing
    Li, M
    Zhou, ZH
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PROCEEDINGS, 2005, 3518 : 611 - 621
  • [3] Deep Bayesian Self-Training
    Fabio De Sousa Ribeiro
    Francesco Calivá
    Mark Swainson
    Kjartan Gudmundsson
    Georgios Leontidis
    Stefanos Kollias
    Neural Computing and Applications, 2020, 32 : 4275 - 4291
  • [4] Confidence Regularized Self-Training
    Zou, Yang
    Yu, Zhiding
    Liu, Xiaofeng
    Kumar, B. V. K. Vijaya
    Wang, Jinsong
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 5981 - 5990
  • [6] Self-Training with Weak Supervision
    Karamanolakis, Giannis
    Mukherjee, Subhabrata
    Zheng, Guoqing
    Awadallah, Ahmed Hassan
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 845 - 863
  • [7] Doubly Robust Self-Training
    Zhu, Banghua
    Ding, Mingyu
    Jacobson, Philip
    Wu, Ming
    Zhan, Wei
    Jordan, Michael I.
    Jiao, Jiantao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [8] KUDOS FOR SELF-TRAINING AIDS
    BRYANT, SF
    COMPUTER DECISIONS, 1984, 16 (14): : 44 - &
  • [9] Deep Bayesian Self-Training
    Ribeiro, Fabio De Sousa
    Caliva, Francesco
    Swainson, Mark
    Gudmundsson, Kjartan
    Leontidis, Georgios
    Kollias, Stefanos
    NEURAL COMPUTING & APPLICATIONS, 2020, 32 (09): : 4275 - 4291
  • [10] RECURSIVE SELF-TRAINING ALGORITHMS
    TSYPKIN, YZ
    KELMANS, GK
    ENGINEERING CYBERNETICS, 1967, (05): : 70 - &