Survey on Self-Supervised Learning: Auxiliary Pretext Tasks and Contrastive Learning Methods in Imaging

被引:74
作者
Albelwi, Saleh [1 ,2 ]
机构
[1] Univ Tabuk, Fac Comp & Informat Technol, Tabuk 47731, Saudi Arabia
[2] Univ Tabuk, Ind Innovat & Robot Ctr IIRC, Tabuk 47731, Saudi Arabia
关键词
self-supervised learning (SSL); auxiliary pretext tasks; contrastive learning; pretext tasks; data augmentation; contrastive loss; encoder; downstream tasks;
D O I
10.3390/e24040551
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Although deep learning algorithms have achieved significant progress in a variety of domains, they require costly annotations on huge datasets. Self-supervised learning (SSL) using unlabeled data has emerged as an alternative, as it eliminates manual annotation. To do this, SSL constructs feature representations using pretext tasks that operate without manual annotation, which allows models trained in these tasks to extract useful latent representations that later improve downstream tasks such as object classification and detection. The early methods of SSL are based on auxiliary pretext tasks as a way to learn representations using pseudo-labels, or labels that were created automatically based on the dataset's attributes. Furthermore, contrastive learning has also performed well in learning representations via SSL. To succeed, it pushes positive samples closer together, and negative ones further apart, in the latent space. This paper provides a comprehensive literature review of the top-performing SSL methods using auxiliary pretext and contrastive learning techniques. It details the motivation for this research, a general pipeline of SSL, the terminologies of the field, and provides an examination of pretext tasks and self-supervised methods. It also examines how self-supervised methods compare to supervised ones, and then discusses both further considerations and ongoing challenges faced by SSL.
引用
收藏
页数:22
相关论文
共 80 条
  • [1] Albelwi Saleh A., 2022, 2022 2nd International Conference on Computing and Information Technology (ICCIT), P349, DOI 10.1109/ICCIT52419.2022.9711630
  • [2] Appalaraju S., 2020, ARXIV201200868
  • [3] Asano Y., 2019, INT C LEARN REPR
  • [4] Bachman P, 2019, ADV NEUR IN, V32
  • [5] Balestriero R., 2022, ARXIV220208325
  • [6] Caron M, 2020, ADV NEUR IN, V33
  • [7] Deep Clustering for Unsupervised Learning of Visual Features
    Caron, Mathilde
    Bojanowski, Piotr
    Joulin, Armand
    Douze, Matthijs
    [J]. COMPUTER VISION - ECCV 2018, PT XIV, 2018, 11218 : 139 - 156
  • [8] Self-supervised learning for medical image analysis using image context restoration
    Chen, Liang
    Bentley, Paul
    Mori, Kensaku
    Misawa, Kazunari
    Fujiwara, Michitaka
    Rueckert, Daniel
    [J]. MEDICAL IMAGE ANALYSIS, 2019, 58
  • [9] DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs
    Chen, Liang-Chieh
    Papandreou, George
    Kokkinos, Iasonas
    Murphy, Kevin
    Yuille, Alan L.
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) : 834 - 848
  • [10] Chen T, 2020, PR MACH LEARN RES, V119