Learning What and Where to Learn: A New Perspective on Self-Supervised Learning

被引:4
|
作者
Zhao, Wenyi [1 ]
Yang, Lu [1 ]
Zhang, Weidong [2 ]
Tian, Yongqin [2 ]
Jia, Wenhe [1 ]
Li, Wei [1 ]
Yang, Mu [3 ]
Pan, Xipeng [4 ]
Yang, Huihua [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Sch Artificial Intelligence, Beijing 100876, Peoples R China
[2] Henan Inst Sci & Technol, Sch Informat Engn, Xinxiang 453003, Peoples R China
[3] Techmach Beijing Ind Technol Co Ltd, Beijing 102676, Peoples R China
[4] Guilin Univ Elect Technol, Sch Comp Sci & Informat Secur, Guilin 541004, Peoples R China
基金
中国国家自然科学基金;
关键词
Semantics; Feature extraction; Task analysis; Computational modeling; Optimization; Self-supervised learning; Training; learning what; learning where; efficient framework; positional information;
D O I
10.1109/TCSVT.2023.3298937
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Self-supervised learning (SSL) has demonstrated its power in generalized model acquisition by leveraging the discriminative semantic and explicit positional information of unlabeled datasets. Unfortunately, mainstream contrastive learning-based methods excessive focus on semantic information and ignore the position is also the carrier of image content, resulting in inadequate data utilization and extensive computational consumption. To address these issues, we present an efficient SSL framework, learning What and Where to learn ( $\text {W}<^>{2} \text {SSL}$ ), to aggregate semantic and position features. Concretely, we devise a spatially-coupled sampling manner to process images through pre-defined rules, which integrates the advantage of semantic (What) and positional (Where) features into framework to enrich the diversity of feature representation capabilities and improve data utilization. Besides, a spectrum of latent vectors is obtained by mapping the positional features, which implicitly explores the relationship between these vectors. Whereafter, the corresponding discriminative and contrastive optimization objectives are seamlessly embedded in the framework via a cascade paradigm to explore semantic and positional features. The proposed $\text {W}<^>{2} \text {SSL}$ is verified on different types of datasets, which demonstrates that it still outperforms state-of-the-art SSL methods even with half the computational consumption. Code will be available at https://github.com/WilyZhao8/W2SSL.
引用
收藏
页码:6620 / 6633
页数:14
相关论文
共 50 条
  • [1] Reduce the Difficulty of Incremental Learning With Self-Supervised Learning
    Guan, Linting
    Wu, Yan
    IEEE ACCESS, 2021, 9 : 128540 - 128549
  • [2] Evolved Hierarchical Masking for Self-Supervised Learning
    Feng, Zhanzhou
    Zhang, Shiliang
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2025, 47 (02) : 1013 - 1027
  • [3] Embedding Global Contrastive and Local Location in Self-Supervised Learning
    Zhao, Wenyi
    Li, Chongyi
    Zhang, Weidong
    Yang, Lu
    Zhuang, Peixian
    Li, Lingqiao
    Fan, Kefeng
    Yang, Huihua
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (05) : 2275 - 2289
  • [4] Self-Supervised Synthesis Ranking for Deep Metric Learning
    Fu, Zheren
    Mao, Zhendong
    Yan, Chenggang
    Liu, An-An
    Xie, Hongtao
    Zhang, Yongdong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (07) : 4736 - 4750
  • [5] Self-Supervised Representation Learning for Document Image Classification
    Siddiqui, Shoaib Ahmed
    Dengel, Andreas
    Ahmed, Sheraz
    IEEE ACCESS, 2021, 9 : 164358 - 164367
  • [6] Self-Supervised Learning Based on Spatial Awareness for Medical Image Analysis
    Nguyen, Xuan-Bac
    Lee, Guee Sang
    Kim, Soo Hyung
    Yang, Hyung Jeong
    IEEE ACCESS, 2020, 8 (08): : 162973 - 162981
  • [7] Pretext Tasks Selection for Multitask Self-Supervised Audio Representation Learning
    Zaiem, Salah
    Parcollet, Titouan
    Essid, Slim
    Heba, Abdelwahab
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2022, 16 (06) : 1439 - 1453
  • [8] Comparing Learning Methodologies for Self-Supervised Audio-Visual Representation Learning
    Terbouche, Hacene
    Schoneveld, Liam
    Benson, Oisin
    Othmani, Alice
    IEEE ACCESS, 2022, 10 : 41622 - 41638
  • [9] Self-Supervised Learning for Electroencephalography
    Rafiei, Mohammad H.
    Gauthier, Lynne V.
    Adeli, Hojjat
    Takabi, Daniel
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (02) : 1457 - 1471
  • [10] Self-Supervised Feature Enhancement: Applying Internal Pretext Task to Supervised Learning
    Xie, Tianshu
    Yang, Yuhang
    Ding, Zilin
    Cheng, Xuan
    Wang, Xiaomin
    Gong, Haigang
    Liu, Ming
    IEEE ACCESS, 2023, 11 : 1708 - 1717