Self-Supervised Learning with an Information Maximization Criterion

被引:0
作者
Ozsoy, Serdar [1 ,2 ]
Hamdan, Shadi [1 ,3 ]
Arik, Sercan O. [4 ]
Yuret, Deniz [1 ,3 ]
Erdogan, Alper T. [1 ,2 ]
机构
[1] Koc Univ, KUIS AI Ctr, Istanbul, Turkiye
[2] Koc Univ, EEE Dept, Istanbul, Turkiye
[3] Koc Univ, CE Dept, Istanbul, Turkiye
[4] Google Cloud Res, Sunnyvale, CA USA
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022) | 2022年
关键词
MATRIX FACTORIZATION; FRAMEWORK;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Self-supervised learning allows AI systems to learn effective representations from large amounts of data using tasks that do not require costly labeling. Mode collapse, i.e., the model producing identical representations for all inputs, is a central problem to many self-supervised learning approaches, making self-supervised tasks, such as matching distorted variants of the inputs, ineffective. In this article, we argue that a straightforward application of information maximization among alternative latent representations of the same input naturally solves the collapse problem and achieves competitive empirical results. We propose a self-supervised learning method, CorInfoMax, that uses a second-order statistics-based mutual information measure that reflects the level of correlation among its arguments. Maximizing this correlative information measure between alternative representations of the same input serves two purposes: (1) it avoids the collapse problem by generating feature vectors with non-degenerate covariances; (2) it establishes relevance among alternative representations by increasing the linear dependence among them. An approximation of the proposed information maximization objective simplifies to a Euclidean distance-based objective function regularized by the log-determinant of the feature covariance matrix. The regularization term acts as a natural barrier against feature space degeneracy. Consequently, beyond avoiding complete output collapse to a single point, the proposed approach also prevents dimensional collapse by encouraging the spread of information across the whole feature space. Numerical experiments demonstrate that CorInfoMax achieves better or competitive performance results relative to the state-of-the-art SSL approaches.
引用
收藏
页数:14
相关论文
共 56 条
  • [1] Asano Y., 2019, INT C LEARN REPR
  • [2] An Algorithmic Framework for Sparse Bounded Component Analysis
    Babatas, Eren
    Erdogan, Alper T.
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2018, 66 (19) : 5194 - 5205
  • [3] Bardes Adrien, 2021, ARXIV210504906
  • [4] SELF-ORGANIZING NEURAL NETWORK THAT DISCOVERS SURFACES IN RANDOM-DOT STEREOGRAMS
    BECKER, S
    HINTON, GE
    [J]. NATURE, 1992, 355 (6356) : 161 - 163
  • [5] AN INFORMATION MAXIMIZATION APPROACH TO BLIND SEPARATION AND BLIND DECONVOLUTION
    BELL, AJ
    SEJNOWSKI, TJ
    [J]. NEURAL COMPUTATION, 1995, 7 (06) : 1129 - 1159
  • [6] BUNCH JR, 1974, MATH COMPUT, V28, P231, DOI 10.1090/S0025-5718-1974-0331751-8
  • [7] Caron M, 2020, ADV NEUR IN, V33
  • [8] Deep Clustering for Unsupervised Learning of Visual Features
    Caron, Mathilde
    Bojanowski, Piotr
    Joulin, Armand
    Douze, Matthijs
    [J]. COMPUTER VISION - ECCV 2018, PT XIV, 2018, 11218 : 139 - 156
  • [9] A Simplex Volume Maximization Framework for Hyperspectral Endmember Extraction
    Chan, Tsung-Han
    Ma, Wing-Kin
    Ambikapathi, ArulMurugan
    Chi, Chong-Yung
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2011, 49 (11): : 4177 - 4193
  • [10] Exploring Simple Siamese Representation Learning
    Chen, Xinlei
    He, Kaiming
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 15745 - 15753