Self-Supervised Learning via Maximum Entropy Coding

被引:0
|
作者
Liu, Xin [1 ]
Wang, Zhongdao [1 ]
Li, Yali [1 ]
Wang, Shengjin [1 ]
机构
[1] Tsinghua Univ, Dept Elect Engn, Beijing Natl Res Ctr Informat Sci & Technol BNRis, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A mainstream type of current self-supervised learning methods pursues a general-purpose representation that can be well transferred to downstream tasks, typically by optimizing on a given pretext task such as instance discrimination. In this work, we argue that existing pretext tasks inevitably introduce biases into the learned representation, which in turn leads to biased transfer performance on various downstream tasks. To cope with this issue, we propose Maximum Entropy Coding (MEC), a more principled objective that explicitly optimizes on the structure of the representation, so that the learned representation is less biased and thus generalizes better to unseen downstream tasks. Inspired by the principle of maximum entropy in information theory, we hypothesize that a generalizable representation should be the one that admits the maximum entropy among all plausible representations. To make the objective end-to-end trainable, we propose to leverage the minimal coding length in lossy data coding as a computationally tractable surrogate for the entropy, and further derive a scalable reformulation of the objective that allows fast computation. Extensive experiments demonstrate that MEC learns a more generalizable representation than previous methods based on specific pretext tasks. It achieves state-of-the-art performance consistently on various downstream tasks, including not only ImageNet linear probe, but also semi-supervised classification, object detection, instance segmentation, and object tracking. Interestingly, we show that existing batch-wise and feature-wise self-supervised objectives could be seen equivalent to low-order approximations of MEC. Code and pre-trained models are available at https://github.com/xinliu20/MEC.
引用
收藏
页数:15
相关论文
共 50 条
  • [31] Graph Self-supervised Learning via Proximity Divergence Minimization
    Zhang, Tianyi
    Dai, Zhenwei
    Xu, Zhaozhuo
    Shrivastava, Anshumali
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 2498 - 2508
  • [32] Evaluate AMR Graph Similarity via Self-supervised Learning
    Shou, Ziyi
    Lin, Fangzhen
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 16112 - 16123
  • [33] Self-supervised graph representation learning via positive mining
    Lee, Namkyeong
    Lee, Junseok
    Park, Chanyoung
    INFORMATION SCIENCES, 2022, 611 : 476 - 493
  • [34] Audio Mixing Inversion via Embodied Self-supervised Learning
    Haotian Zhou
    Feng Yu
    Xihong Wu
    Machine Intelligence Research, 2024, 21 : 55 - 62
  • [35] Repeatable adaptive keypoint detection via self-supervised learning
    Pei YAN
    Yihua TAN
    Yuan TAI
    Science China(Information Sciences), 2022, 65 (11) : 123 - 147
  • [36] SELF-SUPERVISED ACOUSTIC ANOMALY DETECTION VIA CONTRASTIVE LEARNING
    Hojjati, Hadi
    Armanfard, Narges
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 3253 - 3257
  • [37] Enhancing robust VQA via contrastive and self-supervised learning
    Cao, Runlin
    Li, Zhixin
    Tang, Zhenjun
    Zhang, Canlong
    Ma, Huifang
    PATTERN RECOGNITION, 2025, 159
  • [38] Architecture Colorization via Self-supervised Learning and Instance Segmentation
    Liu, Sen
    Chen, Hang
    Li, Li
    PATTERN RECOGNITION AND COMPUTER VISION, PT I, PRCV 2022, 2022, 13534 : 3 - 16
  • [39] Self-Supervised Exploration via Temporal Inconsistency in Reinforcement Learning
    Gao Z.
    Xu K.
    Zhai Y.
    Ding B.
    Feng D.
    Mao X.
    Wang H.
    IEEE Transactions on Artificial Intelligence, 2024, 5 (11): : 1 - 10
  • [40] Self-Supervised Visual Representation Learning via Residual Momentum
    Pham, Trung Xuan
    Niu, Axi
    Zhang, Kang
    Jin, Tee Joshua Tian
    Hong, Ji Woo
    Yoo, Chang D.
    IEEE ACCESS, 2023, 11 : 116706 - 116720