Delving into Inter-Image Invariance for Unsupervised Visual Representations

被引:0
作者
Jiahao Xie
Xiaohang Zhan
Ziwei Liu
Yew-Soon Ong
Chen Change Loy
机构
[1] Nanyang Technological University,
[2] The Chinese University of Hong Kong,undefined
来源
International Journal of Computer Vision | 2022年 / 130卷
关键词
Unsupervised learning; Self-supervised learning; Representation learning; Contrastive learning; Inter-image invariance;
D O I
暂无
中图分类号
学科分类号
摘要
Contrastive learning has recently shown immense potential in unsupervised visual representation learning. Existing studies in this track mainly focus on intra-image invariance learning. The learning typically uses rich intra-image transformations to construct positive pairs and then maximizes agreement using a contrastive loss. The merits of inter-image invariance, conversely, remain much less explored. One major obstacle to exploit inter-image invariance is that it is unclear how to reliably construct inter-image positive pairs, and further derive effective supervision from them since no pair annotations are available. In this work, we present a comprehensive empirical study to better understand the role of inter-image invariance learning from three main constituting components: pseudo-label maintenance, sampling strategy, and decision boundary design. To facilitate the study, we introduce a unified and generic framework that supports the integration of unsupervised intra- and inter-image invariance learning. Through carefully-designed comparisons and analysis, multiple valuable observations are revealed: 1) online labels converge faster and perform better than offline labels; 2) semi-hard negative samples are more reliable and unbiased than hard negative samples; 3) a less stringent decision boundary is more favorable for inter-image invariance learning. With all the obtained recipes, our final model, namely InterCLR, shows consistent improvements over state-of-the-art intra-image invariance learning methods on multiple standard benchmarks. We hope this work will provide useful experience for devising effective unsupervised inter-image invariance learning. Code: https://github.com/open-mmlab/mmselfsup.
引用
收藏
页码:2994 / 3013
页数:19
相关论文
共 24 条
  • [1] Ericsson L(2022)Self-supervised representation learning: Introduction, advances, and challenges IEEE Signal Processing Magazine 39 42-62
  • [2] Gouk H(2010)The pascal visual object classes (voc) challenge IJCV 88 303-338
  • [3] Loy CC(2008)Liblinear: A library for large linear classification JMLR 9 1871-1874
  • [4] Hospedales TM(2008)Visualizing data using t-sne Journal of machine learning research 9 2579-2605
  • [5] Everingham M(2018)Virtual adversarial training: a regularization method for supervised and semi-supervised learning TPAMI 41 1979-1993
  • [6] Van Gool L(2018)Additive margin softmax for face verification IEEE Signal Processing Letters undefined undefined-undefined
  • [7] Williams CK(undefined)undefined undefined undefined undefined-undefined
  • [8] Winn J(undefined)undefined undefined undefined undefined-undefined
  • [9] Zisserman A(undefined)undefined undefined undefined undefined-undefined
  • [10] Fan RE(undefined)undefined undefined undefined undefined-undefined