Self-Supervised Learning of Pretext-Invariant Representations

被引:910
作者
Misra, Ishan [1 ]
van der Maaten, Laurens [1 ]
机构
[1] Facebook AI Res, Menlo Pk, CA 94025 USA
来源
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2020年
关键词
CODE;
D O I
10.1109/CVPR42600.2020.00674
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The goal of self-supervised learning from images is to construct image representations that are semantically meaningful via pretext tasks that do not require semantic annotations. Many pretext tasks lead to representations that are covariant with image transformations. We argue that, instead, semantic representations ought to be invariant under such transformations. Specifically, we develop Pretext-Invariant Representation Learning (PIRL, pronounced as "pearl") that learns invariant representations based on pretext tasks. We use PIRL with a commonly used pretext task that involves solving jigsaw puzzles. We find that PIRL substantially improves the semantic quality of the learned image representations. Our approach sets a new state-of-the-art in self-supervised learning from images on several popular benchmarks for self-supervised learning. Despite being unsupervised, PIRL outperforms supervised pre-training in learning image representations for object detection. Altogether, our results demonstrate the potential of self-supervised representations with good invariance properties.
引用
收藏
页码:6706 / 6716
页数:11
相关论文
共 88 条
[1]  
Ahsan Unaiza, 2019, WACV
[2]  
[Anonymous], 2017, CVPR
[3]  
[Anonymous], 2017, CVPR
[4]  
[Anonymous], 2008, P 25 INT C MACH LEAR, DOI DOI 10.1145/1390156.1390294
[5]  
[Anonymous], 1982, VISION COMPUTATIONAL
[6]  
[Anonymous], 2018, ECCV
[7]  
[Anonymous], 2017, ICCV
[8]  
[Anonymous], 2016, ICLR
[9]  
[Anonymous], 2016, CVPR
[10]  
[Anonymous], 2018, CVPR