Exploring Cross-Image Pixel Contrast for Semantic Segmentation

被引:345
作者
Wang, Wenguan [1 ]
Zhou, Tianfei [1 ]
Yu, Fisher [1 ]
Dai, Jifeng [2 ]
Konukoglu, Ender [1 ]
Van Gool, Luc [1 ]
机构
[1] Swiss Fed Inst Technol, Comp Vis Lab, Zurich, Switzerland
[2] SenseTime Res, Hong Kong, Peoples R China
来源
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021) | 2021年
关键词
D O I
10.1109/ICCV48922.2021.00721
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Current semantic segmentation methods focus only on mining "local" context, i.e., dependencies between pixels within individual images, by context-aggregation modules (e.g., dilated convolution, neural attention) or structure-aware optimization criteria (e.g., IoU-like loss). However, they ignore "global" context of the training data, i.e., rich semantic relations between pixels across different images. Inspired by recent advance in unsupervised contrastive representation learning, we propose a pixel-wise contrastive algorithm for semantic segmentation in the fully supervised setting. The core idea is to enforce pixel embeddings belonging to a same semantic class to be more similar than embeddings from different classes. It raises a pixel-wise metric learning paradigm for semantic segmentation, by explicitly exploring the structures of labeled pixels, which were rarely explored before. Our method can be effortlessly incorporated into existing segmentation frameworks without extra overhead during testing. We experimentally show that, with famous segmentation models (i.e., DeepLabV3, HRNet, OCR) and backbones (i.e., ResNet, HRNet), our method brings performance improvements across diverse datasets (i.e., Cityscapes, PASCAL-Context, COCO-Stuff, CamVid). We expect this work will encourage our community to rethink the current de facto training paradigm in semantic segmentation.
引用
收藏
页码:7283 / 7293
页数:11
相关论文
共 88 条
  • [1] [Anonymous], 2021, CVPR, DOI DOI 10.1109/CVPR46437.2021.00905
  • [2] [Anonymous], 2019, ECCV, DOI DOI 10.1007/S13143-018-0064-5
  • [3] [Anonymous], 2018, CVPR, DOI DOI 10.1109/CVPR.2018.00132
  • [4] [Anonymous], 2020, ICML
  • [5] [Anonymous], 2018, ECCV, DOI DOI 10.1007/978-3-030-01246-5_36
  • [6] [Anonymous], 2020, ICML
  • [7] [Anonymous], 2019, SYMMETRY-BASEL, DOI DOI 10.3390/sym11091066
  • [8] [Anonymous], 2019, 2019 AES INT C
  • [9] Bai Min, 2017, CVPR
  • [10] The Lovasz-Softmax loss: A tractable surrogate for the optimization of the intersection-over-union measure in neural networks
    Berman, Maxim
    Triki, Amal Rannen
    Blaschko, Matthew B.
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 4413 - 4421