Pixel-Wise and Class-Wise Semantic Cues for Few-Shot Segmentation in Astronaut Working Scenes

被引:1
作者
Sun, Qingwei [1 ,2 ]
Chao, Jiangang [2 ,3 ]
Lin, Wanhong [2 ,3 ]
Wang, Dongyang [2 ,3 ]
Chen, Wei [2 ,3 ]
Xu, Zhenying [2 ,3 ]
Xie, Shaoli [2 ]
机构
[1] Space Engn Univ, Dept Aerosp Sci & Technol, Beijing 101416, Peoples R China
[2] China Astronaut Res & Training Ctr, Beijing 100094, Peoples R China
[3] China Astronaut Res & Training Ctr, Natl Key Lab Human Factors Engn, Beijing 100094, Peoples R China
关键词
few-shot semantic segmentation; astronaut working scenes; intelligent parsing; image processing; AGGREGATION; NETWORK;
D O I
10.3390/aerospace11060496
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
Few-shot segmentation (FSS) is a cutting-edge technology that can meet requirements using a small workload. With the development of China Aerospace Engineering, FSS plays a fundamental role in astronaut working scene (AWS) intelligent parsing. Although mainstream FSS methods have made considerable breakthroughs in natural data, they are not suitable for AWSs. AWSs are characterized by a similar foreground (FG) and background (BG), indistinguishable categories, and the strong influence of light, all of which place higher demands on FSS methods. We design a pixel-wise and class-wise network (PCNet) to match support and query features using pixel-wise and class-wise semantic cues. Specifically, PCNet extracts pixel-wise semantic information at each layer of the backbone using novel cross-attention. Dense prototypes are further utilized to extract class-wise semantic cues as a supplement. In addition, the deep prototype is distilled in reverse to the shallow layer to improve its quality. Furthermore, we customize a dataset for AWSs and conduct abundant experiments. The results indicate that PCNet outperforms the published best method by 4.34% and 5.15% in accuracy under one-shot and five-shot settings, respectively. Moreover, PCNet compares favorably with the traditional semantic segmentation model under the 13-shot setting.
引用
收藏
页数:18
相关论文
共 55 条
  • [1] SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation
    Badrinarayanan, Vijay
    Kendall, Alex
    Cipolla, Roberto
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) : 2481 - 2495
  • [2] Conceptual design and flight simulation of space stations
    Bertrand, R
    [J]. AEROSPACE SCIENCE AND TECHNOLOGY, 2001, 5 (02) : 147 - 163
  • [3] Boyu Yang, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12353), P763, DOI 10.1007/978-3-030-58598-3_45
  • [4] Chen J., 2021, arXiv preprint: arXiv:2102.04306
  • [5] Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation
    Chen, Liang-Chieh
    Zhu, Yukun
    Papandreou, George
    Schroff, Florian
    Adam, Hartwig
    [J]. COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 : 833 - 851
  • [6] Image Deformation Meta-Networks for One-Shot Learning
    Chen, Zitian
    Fu, Yanwei
    Wang, Yu-Xiong
    Ma, Lin
    Liu, Wei
    Hebert, Martial
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 8672 - 8681
  • [7] Dong E.P., 2018, P BRIT MACH VIS C, P79
  • [8] Dosovitskiy A, 2021, Arxiv, DOI arXiv:2010.11929
  • [9] The Pascal Visual Object Classes (VOC) Challenge
    Everingham, Mark
    Van Gool, Luc
    Williams, Christopher K. I.
    Winn, John
    Zisserman, Andrew
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2010, 88 (02) : 303 - 338
  • [10] Rethinking BiSeNet For Real-time Semantic Segmentation
    Fan, Mingyuan
    Lai, Shenqi
    Huang, Junshi
    Wei, Xiaoming
    Chai, Zhenhua
    Luo, Junfeng
    Wei, Xiaolin
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 9711 - 9720