Part-Aware Prototype Network for Few-Shot Semantic Segmentation

被引:227
作者
Liu, Yongfei [1 ]
Zhang, Xiangyi [1 ]
Zhang, Songyang [1 ]
He, Xuming [1 ,2 ]
机构
[1] ShanghaiTech Univ, Sch Informat Sci & Technol, Shanghai, Peoples R China
[2] Shanghai Engn Res Ctr Intelligent Vis & Imaging, Shanghai, Peoples R China
来源
COMPUTER VISION - ECCV 2020, PT IX | 2020年 / 12354卷
关键词
D O I
10.1007/978-3-030-58545-7_9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Few-shot semantic segmentation aims to learn to segment new object classes with only a few annotated examples, which has a wide range of real-world applications. Most existing methods either focus on the restrictive setting of one-way few-shot segmentation or suffer from incomplete coverage of object regions. In this paper, we propose a novel few-shot semantic segmentation framework based on the prototype representation. Our key idea is to decompose the holistic class representation into a set of part-aware prototypes, capable of capturing diverse and fine-grained object features. In addition, we propose to leverage unlabeled data to enrich our part-aware prototypes, resulting in better modeling of intra-class variations of semantic objects. We develop a novel graph neural network model to generate and enhance the proposed part-aware prototypes based on labeled and unlabeled images. Extensive experimental evaluations on two benchmarks show that our method outperforms the prior art with a sizable margin (Code is available at: https://github.com/Xiangyi1996/PPNet-PyTorch).
引用
收藏
页码:142 / 158
页数:17
相关论文
共 37 条
[1]  
Ayyad A., 2019, arXiv
[2]   SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation [J].
Badrinarayanan, Vijay ;
Kendall, Alex ;
Cipolla, Roberto .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) :2481-2495
[3]  
Boots Z.L.I.E.B., 2017, BRIT MACH VIS C BMVC
[4]  
Brabandere B.D., 2017, P IEEE C COMP VIS PA
[5]  
Chen LB, 2017, IEEE INT SYMP NANO, P1, DOI 10.1109/NANOARCH.2017.8053709
[6]  
Chung Y.A., 2017, NIPS MACH LEARN HLTH
[7]  
Dong N., 2018, BMVC, V3
[8]  
Finn C, 2017, PR MACH LEARN RES, V70
[9]  
Garcia Victor, 2017, arXiv
[10]  
Gori M, 2005, IEEE IJCNN, P729