Learning from the Target: Dual Prototype Network for Few Shot Semantic Segmentation

被引:0
作者
Mao, Binjie [1 ,2 ]
Zhang, Xinbang [1 ,2 ]
Wang, Lingfeng [1 ]
Zhang, Qian [3 ]
Xiang, Shiming [1 ,2 ]
Pan, Chunhong [1 ]
机构
[1] Chinese Acad Sci, Inst Automat, Natl Lab Pattern Recognit, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing, Peoples R China
[3] Horizon Robot, Beijing, Peoples R China
来源
THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE | 2022年
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Due to the scarcity of annotated samples, the diversity between support set and query set becomes the main obstacle for few shot semantic segmentation. Most existing prototype-based approaches only exploit the prototype from the support feature and ignore the information from the query sample, failing to remove this obstacle.In this paper, we proposes a dual prototype network (DPNet) to dispose of few shot semantic segmentation from a new perspective. Along with the prototype extracted from the support set, we propose to build the pseudo-prototype based on foreground features in the query image. To achieve this goal, the cycle comparison module is developed to select reliable foreground features and generate the pseudo-prototype with them. Then, a prototype interaction module is utilized to integrate the information of the prototype and the pseudo-prototype based on their underlying correlation. Finally, a multi-scale fusion module is introduced to capture contextual information during the dense comparison between prototype (pseudo-prototype) and query feature. Extensive experiments conducted on two benchmarks demonstrate that our method exceeds previous state-of-the-arts with a sizable margin, verifying the effectiveness of the proposed method.
引用
收藏
页码:1953 / 1961
页数:9
相关论文
共 38 条
[1]  
[Anonymous], 2019, CVPR, DOI DOI 10.1109/CVPR.2019.01199
[2]  
[Anonymous], 2019, CVPR, DOI DOI 10.1109/CVPR.2019.00536
[3]  
[Anonymous], 2018, NERUIPS
[4]  
[Anonymous], 2020, CVPR, DOI DOI 10.1109/CVPR42600.2020.00422
[5]  
[Anonymous], 2021, ICASSP, DOI DOI 10.1109/ICASSP39728.2021.9414786
[6]  
Antoniou A., 2017, CoRR abs/1711.04340
[7]  
Cermelli F., 2020, FEW GUIDELINES INCRE
[8]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[9]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[10]   Temporal Cycle-Consistency Learning [J].
Dwibedi, Debidatta ;
Aytar, Yusuf ;
Tompson, Jonathan ;
Sermanet, Pierre ;
Zisserman, Andrew .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :1801-1810