Prototype-wise self-knowledge distillation for few-shot segmentation

被引:1
作者
Chen, Yadang [1 ,2 ]
Xu, Xinyu [1 ,2 ]
Wei, Chenchen [1 ,2 ]
Lu, Chuhan [3 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Comp & Software, Nanjing 210044, Peoples R China
[2] Nanjing Univ Informat Sci & Technol, Engn Res Ctr Digital Forens, Minist Educ, Nanjing 210044, Peoples R China
[3] Nanjing Univ Informat Sci & Technol, Sch Atmospher Sci, Nanjing 210044, Peoples R China
关键词
Few-shot segmentation; Data augmentation; Self-knowledge distillation;
D O I
10.1016/j.image.2024.117186
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Few-shot segmentation was proposed to obtain segmentation results for a image with an unseen class by referring to a few labeled samples. However, due to the limited number of samples, many few-shot segmentation models suffer from poor generalization. Prototypical network-based few-shot segmentation still has issues with spatial inconsistency and prototype bias. Since the target class has different appearance in each image, some specific features in the prototypes generated from the support image and its mask do not accurately reflect the generalized features of the target class. To address the support prototype consistency issue, we put forward two modules: Data Augmentation Self-knowledge Distillation (DASKD) and Prototype-wise Regularization (PWR). The DASKD module focuses on enhancing spatial consistency by using data augmentation and self-knowledge distillation. Self-knowledge distillation helps the model acquire generalized features of the target class and learn hidden knowledge from the support images. The PWR module focuses on obtaining a more representative support prototype by conducting prototype-level loss to obtain support prototypes closer to the category center. Broad evaluation experiments on PASCAL-5(t) and COCO-20(t) demonstrate that our model outperforms the prior works on few-shot segmentation. Our approach surpasses the state of the art by 7.5% in PASCAL-5(t) and 4.2% in COCO-20(t).
引用
收藏
页数:9
相关论文
共 45 条
[1]   Few-Shot Segmentation Without Meta-Learning: A Good Transductive Inference Is All You Need? [J].
Boudiaf, Malik ;
Kervadec, Hoel ;
Masud, Ziko Imtiaz ;
Piantanida, Pablo ;
Ben Ayed, Ismail ;
Dolz, Jose .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :13974-13983
[2]  
Chaurasia Abhishek, 2017, arXiv
[3]   Distilling knowledge from ensembles of neural networks for speech recognition [J].
Chebotar, Yevgen ;
Waters, Austin .
17TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2016), VOLS 1-5: UNDERSTANDING SPEECH PROCESSING IN HUMANS AND MACHINES, 2016, :3439-3443
[4]   Holistic Prototype Activation for Few-Shot Segmentation [J].
Cheng, Gong ;
Lang, Chunbo ;
Han, Junwei .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (04) :4650-4666
[5]   The Pascal Visual Object Classes (VOC) Challenge [J].
Everingham, Mark ;
Van Gool, Luc ;
Williams, Christopher K. I. ;
Winn, John ;
Zisserman, Andrew .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2010, 88 (02) :303-338
[6]   Self-support Few-Shot Semantic Segmentation [J].
Fan, Qi ;
Pei, Wenjie ;
Tai, Yu-Wing ;
Tang, Chi-Keung .
COMPUTER VISION, ECCV 2022, PT XIX, 2022, 13679 :701-719
[7]   An adaptive infrared image segmentation method based on fusion SPCNN [J].
Guo, Zhengkun ;
Song, Yong ;
Zhao, Yufei ;
Yang, Xin ;
Wang, Fengning .
SIGNAL PROCESSING-IMAGE COMMUNICATION, 2020, 87
[8]   Simultaneous Detection and Segmentation [J].
Hariharan, Bharath ;
Arbelaez, Pablo ;
Girshick, Ross ;
Malik, Jitendra .
COMPUTER VISION - ECCV 2014, PT VII, 2014, 8695 :297-312
[9]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[10]  
Hinton G., NIPS 2014 DEEP LEARN