Query-support semantic correlation mining for few-shot segmentation

被引:9
作者
Shao, Ji [1 ,2 ,3 ,4 ]
Gong, Bo [1 ,2 ,3 ,4 ]
Dai, Kanyuan [2 ,3 ,4 ,5 ]
Li, Daoliang [1 ,2 ,3 ,4 ]
Jing, Ling [1 ,2 ,3 ,4 ,5 ]
Chen, Yingyi [1 ,2 ,3 ,4 ]
机构
[1] China Agr Univ, Coll Informat & Elect Engn, Beijing 100083, Peoples R China
[2] China Agr Univ, Natl Innovat Ctr Digital Fishery, Beijing 100083, Peoples R China
[3] Minist Agr & Rural Affairs, Key Lab Smart Farming Technol Aquat Anim & Livesto, Beijing 100083, Peoples R China
[4] China Agr Univ, Beijing Engn & Technol Res Ctr Internet Things Agr, Beijing 100083, Peoples R China
[5] China Agr Univ, Coll Sci, Beijing 100083, Peoples R China
关键词
Few-shot semantic segmentation; Semantic segmentation; Few-shot learning; Semantic correlation mining; Attention mechanism;
D O I
10.1016/j.engappai.2023.106797
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Given one or several annotation samples, few-shot segmentation has the capability to complete the segmentation of unseen target objects, thereby reducing the need for extensive data annotation. However, due to the scarcity and diversity of data, it becomes challenging to employ a few support samples effectively for guiding the segmentation of one query sample. Most existing approaches are committed to excavating support samples' information while ignoring the crucial relationship between query and support samples. To address this limitation and capture richer mutual information, we propose a query-support semantic correlation mining module that serves as an elaborate guide for subsequent dense matching. Specifically, two kinds of query guide maps are generated that leverage global and local similarities to learn the comprehensive correlation between multi-scale support and query features. By incorporating this module, the network can better execute dense matching and mitigate the possibility of misclassification effectively. In addition, a dual-attention module is introduced to obtain a discriminating initial prototype representation to avert prototype bias. Furthermore, our method generalizes well in the k-shot setting. Extensive experiments on PASCAL -5i and COCO-20i datasets demonstrate that the proposed method is potent and successful, which yields competitive segmentation results with state-of-the-art methods.
引用
收藏
页数:11
相关论文
共 51 条
[1]   SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation [J].
Badrinarayanan, Vijay ;
Kendall, Alex ;
Cipolla, Roberto .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) :2481-2495
[2]   MGNet: Mutual-guidance network for few-shot semantic segmentation [J].
Chang, Zhaobin ;
Lu, Yonggang ;
Wang, Xiangwen ;
Ran, Xingcheng .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2022, 116
[3]   RRCNet: Refinement residual convolutional network for breast ultrasound images segmentation [J].
Chen, Gongping ;
Dai, Yu ;
Zhang, Jianxun .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 117
[4]  
Chen LC, 2017, Arxiv, DOI arXiv:1706.05587
[5]   Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation [J].
Chen, Liang-Chieh ;
Zhu, Yukun ;
Papandreou, George ;
Schroff, Florian ;
Adam, Hartwig .
COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 :833-851
[6]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[7]   Self-regularized prototypical network for few-shot semantic segmentation [J].
Ding, Henghui ;
Zhang, Hui ;
Jiang, Xudong .
PATTERN RECOGNITION, 2023, 133
[8]   The Pascal Visual Object Classes (VOC) Challenge [J].
Everingham, Mark ;
Van Gool, Luc ;
Williams, Christopher K. I. ;
Winn, John ;
Zisserman, Andrew .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2010, 88 (02) :303-338
[9]  
Finn C, 2017, PR MACH LEARN RES, V70
[10]   Dual Attention Network for Scene Segmentation [J].
Fu, Jun ;
Liu, Jing ;
Tian, Haijie ;
Li, Yong ;
Bao, Yongjun ;
Fang, Zhiwei ;
Lu, Hanqing .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :3141-3149