SVDML: Semantic and Visual Space Deep Mutual Learning for Zero-Shot Learning

被引:0
作者
Lu, Nannan [1 ]
Luo, Yi [1 ]
Qiu, Mingkai [1 ]
机构
[1] China Univ Min & Technol, Sch Informat & Control Engn, Xuzhou 221100, Jiangsu, Peoples R China
来源
PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT IX | 2024年 / 14433卷
基金
中国国家自然科学基金;
关键词
Zero-shot Learning; Semantic Representation; Visual Representation; Mutual Learning;
D O I
10.1007/978-981-99-8546-3_31
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The key challenge of zero-shot learning (ZSL) is how to identify novel objects for which no samples are available during the training process. Current approaches either align the global features of images to the corresponding class semantic vectors or use unidirectional attentions to locate the local visual features of images via semantic attributes to avoid interference from other noise in the image. However, they still have not found a way to establish a robust correlation between the semantic and visual representation. To solve the issue, we propose a Semantic and Visual space Deep Mutual Learning (SVDML), which consists of three modules: class representation learning, attribute embedding, and mutual learning, to establish the intrinsic semantic relations between visual features and attribute features. SVDML uses two kinds of prototype generators to separately guide the learning of global and local features of images and achieves interaction between two learning pipelines by mutual learning, so that promotes the recognition of the fine-grained features and strengthens the knowledge generalization ability in zero-shot learning. The proposed SVDML yields significant improvements over the strong baselines, leading to the new state-of the-art performances on three popular challenging benchmarks.
引用
收藏
页码:383 / 395
页数:13
相关论文
共 24 条
[1]  
Chen S., 2022, DSP: dynamic semantic prototype for generative zero-shot learning
[2]  
Chen SM, 2021, ADV NEUR IN, V34
[3]   MSDN: Mutually Semantic Distillation Network for Zero-Shot Learning [J].
Chen, Shiming ;
Hong, Ziming ;
Xie, Guo-Sen ;
Yang, Wenhan ;
Peng, Qinmu ;
Wang, Kai ;
Zhao, Jian ;
You, Xinge .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, :7602-7611
[4]   FREE: Feature Refinement for Generalized Zero-Shot Learning [J].
Chen, Shiming ;
Wang, Wenjie ;
Xia, Beihao ;
Peng, Qinmu ;
You, Xinge ;
Zheng, Feng ;
Shao, Ling .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :122-131
[5]   Fine-Grained Generalized Zero-Shot Learning via Dense Attribute-Based Attention [J].
Dat Huynh ;
Elhamifar, Ehsan .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :4482-4492
[6]   Link the head to the "beak": Zero Shot Learning from Noisy Text Description at Part Precision [J].
Elhoseiny, Mohamed ;
Zhu, Yizhe ;
Zhang, Han ;
Elgammal, Ahmed .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6288-6297
[7]  
Farhadi A, 2009, PROC CVPR IEEE, P1778, DOI 10.1109/CVPRW.2009.5206772
[8]  
Fu YW, 2014, LECT NOTES COMPUT SC, V8690, P584, DOI 10.1007/978-3-319-10605-2_38
[9]  
Guo JC, 2023, AAAI CONF ARTIF INTE, P7775
[10]   Learning Classifiers of Prototypes and Reciprocal Points for Universal Domain Adaptation [J].
Hur, Sungsu ;
Shin, Inkyu ;
Park, Kwanyong ;
Woo, Sanghyun ;
Kweon, In So .
2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, :531-540