Co-GZSL: Feature Contrastive Optimization for Generalized Zero-Shot Learning

被引:0
作者
Qun Li
Zhuxi Zhan
Yaying Shen
Bir Bhanu
机构
[1] Nanjing University of Posts and Telecommunications,School of Computer Science
[2] University of California at Riverside,Department of Electrical and Computer Engineering
来源
Neural Processing Letters | / 56卷
关键词
Generalized zero-shot learning; Contrastive loss; Semantic features; Visual features; Synthesized features;
D O I
暂无
中图分类号
学科分类号
摘要
Generalized Zero-Shot Learning (GZSL) learns from only labeled seen classes during training but discriminates both seen and unseen classes during testing. In GZSL tasks, most of the existing methods commonly utilize visual and semantic features for training. Due to the lack of visual features for unseen classes, recent works generate real-like visual features by using semantic features. However, the synthesized features in the original feature space lack discriminative information. It is important that the synthesized visual features should be similar to the ones in the same class, but different from the other classes. One way to solve this problem is to introduce the embedding space after generating visual features. Following this situation, the embedded features from the embedding space can be inconsistent with the original semantic features. For another way, some recent methods constrain the representation by reconstructing the semantic features using the original visual features and the synthesized visual features. In this paper, we propose a hybrid GZSL model, named feature Contrastive optimization for GZSL (Co-GZSL), to reconstruct the semantic features from the embedded features, which ensures that the embedded features are close to the original semantic features indirectly by comparing reconstructed semantic features with original semantic features. In addition, to settle the problem that the synthesized features lack discrimination and semantic consistency, we introduce a Feature Contrastive Optimization Module (FCOM) and jointly utilize contrastive and semantic cycle-consistency losses in the FCOM to strengthen the intra-class compactness and the inter-class separability and to encourage the model to generate semantically consistent and discriminative visual features. By combining the generative module, the embedding module, and the FCOM, we achieve Co-GZSL. We evaluate the proposed Co-GZSL model on four benchmarks, and the experimental results indicate that our model is superior over current methods. Code is available at: https://github.com/zhanzhuxi/Co-GZSL.
引用
收藏
相关论文
共 33 条
[1]  
Akata Z(2015)Label-embedding for image classification IEEE Trans Pattern Anal Mach Intell 38 1425-1438
[2]  
Perronnin F(2022)Cross-modal propagation network for generalized zero-shot learning Pattern Recognit Lett 159 125-131
[3]  
Harchaoui Z(2021)Inference guided feature generation for generalized zero-shot learning Neurocomputing 430 150-158
[4]  
Guo T(2022)Semantic contrastive embedding for generalized zero-shot learning Int J Comput Vis 130 2606-2622
[5]  
Liang J(2023)Dual-focus transfer network for zero-shot learning Neurocomputing 541 77-83
[6]  
Liang J(2022)GAN-MVAE: a discriminative latent feature generation framework for generalized zero-shot learning Pattern Recognit Lett 155 125-131
[7]  
Han Z(2022)Cross-modal propagation network for generalized zero-shot learning Pattern Recognit Lett 159 2251-2265
[8]  
Fu Z(2018)Zero-shot learning—a comprehensive evaluation of the good, the bad and the ugly IEEE Trans Pattern Anal Mach Intell 41 1331-1345
[9]  
Li G(2023)Semantics-guided intra-category knowledge transfer for generalized zero-shot learning Int J Comput Vis 131 112-118
[10]  
Han Z(2022)Attributes learning network for generalized zero-shot learning Neural Netw 150 undefined-undefined