Learning discriminative and representative feature with cascade GAN for generalized zero-shot learning

被引:15
作者
Liu, Jingren [1 ]
Fu, Liyong [2 ]
Zhang, Haofeng [1 ]
Ye, Qiaolin [3 ]
Yang, Wankou [4 ]
Liu, Li [5 ]
机构
[1] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing, Peoples R China
[2] Chinese Acad Forestry, Inst Forest Resource Informat Tech, Beijing, Peoples R China
[3] Nanjing Forestry Univ, E Coll Informat Sci & Technol, Nanjing, Peoples R China
[4] Southeast Univ, Sch Automat, Nanjing, Peoples R China
[5] Incept Inst Artificial Intelligence, Abu Dhabi, U Arab Emirates
基金
中国国家自然科学基金;
关键词
Generalized zero-shot learning; Generative models; Orthogonality; Cascade GAN;
D O I
10.1016/j.knosys.2021.107780
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Zero-Shot Learning (ZSL) aims to employ seen images and their related semantics to identify unseen images through knowledge transfer. Among past numerous methods, the generative methods are more prominent and achieve better results than other methods. However, we find the input for generating samples is too monotonous, there are only semantics of each class and artificially defined noise, which makes the generated visual features non-discriminative and the classifier cannot effectively distinguish them. In order to solve this problem, we propose a novel approach with cascade Generative Adversarial Network (GAN) to generate discriminative and representative features. In this method, we define a latent space where the features from different categories are orthogonal to each other and the generator for this latent space is learned with a Wasserstein GAN. In addition, in order to make up for the deficiency that the features in this latent space cannot accurately simulate the true distribution of species, we utilize another Wasserstein GAN or Cramer GAN cascaded with the previous one to generate more discriminative and representative visual features. In this way, we can not only expand the content used as input in the generation process, but also make the final generated visual features clear and separable under the influence of latent spatial orthogonality. Extensive experiments on five benchmark datasets, i.e., AWA1, AWA2, CUB, SUN and APY, demonstrate that our proposed method can outperform most of the state-of-the-art methods on both conventional and generalized zero-shot learning settings. (C) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页数:13
相关论文
共 62 条
[1]  
Akata Z, 2015, PROC CVPR IEEE, P2927, DOI 10.1109/CVPR.2015.7298911
[2]   Label-Embedding for Attribute-Based Classification [J].
Akata, Zeynep ;
Perronnin, Florent ;
Harchaoui, Zaid ;
Schmid, Cordelia .
2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, :819-826
[3]  
[Anonymous], NeurIPS
[4]   Adaptive Confidence Smoothing for Generalized Zero-Shot Learning [J].
Atzmon, Yuval ;
Chechik, Gal .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :11663-11672
[5]  
Bellemare M. G., 2017, arXiv
[6]   Synthesized Classifiers for Zero-Shot Learning [J].
Changpinyo, Soravit ;
Chao, Wei-Lun ;
Gong, Boqing ;
Sha, Fei .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :5327-5336
[7]   An Empirical Study and Analysis of Generalized Zero-Shot Learning for Object Recognition in the Wild [J].
Chao, Wei-Lun ;
Changpinyo, Soravit ;
Gong, Boqing ;
Sha, Fei .
COMPUTER VISION - ECCV 2016, PT II, 2016, 9906 :52-68
[8]   Zero-Shot Visual Recognition using Semantics-Preserving Adversarial Embedding Networks [J].
Chen, Long ;
Zhang, Hanwang ;
Xiao, Jun ;
Liu, Wei ;
Chang, Shih-Fu .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :1043-1052
[9]  
Chen S., 2021, CVPR, P122
[10]   Fine-Grained Generalized Zero-Shot Learning via Dense Attribute-Based Attention [J].
Dat Huynh ;
Elhamifar, Ehsan .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :4482-4492