Zero-Shot Learning via Latent Space Encoding

被引:48
|
作者
Yu, Yunlong [1 ]
Ji, Zhong [1 ]
Guo, Jichang [1 ]
Zhang, Zhongfei [2 ]
机构
[1] Tianjin Univ, Sch Elect & Informat Engn, Tianjin 300072, Peoples R China
[2] SUNY Binghamton, Watson Sch, Dept Comp Sci, Binghamton, NY 13902 USA
基金
中国国家自然科学基金;
关键词
Encoder-decoder framework; latent space encoding (LSE); zero-shot learning (ZSL); OBJECTS;
D O I
10.1109/TCYB.2018.2850750
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Zero-shot learning (ZSL) is typically achieved by resorting to a class semantic embedding space to transfer the knowledge from the seen classes to unseen ones. Capturing the common semantic characteristics between the visual modality and the class semantic modality (e.g., attributes or word vector) is a key to the success of ZSL. In this paper, we propose a novel encoder-decoder approach, namely latent space encoding (LSE), to connect the semantic relations of different modalities. Instead of requiring a projection function to transfer information across different modalities like most previous work, LSE performs the interactions of different modalities via a feature aware latent space, which is learned in an implicit way. Specifically, different modalities are modeled separately but optimized jointly. For each modality, an encoder-decoder framework is performed to learn a feature aware latent space via jointly maximizing the recoverability of the original space from the latent space and the predictability of the latent space from the original space. To relate different modalities together, their features referring to the same concept are enforced to share the same latent codings. In this way, the common semantic characteristics of different modalities are generalized with the latent representations. Another property of the proposed approach is that it is easily extended to more modalities. Extensive experimental results on four benchmark datasets [animal with attribute, Caltech UCSID birds, aPY, and ImageNet] clearly demonstrate the superiority of the proposed approach on several ZSL tasks, including traditional ZSL, generalized ZSL, and zero-shot retrieval.
引用
收藏
页码:3755 / 3766
页数:12
相关论文
共 50 条
  • [1] Zero-Shot Learning via Joint Latent Similarity Embedding
    Zhang, Ziming
    Saligrama, Venkatesh
    2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 6034 - 6042
  • [2] Contrastive semantic disentanglement in latent space for generalized zero-shot learning
    Fan, Wentao
    Liang, Chen
    Wang, Tian
    KNOWLEDGE-BASED SYSTEMS, 2022, 257
  • [3] Contrastive semantic disentanglement in latent space for generalized zero-shot learning
    Fan, Wentao
    Liang, Chen
    Wang, Tian
    Knowledge-Based Systems, 2022, 257
  • [4] Zero-Shot Learning via Robust Latent Representation and Manifold Regularization
    Meng, Min
    Yu, Jun
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (04) : 1824 - 1836
  • [5] Salient Latent Features For Zero-shot Learning
    Pan, Zongrong
    Li, Jian
    Zhu, Anna
    PROCEEDINGS OF 2020 3RD INTERNATIONAL CONFERENCE ON ROBOT SYSTEMS AND APPLICATIONS, ICRSA2020, 2020, : 40 - 44
  • [6] Discriminative Learning of Latent Features for Zero-Shot Recognition
    Li, Yan
    Zhang, Junge
    Zhang, Jianguo
    Huang, Kaiqi
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 7463 - 7471
  • [7] Zero-shot recognition with latent visual attributes learning
    Xie, Yurui
    He, Xiaohai
    Zhang, Jing
    Luo, Xiaodong
    MULTIMEDIA TOOLS AND APPLICATIONS, 2020, 79 (37-38) : 27321 - 27335
  • [8] Open zero-shot learning via asymmetric VAE with dissimilarity space
    Zhai, Zhibo
    Li, Xiao
    Chang, Zhonghao
    INFORMATION SCIENCES, 2023, 647
  • [9] Discriminative Latent Attribute Autoencoder for Zero-Shot Learning
    Chen, Runqing
    Wu, Songsong
    Sun, Guangcheng
    PROCEEDINGS OF 2018 5TH IEEE INTERNATIONAL CONFERENCE ON CLOUD COMPUTING AND INTELLIGENCE SYSTEMS (CCIS), 2018, : 873 - 877
  • [10] Marginalized Latent Semantic Encoder for Zero-Shot Learning
    Ding, Zhengming
    Liu, Hongfu
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 6184 - 6192