Transductive Visual-Semantic Embedding for Zero-shot Learning

被引:8
|
作者
Xu, Xing [1 ,2 ]
Shen, Fumin [1 ,2 ]
Yang, Yang [1 ,2 ]
Shao, Jie [1 ,2 ]
Huang, Zi [3 ]
机构
[1] Univ Elect Sci & Technol China, Ctr Future Media, Chengdu, Peoples R China
[2] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu, Peoples R China
[3] Univ Queensland, Sch Informat Technol & Elect Engn, Brisbane, Qld, Australia
基金
中国国家自然科学基金;
关键词
Zero-shot learning; transductive learning; matrix factorization; manifold learning;
D O I
10.1145/3078971.3078977
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Zero-shot learning (ZSL) aims to bridge the knowledge transfer via available semantic representations (e.g., attributes) between labeled source instances of seen classes and unlabelled target instances of unseen classes. Most existing ZSL approaches achieve this by learning a projection from the visual feature space to the semantic representation space based on the source instances, and directly applying it to the target instances. However, the intrinsic manifold structures residing in both semantic representations and visual features are not effectively incorporated into the learned projection function. Moreover, these methods may suffer from the inherent projection shift problem, due to the disjointness between seen and unseen classes. To overcome these drawbacks, we propose a novel framework termed transductive visual-semantic embedding (TVSE) for ZSL. In specific, TVSE first learns a latent embedding space to incorporate the manifold structures in both labeled source instances and unlabeled target instances under the transductive setting. In the learned space, each instance is viewed as a mixture of seen class scores. TVSE then effectively constructs the relational mapping between seen and unseen classes using the available semantic representations, and applies it to map the seen class scores of the target instances to their predictions of unseen classes. Extensive experiments on four benchmark datasets demonstrate that the proposed TVSE achieves competitive performance compared with the state-of-the-arts for zero-shot recognition and retrieval tasks.
引用
收藏
页码:41 / 49
页数:9
相关论文
共 50 条
  • [1] Survey of Visual-Semantic Embedding Methods for Zero-Shot Image Retrieval
    Ueki, Kazuya
    20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 628 - 634
  • [2] Spatiotemporal visual-semantic embedding network for zero-shot action recognition
    An, Rongqiao
    Miao, Zhenjiang
    Li, Qingyu
    Xu, Wanru
    Zhang, Qiang
    JOURNAL OF ELECTRONIC IMAGING, 2019, 28 (02)
  • [3] Visual-Semantic Aligned Bidirectional Network for Zero-Shot Learning
    Gao, Rui
    Hou, Xingsong
    Qin, Jie
    Shen, Yuming
    Long, Yang
    Liu, Li
    Zhang, Zhao
    Shao, Ling
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 1649 - 1664
  • [4] Zero-shot learning via visual-semantic aligned autoencoder
    Wei, Tianshu
    Huang, Jinjie
    Jin, Cong
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2023, 20 (08) : 14081 - 14095
  • [5] Hierarchical Semantic Loss and Confidence Estimator for Visual-Semantic Embedding-Based Zero-Shot Learning
    Seo, Sanghyun
    Kim, Juntae
    APPLIED SCIENCES-BASEL, 2019, 9 (15):
  • [6] Visual-Semantic Graph Matching Net for Zero-Shot Learning
    Duan, Bowen
    Chen, Shiming
    Guo, Yufei
    Xie, Guo-Sen
    Ding, Weiping
    Wang, Yisong
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [7] Transductive Unbiased Embedding for Zero-Shot Learning
    Song, Jie
    Shen, Chengchao
    Yang, Yezhou
    Liu, Yang
    Song, Mingli
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 1024 - 1033
  • [8] Scalable Zero-Shot Learning via Binary Visual-Semantic Embeddings
    Shen, Fumin
    Zhou, Xiang
    Yu, Jun
    Yang, Yang
    Liu, Li
    Shen, Heng Tao
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (07) : 3662 - 3674
  • [9] Visual-semantic consistency matching network for generalized zero-shot learning
    Zhang, Zhenqi
    Cao, Wenming
    NEUROCOMPUTING, 2023, 536 : 30 - 39
  • [10] Zero-shot learning with visual-semantic mutual reinforcement for image recognition
    Zhang, Yuhong
    Chen, Taohong
    Yu, Kui
    Hua, Xuegang
    JOURNAL OF ELECTRONIC IMAGING, 2024, 33 (05)