Generalized Zero-Shot Learning using Generated Proxy Unseen Samples and Entropy Separation

被引:9
作者
Gune, Omkar [1 ]
Banerjee, Biplab [1 ]
Chaudhuri, Subhasis [1 ]
Cuzzolin, Fabio [2 ]
机构
[1] Indian Inst Technol, Mumbai, Maharashtra, India
[2] Oxford Brookes Univ, Oxford, England
来源
MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA | 2020年
关键词
Generalized zero-shot learning; generative models;
D O I
10.1145/3394171.3413657
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The recent generative model-driven Generalized Zero-shot Learning (GZSL) techniques overcome the prevailing issue of the model bias towards the seen classes by synthesizing the visual samples of the unseen classes through leveraging the corresponding semantic prototypes. Although such approaches significantly improve the GZSL performance due to data augmentation, they violate the principal assumption of GZSL regarding the unavailability of semantic information of unseen classes during training. In this work, we propose to use a generative model (GAN) for synthesizing the visual proxy samples while strictly adhering to the standard assumptions of the GZSL. The aforementioned proxy samples are generated by exploring the early training regime of the GAN. We hypothesize that such proxy samples can effectively be used to characterize the average entropy of the label distribution of the samples from the unseen classes. Further, we train a classifier on the visual samples from the seen classes and proxy samples using entropy separation criterion such that an average entropy of the label distribution is low and high, respectively, for the visual samples from the seen classes and the proxy samples. Such entropy separation criterion generalizes well during testing where the samples from the unseen classes exhibit higher entropy than the entropy of the samples from the seen classes. Subsequently, low and high entropy samples are classified using supervised learning and ZSL rather than GZSL. We show the superiority of the proposed method by experimenting on AWA1, CUB, HMDB51, and UCF101 datasets.
引用
收藏
页码:4262 / 4270
页数:9
相关论文
共 50 条
  • [41] Transductive Learning With Prior Knowledge for Generalized Zero-Shot Action Recognition
    Su, Taiyi
    Wang, Hanli
    Qi, Qiuping
    Wang, Lei
    He, Bin
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (01) : 260 - 273
  • [42] Attribute disentanglement and re-entanglement for generalized zero-shot learning
    Zhou, Quan
    Liang, Yucuan
    Zhang, Zhenqi
    Cao, Wenming
    PATTERN RECOGNITION LETTERS, 2024, 186 : 1 - 7
  • [43] Contrastive semantic disentanglement in latent space for generalized zero-shot learning
    Fan, Wentao
    Liang, Chen
    Wang, Tian
    KNOWLEDGE-BASED SYSTEMS, 2022, 257
  • [44] Self-Assembled Generative Framework for Generalized Zero-Shot Learning
    Gao, Mengyu
    Dong, Qiulei
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2025, 34 : 914 - 924
  • [45] Residual-Prototype Generating Network for Generalized Zero-Shot Learning
    Zhang, Zeqing
    Li, Xiaofan
    Ma, Tai
    Gao, Zuodong
    Li, Cuihua
    Lin, Weiwei
    MATHEMATICS, 2022, 10 (19)
  • [46] Generalized zero-shot learning via discriminative and transferable disentangled representations
    Zhang, Chunyu
    Li, Zhanshan
    NEURAL NETWORKS, 2025, 183
  • [47] RE-GZSL: Relation Extrapolation for Generalized Zero-Shot Learning
    Wu, Yao
    Kong, Xia
    Xie, Yuan
    Qu, Yanyun
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (03) : 1973 - 1986
  • [48] A Generalized Zero-Shot Learning Framework for PolSAR Land Cover Classification
    Gui, Rong
    Xu, Xin
    Wang, Lei
    Yang, Rui
    Pu, Fangling
    REMOTE SENSING, 2018, 10 (08)
  • [49] Enhancing Domain-Invariant Parts for Generalized Zero-Shot Learning
    Zhang, Yang
    Feng, Songhe
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 6283 - 6291
  • [50] Domain-Aware Prototype Network for Generalized Zero-Shot Learning
    Hu, Yongli
    Feng, Lincong
    Jiang, Huajie
    Liu, Mengting
    Yin, Baocai
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (05) : 3180 - 3191