Learning Modality-Invariant Latent Representations for Generalized Zero-shot Learning

被引:25
|
作者
Li, Jingjing [1 ]
Jing, Mengmeng [1 ]
Zhu, Lei [2 ]
Ding, Zhengming [3 ]
Lu, Ke [1 ]
Yang, Yang [1 ]
机构
[1] Univ Elect Sci & Technol China, Chengdu, Peoples R China
[2] Shandong Normal Univ, Jinan, Shandong, Peoples R China
[3] Indiana Univ Purdue Univ, Indianapolis, IN 46202 USA
来源
MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA | 2020年
基金
中国国家自然科学基金;
关键词
Zero-shot learning; mutual information estimation; generalized ZSL; variational autoencoders;
D O I
10.1145/3394171.3413503
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, feature generating methods have been successfully applied to zero-shot learning (ZSL). However, most previous approaches only generate visual representations for zero-shot recognition. In fact, typical ZSL is a classic multi-modal learning protocol which consists of a visual space and a semantic space. In this paper, therefore, we present a new method which can simultaneously generate both visual representations and semantic representations so that the essential multi-modal information associated with unseen classes can be captured. Specifically, we address the most challenging issue in such a paradigm, i.e., how to handle the domain shift and thus guarantee that the learned representations are modality-invariant. To this end, we propose two strategies: 1) leveraging the mutual information between the latent visual representations and the semantic representations; 2) maximizing the entropy of the joint distribution of the two latent representations. By leveraging the two strategies, we argue that the two modalities can be well aligned. At last, extensive experiments on five widely used datasets verify that the proposed method is able to significantly outperform previous the state-of-the-arts.
引用
收藏
页码:1348 / 1356
页数:9
相关论文
共 50 条
  • [31] Practical Aspects of Zero-Shot Learning
    Saad, Elie
    Paprzycki, Marcin
    Ganzha, Maria
    COMPUTATIONAL SCIENCE, ICCS 2022, PT II, 2022, : 88 - 95
  • [32] Research progress of zero-shot learning
    Sun, Xiaohong
    Gu, Jinan
    Sun, Hongying
    APPLIED INTELLIGENCE, 2021, 51 (06) : 3600 - 3614
  • [33] Research progress of zero-shot learning
    Xiaohong Sun
    Jinan Gu
    Hongying Sun
    Applied Intelligence, 2021, 51 : 3600 - 3614
  • [34] Zero-Shot Learning With Transferred Samples
    Guo, Yuchen
    Ding, Guiguang
    Han, Jungong
    Gao, Yue
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2017, 26 (07) : 3277 - 3290
  • [35] Research and Development on Zero-Shot Learning
    Zhang L.-N.
    Zuo X.
    Liu J.-W.
    Zidonghua Xuebao/Acta Automatica Sinica, 2020, 46 (01): : 1 - 23
  • [36] Towards Open Zero-Shot Learning
    Marmoreo, Federico
    Carrazco, Julio Ivan Davila
    Cavazza, Jacopo
    Murino, Vittorio
    IMAGE ANALYSIS AND PROCESSING, ICIAP 2022, PT II, 2022, 13232 : 564 - 575
  • [37] Language-Augmented Pixel Embedding for Generalized Zero-Shot Learning
    Wang, Ziyang
    Gou, Yunhao
    Li, Jingjing
    Zhu, Lei
    Shen, Heng Tao
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (03) : 1019 - 1030
  • [38] From Classical to Generalized Zero-Shot Learning: A Simple Adaptation Process
    Le Cacheux, Yannick
    Le Borgne, Herve
    Crucianu, Michel
    MULTIMEDIA MODELING, MMM 2019, PT II, 2019, 11296 : 465 - 477
  • [39] Audio-Visual Generalized Zero-Shot Learning the Easy Way
    Mo, Shentong
    Morgado, Pedro
    COMPUTER VISION - ECCV 2024, PT LXXI, 2025, 15129 : 377 - 395
  • [40] Self-Assembled Generative Framework for Generalized Zero-Shot Learning
    Gao, Mengyu
    Dong, Qiulei
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2025, 34 : 914 - 924