Information bottleneck and selective noise supervision for zero-shot learning

被引:0
作者
Lei Zhou
Yang Liu
Pengcheng Zhang
Xiao Bai
Lin Gu
Jun Zhou
Yazhou Yao
Tatsuya Harada
Jin Zheng
Edwin Hancock
机构
[1] Beihang University,School of Computer Science and Engineering, State Key Laboratory of Software Development Environment, Jiangxi Research Institute
[2] RIKEN AIP,School of Information and Communication Technology
[3] The University of Tokyo,School of Computer Science and Engineering
[4] Griffith University,Department of Computer Science
[5] Nanjing University of Science and Technology,undefined
[6] University of York,undefined
来源
Machine Learning | 2023年 / 112卷
关键词
Zero-shot learning; Information bottleneck; Uncertainty estimation; Label-noise learning; Transductive ZSL;
D O I
暂无
中图分类号
学科分类号
摘要
Zero-shot learning (ZSL) aims to recognize novel classes by transferring semantic knowledge from seen classes to unseen classes. Though many ZSL methods rely on a direct mapping between the visual and the semantic space, the calibration deviation and hubness problem limit the generalization capability to unseen classes. Recently emerged generative ZSL methods generate unseen image features to transform ZSL into a supervised classification problem. However, most generative models still suffer from the seen-unseen bias problem as only seen data is used for training. To address these issues, we propose a novel bidirectional embedding based generative model with a tight visual-semantic coupling constraint. We learn a unified latent space that calibrates the embedded parametric distributions of both visual and semantic spaces. Since the embedding from high-dimensional visual features comprises much non-semantic information, the alignment of visual and semantic in latent space would inevitably be deviated. Therefore, we introduce an information bottleneck constraint to ZSL for the first time to preserve essential attribute information during the mapping. Specifically, we utilize the uncertainty estimation and the wake-sleep procedure to alleviate the feature noises and improve model abstraction capability. In addition, our method can be easily extended to the transductive ZSL setting by generating labels for unseen images. We then introduce a robust self-training loss to solve this label-noise problem. Extensive experimental results show that our method outperforms the state-of-the-art methods in different ZSL settings on most benchmark datasets.
引用
收藏
页码:2239 / 2261
页数:22
相关论文
共 37 条
[1]  
Chen LC(2017)Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs IEEE Transactions on Pattern Analysis and Machine Intelligence 40 834-848
[2]  
Papandreou G(2017)Imagenet classification with deep convolutional neural networks Communications of the ACM 60 84-90
[3]  
Kokkinos I(2013)Attribute-based classification for zero-shot visual object categorization IEEE Transactions on Pattern Analysis and Machine Intelligence 36 453-465
[4]  
Murphy K(2008)Visualizing data using t-SNE Journal of Machine Learning Research 9 2579-2605
[5]  
Yuille AL(2010)Hubs in space: Popular nearest neighbors in high-dimensional data Journal of Machine Learning Research 11 2487-2531
[6]  
Krizhevsky A(2018)Multiscale visual attention networks for object detection in VHR remote sensing images IEEE Geoscience and Remote Sensing Letters 16 310-314
[7]  
Sutskever I(2017)Zero-shot visual recognition via bidirectional latent embedding International Journal of Computer Vision 124 356-383
[8]  
Hinton GE(2019)Are anchor points really indispensable in label-noise learning? Advances in Neural Information Processing Systems 32 6838-6849
[9]  
Lampert CH(2018)Generalized cross entropy loss for training deep neural networks with noisy labels Advances in Neural Information Processing Systems 31 8778-8788
[10]  
Nickisch H(2020)Learning binary code for fast nearest subspace search Pattern Recognition 98 undefined-undefined