Estimation of Near-Instance-Level Attribute Bottleneck for Zero-Shot Learning

被引:3
作者
Jiang, Chenyi [1 ]
Shen, Yuming [2 ]
Chen, Dubing [1 ]
Zhang, Haofeng [1 ]
Shao, Ling [3 ]
Torr, Philip H. S. [2 ]
机构
[1] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing 210094, Peoples R China
[2] Univ Oxford, Dept Engn Sci, Oxford OX1 3PJ, England
[3] Univ Chinese Acad Sci, Terminus AI Lab, Beijing 100190, Peoples R China
基金
中国国家自然科学基金;
关键词
Generalized zero-shot learning; Attribute bottleneck; Visual-semantic relationship; Near-instance-level; MODEL;
D O I
10.1007/s11263-024-02021-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Zero-Shot Learning (ZSL) involves transferring knowledge from seen classes to unseen classes by establishing connections between visual and semantic spaces. Traditional ZSL methods identify novel classes by class-level attribute vectors, which implies an information bottleneck. These approaches often use class-level attribute vectors as the fitting target during training, disregarding the individual variations within a class. Moreover, the attributes used for training lack location information and are prone to mismatch with local regions of visual features. To this end, we introduce a Near-Instance-Level Attribute Bottleneck (IAB) to alter class-level attribute vectors as well as visual features throughout the training phase to better reflect their naturalistic correspondences. Specifically, our Near-Instance-Wise Attribute Adaptation (NAA) modifies class attribute vectors to obtain multiple attribute basis vectors, generating a subspace that is more relevant to instance-level samples. Additionally, our Vision Attribute Relation Strengthening (VARS) module searches for attribute-related regions within the features, offering additional location information during the training phase. The proposed method is evaluated on four ZSL benchmarks, revealing that it is superior or competitive to the state-of-the-art methods on ZSL and the more challenging Generalized Zero-Shot Learning (GZSL) settings. Extensive experiments corroborate the sustainability of this study as one of the most potential directions for ZSL, i.e., the effectiveness of enhancing the visual-semantic relationships formed during training using a simple model structure. Code is available at: https://github.com/LanchJL/IAB-GZSL.
引用
收藏
页码:2962 / 2988
页数:27
相关论文
共 108 条
[1]  
Adler Jonas, 2018, NeurIPS
[2]  
Akata Z, 2015, PROC CVPR IEEE, P2927, DOI 10.1109/CVPR.2015.7298911
[3]   Label-Embedding for Attribute-Based Classification [J].
Akata, Zeynep ;
Perronnin, Florent ;
Harchaoui, Zaid ;
Schmid, Cordelia .
2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, :819-826
[4]  
Alemi A. A., 2017, ICLR
[5]   Adaptive Confidence Smoothing for Generalized Zero-Shot Learning [J].
Atzmon, Yuval ;
Chechik, Gal .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :11663-11672
[6]  
Brown TB, 2020, ADV NEUR IN, V33
[7]   No Adversaries to Zero-Shot Learning: Distilling an Ensemble of Gaussian Feature Generators [J].
Cavazza, Jacopo ;
Murino, Vittorio ;
Bue, Alessio Del .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (10) :12167-12178
[8]   Classifier and Exemplar Synthesis for Zero-Shot Learning [J].
Changpinyo, Soravit ;
Chao, Wei-Lun ;
Gong, Boqing ;
Sha, Fei .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2020, 128 (01) :166-201
[9]   Synthesized Classifiers for Zero-Shot Learning [J].
Changpinyo, Soravit ;
Chao, Wei-Lun ;
Gong, Boqing ;
Sha, Fei .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :5327-5336
[10]   An Empirical Study and Analysis of Generalized Zero-Shot Learning for Object Recognition in the Wild [J].
Chao, Wei-Lun ;
Changpinyo, Soravit ;
Gong, Boqing ;
Sha, Fei .
COMPUTER VISION - ECCV 2016, PT II, 2016, 9906 :52-68