Multi-domain feature-enhanced attribute updater for generalized zero-shot learning

被引:0
作者
Yuyan Shi [1 ]
Chenyi Jiang [1 ]
Feifan Song [1 ]
Qiaolin Ye [2 ]
Yang Long [3 ]
Haofeng Zhang [1 ]
机构
[1] School of Computer Science and Engineering, Nanjing University of Science and Technology, Jiangsu, Nanjing
[2] School of Information Science and Technology, Nanjing Forestry University, Jiangsu, Nanjing
[3] Department of Computer Science, Durham University, Durham
基金
中国国家自然科学基金;
关键词
Attribute updater; Feature learning; Generalized zero-shot learning; Image classification;
D O I
10.1007/s00521-025-11005-y
中图分类号
学科分类号
摘要
The goal of generalized zero-shot learning (GZSL) is to transfer knowledge from seen classes to unseen classes. However, a significant challenge is the single-category attributes are often inadequate to capture the intra-class variability of diverse image samples. During training, the model tends to cluster samples of seen categories based solely on semantic descriptions, overlooking subtle yet important visual differences. To address these challenges, this paper introduces an instance-level attribute updating method, termed as multi-domain feature-enhanced attribute updater (MDAU), which establishes a one-to-one alignment between visual and semantic by obtaining multiple instance-level attributes that closely correspond to the natural intra-class distribution of the image, in contrast to the many-to-one alignment typically used in traditional GZSL approaches. Specifically, we initially compute feature weights for each individual image, and then, the visual information is incorporated as conditions to facilitate the personalized updating of class attributes. We evaluate our model on three benchmark datasets, the results show that our model can achieve the state-of-the-art performance and detailed analysis also illustrates the effectiveness of each proposed module. Code is available at https://github.com/S-Silvia/MDAU. © The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2025.
引用
收藏
页码:8397 / 8414
页数:17
相关论文
共 69 条
[1]  
Atzmon Y., Chechik G., Adaptive confidence smoothing for generalized zero-shot learning, Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 11671-11680, (2019)
[2]  
Chao W.L., Changpinyo S., Gong B., Et al., An empirical study and analysis of generalized zero-shot learning for object recognition in the wild, Proceedings of the European conference on computer vision (ECCV), pp. 52-68, (2016)
[3]  
Chen L., Zhang H., Xiao J., Et al., Zero-shot visual recognition using semantics-preserving adversarial embedding networks, Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 1043-1052, (2018)
[4]  
Chen S., Hong Z., Liu Y., Et al., Transzero: attribute-guided transformer for zero-shot learning, Proceedings of the AAAI conference on artificial intelligence, pp. 330-338, (2022)
[5]  
Chen Z., Zhang P., Li J., Et al., Zero-shot learning by harnessing adversarial samples, Proceedings of the 31st ACM international conference on multimedia, pp. 4138-4146, (2023)
[6]  
Dai H., Dai B., Song L., Discriminative embeddings of latent variable models for structured data, International conference on machine learning. PMLR, pp. 2702-2711, (2016)
[7]  
Ding Z., Liu H., Marginalized latent semantic encoder for zero-shot learning, Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 6191-6199, (2019)
[8]  
Dosovitskiy A., Beyer L., Kolesnikov A., Et al., An Image is Worth 16X16 Words: Transformers for Image Recognition at Scale, (2020)
[9]  
Feng Y., Huang X., Yang P., Et al., Non-generative generalized zero-shot learning via task-correlated disentanglement and controllable samples synthesis, Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 9346-9355, (2022)
[10]  
Guo T., Liang J., Xie G.S., Group-wise interactive region learning for zero-shot recognition, Inf Sci, 642, (2023)