Leveraging Self-Distillation and Disentanglement Network to Enhance Visual-Semantic Feature Consistency in Generalized Zero-Shot Learning

被引:0
作者
Liu, Xiaoming [1 ,2 ,3 ]
Wang, Chen [1 ,2 ]
Yang, Guan [1 ,2 ]
Wang, Chunhua [4 ]
Long, Yang [5 ]
Liu, Jie [3 ,6 ]
Zhang, Zhiyuan [1 ,2 ]
机构
[1] Zhongyuan Univ Technol, Sch Comp Sci, Zhengzhou 450007, Peoples R China
[2] Zhengzhou Key Lab Text Proc & Image Understanding, Zhengzhou 450007, Peoples R China
[3] Res Ctr Language Intelligence China, Beijing 100089, Peoples R China
[4] Huanghuai Univ, Sch Animat Acad, Zhumadian 463000, Peoples R China
[5] Univ Durham, Dept Comp Sci, Durham DH1 3LE, England
[6] North China Univ Technol, Sch Informat Sci, Beijing 100144, Peoples R China
基金
中国国家自然科学基金;
关键词
generalized zero-shot learning; self-distillation; disentanglement network; visual-semantic feature consistency;
D O I
10.3390/electronics13101977
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Generalized zero-shot learning (GZSL) aims to simultaneously recognize both seen classes and unseen classes by training only on seen class samples and auxiliary semantic descriptions. Recent state-of-the-art methods infer unseen classes based on semantic information or synthesize unseen classes using generative models based on semantic information, all of which rely on the correct alignment of visual-semantic features. However, they often overlook the inconsistency between original visual features and semantic attributes. Additionally, due to the existence of cross-modal dataset biases, the visual features extracted and synthesized by the model may also mismatch with some semantic features, which could hinder the model from properly aligning visual-semantic features. To address this issue, this paper proposes a GZSL framework that enhances the consistency of visual-semantic features using a self-distillation and disentanglement network (SDDN). The aim is to utilize the self-distillation and disentanglement network to obtain semantically consistent refined visual features and non-redundant semantic features to enhance the consistency of visual-semantic features. Firstly, SDDN utilizes self-distillation technology to refine the extracted and synthesized visual features of the model. Subsequently, the visual-semantic features are then disentangled and aligned using a disentanglement network to enhance the consistency of the visual-semantic features. Finally, the consistent visual-semantic features are fused to jointly train a GZSL classifier. Extensive experiments demonstrate that the proposed method achieves more competitive results on four challenging benchmark datasets (AWA2, CUB, FLO, and SUN).
引用
收藏
页数:18
相关论文
共 46 条
  • [21] A Semantic Encoding Out-of-Distribution Classifier for Generalized Zero-Shot Learning
    Ding, Jiayu
    Hu, Xiao
    Zhong, Xiaorong
    IEEE SIGNAL PROCESSING LETTERS, 2021, 28 : 1395 - 1399
  • [22] Dual-level contrastive learning network for generalized zero-shot learning
    Jiaqi Guan
    Min Meng
    Tianyou Liang
    Jigang Liu
    Jigang Wu
    The Visual Computer, 2022, 38 : 3087 - 3095
  • [23] Dual-level contrastive learning network for generalized zero-shot learning
    Guan, Jiaqi
    Meng, Min
    Liang, Tianyou
    Liu, Jigang
    Wu, Jigang
    VISUAL COMPUTER, 2022, 38 (9-10) : 3087 - 3095
  • [24] Self-Assembled Generative Framework for Generalized Zero-Shot Learning
    Gao, Mengyu
    Dong, Qiulei
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2025, 34 : 914 - 924
  • [25] Contrastive embedding-based feature generation for generalized zero-shot learning
    Wang, Han
    Zhang, Tingting
    Zhang, Xiaoxuan
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2023, 14 (05) : 1669 - 1681
  • [26] Co-GZSL: Feature Contrastive Optimization for Generalized Zero-Shot Learning
    Qun Li
    Zhuxi Zhan
    Yaying Shen
    Bir Bhanu
    Neural Processing Letters, 56
  • [27] Residual-Prototype Generating Network for Generalized Zero-Shot Learning
    Zhang, Zeqing
    Li, Xiaofan
    Ma, Tai
    Gao, Zuodong
    Li, Cuihua
    Lin, Weiwei
    MATHEMATICS, 2022, 10 (19)
  • [28] Contrastive embedding-based feature generation for generalized zero-shot learning
    Han Wang
    Tingting Zhang
    Xiaoxuan Zhang
    International Journal of Machine Learning and Cybernetics, 2023, 14 : 1669 - 1681
  • [29] Co-GZSL: Feature Contrastive Optimization for Generalized Zero-Shot Learning
    Li, Qun
    Zhan, Zhuxi
    Shen, Yaying
    Bhanu, Bir
    NEURAL PROCESSING LETTERS, 2024, 56 (02)
  • [30] Domain-Aware Prototype Network for Generalized Zero-Shot Learning
    Hu, Yongli
    Feng, Lincong
    Jiang, Huajie
    Liu, Mengting
    Yin, Baocai
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (05) : 3180 - 3191