Generating generalized zero-shot learning based on dual-path feature enhancement

被引:0
作者
Chang, Xinyi [1 ]
Wang, Zhen [1 ,2 ]
Liu, Wenhao [1 ]
Gao, Limeng [1 ]
Yan, Bingshuai [1 ]
机构
[1] Shandong Univ Technol, Sch Comp Sci & Technol, Zibo 255000, Peoples R China
[2] Jilin Univ, Minist Educ, Key Lab Symbol Computat & Knowledge Engn, Qianjin St, Changchun 130012, Jilin, Peoples R China
基金
中国国家自然科学基金;
关键词
Generalized zero-shot learning; Variational autoencoders; Generative adversarial networks; Feature enhancement; NETWORK;
D O I
10.1007/s00530-024-01485-8
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Generalized zero-shot learning (GZSL) can classify both seen and unseen class samples, which plays a significant role in practical applications such as emerging species recognition and medical image recognition. However, most existing GZSL methods directly use the pre-trained deep model to learn the image feature. Due to the data distribution inconsistency between the GZSL dataset and the pre-training dataset, the obtained image features have an inferior performance. The distribution of different class image features is similar, which makes them difficult to distinguish. To solve this problem, we propose a dual-path feature enhancement (DPFE) model, which consists of four modules: the feature generation network (FGN), the local fine-grained feature enhancement (LFFE) module, the global coarse-grained feature enhancement (GCFE) module, and the feedback module (FM). The feature generation network can synthesize unseen class image features. We enhance the image features' discriminative and semantic relevance from both local and global perspectives. To focus on the image's local discriminative regions, the LFFE module processes the image in blocks and minimizes the semantic cycle-consistency loss to ensure that the region block features contain key classification semantic information. To prevent information loss caused by image blocking, we design the GCFE module. It ensures the consistency between the global image features and the semantic centers, thereby improving the discriminative power of the features. In addition, the feedback module feeds the discriminator network's middle layer information back to the generator network. As a result, the synthesized image features are more similar to the real features. Experimental results demonstrate that the proposed DPFE method outperforms the state-of-the-arts on four zero-shot learning benchmark datasets.
引用
收藏
页数:17
相关论文
共 39 条
[1]   Spatiotemporal visual-semantic embedding network for zero-shot action recognition [J].
An, Rongqiao ;
Miao, Zhenjiang ;
Li, Qingyu ;
Xu, Wanru ;
Zhang, Qiang .
JOURNAL OF ELECTRONIC IMAGING, 2019, 28 (02)
[2]   Preserving Semantic Relations for Zero-Shot Learning [J].
Annadani, Yashas ;
Biswas, Soma .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :7603-7612
[3]  
Chao W.-L., 2016, An empirical study and analysis of generalized zero-shot learning for object recognition in the wild
[4]   FREE: Feature Refinement for Generalized Zero-Shot Learning [J].
Chen, Shiming ;
Wang, Wenjie ;
Xia, Beihao ;
Peng, Qinmu ;
You, Xinge ;
Zheng, Feng ;
Shao, Ling .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :122-131
[5]   Entropy-Based Uncertainty Calibration for Generalized Zero-Shot Learning [J].
Chen, Zhi ;
Huang, Zi ;
Li, Jingjing ;
Zhang, Zheng .
DATABASES THEORY AND APPLICATIONS (ADC 2021), 2021, 12610 :139-151
[6]   Fine-Grained Generalized Zero-Shot Learning via Dense Attribute-Based Attention [J].
Dat Huynh ;
Elhamifar, Ehsan .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :4482-4492
[7]   Enhanced VAEGAN: a zero-shot image classification method [J].
Ding, Bo ;
Fan, Yufei ;
He, Yongjun ;
Zhao, Jing .
APPLIED INTELLIGENCE, 2023, 53 (08) :9235-9246
[8]   Improving GAN-based feature extraction for hyperspectral images classification [J].
Ding, Fanchang ;
Guo, Baofeng ;
Jia, Xiangxiang ;
Chi, Haoyu ;
Xu, Wenjie .
JOURNAL OF ELECTRONIC IMAGING, 2021, 30 (06)
[9]  
Erhan D, 2008, AAAI, P646, DOI DOI 10.5555/1620163.1620172
[10]  
Fang ZY, 2022, AAAI CONF ARTIF INTE, P6605