Dual insurance for generalized zero-shot learning

被引:0
作者
Liang, Jiahao [1 ]
Fang, Xiaozhao [1 ]
Kang, Peipei [2 ]
Han, Na [3 ]
Li, Chuang [4 ]
机构
[1] Guangdong Univ Technol, Sch Automat, Guangzhou 510006, Peoples R China
[2] Guangdong Univ Technol, Sch Comp Sci & Technol, Guangzhou 510006, Peoples R China
[3] Guangdong Polytech Normal Univ, Sch Comp Sci, Guangzhou 510665, Peoples R China
[4] Guangxi Minzu Univ, Sch Financial Technol, Xiangsihu Coll, Nanning 530225, Peoples R China
基金
中国国家自然科学基金; 美国国家科学基金会;
关键词
Traditional zero-shot learning; Generalized zero-shot learning; Generative model; Semantic separation; GENERATIVE NETWORK; CODES;
D O I
10.1007/s13042-024-02381-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Traditional zero-shot learning aims to use the trained model to accurately classify samples from unseen classes, while for the more difficult task of generalized zero-shot learning, the trained model needs to classify samples from both seen and unseen classes into the correct classes. Because only seen class samples are available during training, generalized zero-shot learning meets great challenges in classification. Generative model is one of the good methods to solve this problem. However, the samples generated by the generative model are often of poor quality. In addition, there are semantic redundancies in the generated samples that are not conducive to classification. To solve these problems, we proposed the dual insurance model (DI-GAN) for generalized zero-shot learning in this paper, including a feature generation module and a semantic separation module. They guarantee the high quality of generated features and the good classification performance respectively. Specifically, the first insurance is based on generative adversarial network, whose generator is constrained by a clustering method to make the generated samples close to the real samples. The second insurance is based on variational autoencoder, including semantic separation, instance network and classification network. Semantic separation is designed to extract the semantically related parts which are beneficial to classification, while instance network acting on the semantically related parts is used to ensure the classification performance. Extensive experiments on four benchmark datasets show the competitiveness of the proposed DI-GAN.
引用
收藏
页码:2111 / 2125
页数:15
相关论文
共 57 条
[1]  
Akata Z, 2013, NIPS 2013 WORKSH OUT
[2]   Label-Embedding for Image Classification [J].
Akata, Zeynep ;
Perronnin, Florent ;
Harchaoui, Zaid ;
Schmid, Cordelia .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2016, 38 (07) :1425-1438
[3]  
Akata Z, 2015, PROC CVPR IEEE, P2927, DOI 10.1109/CVPR.2015.7298911
[4]   Label-Embedding for Attribute-Based Classification [J].
Akata, Zeynep ;
Perronnin, Florent ;
Harchaoui, Zaid ;
Schmid, Cordelia .
2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, :819-826
[5]   Preserving Semantic Relations for Zero-Shot Learning [J].
Annadani, Yashas ;
Biswas, Soma .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :7603-7612
[6]  
Arjovsky M, 2017, PR MACH LEARN RES, V70
[7]  
Brock Andrew, 2018, arXiv
[8]   An Empirical Study and Analysis of Generalized Zero-Shot Learning for Object Recognition in the Wild [J].
Chao, Wei-Lun ;
Changpinyo, Soravit ;
Gong, Boqing ;
Sha, Fei .
COMPUTER VISION - ECCV 2016, PT II, 2016, 9906 :52-68
[9]   FREE: Feature Refinement for Generalized Zero-Shot Learning [J].
Chen, Shiming ;
Wang, Wenjie ;
Xia, Beihao ;
Peng, Qinmu ;
You, Xinge ;
Zheng, Feng ;
Shao, Ling .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :122-131
[10]   Generalized Zero-Shot Learning Via Multi-Modal Aggregated Posterior Aligning Neural Network [J].
Chen, Xingyu ;
Li, Jin ;
Lan, Xuguang ;
Zheng, Nanning .
IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 :177-187