Multi-Modality Adversarial Auto-Encoder for Zero-Shot Learning

被引:3
作者
Ji, Zhong [1 ]
Dai, Guangwen [1 ]
Yu, Yunlong [2 ]
机构
[1] Tianjin Univ, Sch Elect & Informat Engn, Tianjin 300072, Peoples R China
[2] Zhejiang Univ, Coll Informat Sci & Elect Engn, Hangzhou 310027, Peoples R China
基金
中国国家自然科学基金;
关键词
Zero-shot learning; adversarial network; auto-encoder; image recognition;
D O I
10.1109/ACCESS.2019.2962298
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The existing generative Zero-Shot Learning (ZSL) methods only consider the unidirectional alignment from the class semantics to the visual features while ignoring the alignment from the visual features to the class semantics, which fails to construct the visual-semantic interactions well. In this paper, we propose to generate visual features based on an auto-encoder framework paired with multi-modality adversarial networks respectively for visual and semantic modalities to reinforce the visual-semantic interactions with a bidirectional alignment, which ensures the generated visual features to fit the real visual distribution and to be highly related to the semantics. The encoder aims at generating real-like visual features while the decoder forces both the real and the generated visual features to be more related to the class semantics. To further capture the discriminative information of the generated visual features, both the real and generated visual features are forced to be classified into the correct classes via a classification network. Experimental results on four benchmark datasets show that the proposed approach is particularly competitive on both the traditional ZSL and the generalized ZSL tasks.
引用
收藏
页码:9287 / 9295
页数:9
相关论文
共 53 条
[21]  
Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672
[22]   SELF-ATTENTION RELATION NETWORK FOR FEW-SHOT LEARNING [J].
Hui, Binyuan ;
Zhu, Pengfei ;
Hu, Qinghua ;
Wang, Qilong .
2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2019, :198-203
[23]  
Gulrajani I, 2017, ADV NEUR IN, V30
[24]   Adaptive Metric Learning For Zero-Shot Recognition [J].
Jiang, Huajie ;
Wang, Ruiping ;
Shan, Shiguang ;
Chen, Xilin .
IEEE SIGNAL PROCESSING LETTERS, 2019, 26 (09) :1270-1274
[25]   Semantic Autoencoder for Zero-Shot Learning [J].
Kodirov, Elyor ;
Xiang, Tao ;
Gong, Shaogang .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :4447-4456
[26]   Attribute-Based Classification for Zero-Shot Visual Object Categorization [J].
Lampert, Christoph H. ;
Nickisch, Hannes ;
Harmeling, Stefan .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2014, 36 (03) :453-465
[27]  
Lazaridou A, 2014, PROCEEDINGS OF THE 52ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, P1403
[28]   Discriminative Learning of Latent Features for Zero-Shot Recognition [J].
Li, Yan ;
Zhang, Junge ;
Zhang, Jianguo ;
Huang, Kaiqi .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :7463-7471
[29]  
Li YJ, 2015, PR MACH LEARN RES, V37, P1718
[30]   From Zero-shot Learning to Conventional Supervised Classification: Unseen Visual Data Synthesis [J].
Long, Yang ;
Liu, Li ;
Shao, Ling ;
Shen, Fumin ;
Ding, Guiguang ;
Han, Jungong .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6165-6174