Dual-stream generative adversarial networks for distributionally robust zero-shot learning

被引:22
|
作者
Liu, Huan [1 ]
Yao, Lina [2 ]
Zheng, Qinghua [1 ]
Luo, Minnan [1 ]
Zhao, Hongke [3 ]
Lyu, Yanzhang [1 ]
机构
[1] Xi An Jiao Tong Univ, Dept Comp Sci & Technol, MOEKLINNS Lab, Xian, Peoples R China
[2] Univ New South Wales, Sch Comp Sci & Engn, Sydney, NSW, Australia
[3] Tianjin Univ, Coll Management & Econ, Tianjin, Peoples R China
基金
中国国家自然科学基金;
关键词
Zero-shot classification; Feature synthesis; Inter-class discrepancy; Intra-class diversity; Semantic consistency;
D O I
10.1016/j.ins.2020.01.025
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Zero-shot learning (ZSL) in visual classification aims to recognize novel categories for which few or even no training samples are available. Through recent advances using generative adversarial networks (GANs) for cross-modal generation, several generative methods have been investigated for ZSL to classify unseen categories with synthetic samples. However, these GAN-based ZSL approaches still struggle to generate samples with semantic consistency and significant between-class discrepancy while preserving within-class diversity, which are vital to building classifiers for unseen classes. Accordingly, in this paper, we propose a robust dual-stream GAN to synthesize satisfactory samples for zero-shot visual classification. In more detail, the inter-class discrepancy is maximized by a backbone compatibility loss, which drives the center of the synthesized samples to move towards the center of real samples of the same class while moving further away from samples of different classes. Secondly, in order to preserve the intra-class diversity ignored by most extant paradigms, we propose a stochastic dispersion regularization to encourage the synthesized samples to be distributed at arbitrary points in the visual space of their categories. Finally, unlike previous methods that project visual samples back into semantic space and consequently cause an information degradation problem, we design a dual-stream generator to synthesize visual samples and reconstruct semantic embedding simultaneously, thereby ensuring semantic consistency. Our model outperforms the state-of-the-arts by 4.7% and 3.0% on average in two metrics over four real-world datasets, demonstrating its effectiveness and superiority. (C) 2020 Published by Elsevier Inc.
引用
收藏
页码:407 / 422
页数:16
相关论文
共 50 条
  • [1] Zero-Shot Learning with Joint Generative Adversarial Networks
    Zhang, Minwan
    Wang, Xiaohua
    Shi, Yueting
    Ren, Shiwei
    Wang, Weijiang
    ELECTRONICS, 2023, 12 (10)
  • [2] Dual-Stream Contrastive Learning for Compositional Zero-Shot Recognition
    Yang, Yanhua
    Pan, Rui
    Li, Xiangyu
    Yang, Xu
    Deng, Cheng
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 1909 - 1919
  • [3] Generative Dual Adversarial Network for Generalized Zero-shot Learning
    Huang, He
    Wang, Changhu
    Yu, Philip S.
    Wang, Chang-Dong
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 801 - 810
  • [4] ZeroNAS: Differentiable Generative Adversarial Networks Search for Zero-Shot Learning
    Yan, Caixia
    Chang, Xiaojun
    Li, Zhihui
    Guan, Weili
    Ge, Zongyuan
    Zhu, Lei
    Zheng, Qinghua
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (12) : 9733 - 9740
  • [5] Zero-shot counting with a dual-stream neural network model
    Thompson, Jessica A. F.
    Sheahan, Hannah
    Dumbalska, Tsvetomira
    Sandbrink, Julian D.
    Piazza, Manuela
    Summerfield, Christopher
    NEURON, 2024, 112 (24)
  • [6] Leveraging Dual Variational Autoencoders and Generative Adversarial Networks for Enhanced Multimodal Interaction in Zero-Shot Learning
    Li, Ning
    Chen, Jie
    Fu, Nanxin
    Xiao, Wenzhuo
    Ye, Tianrun
    Gao, Chunming
    Zhang, Ping
    ELECTRONICS, 2024, 13 (03)
  • [7] Generative Mixup Networks for Zero-Shot Learning
    Xu, Bingrong
    Zeng, Zhigang
    Lian, Cheng
    Ding, Zhengming
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022,
  • [8] A Generative Framework for Zero-Shot Learning with Adversarial Domain Adaptation
    Khare, Varun
    Mahajan, Divyat
    Bharadhwaj, Homanga
    Verma, Vinay Kumar
    Rai, Piyush
    2020 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2020, : 3090 - 3099
  • [9] Generative Adversarial Zero-Shot Relational Learning for Knowledge Graphs
    Qin, Pengda
    Wang, Xin
    Chen, Wenhu
    Zhang, Chunyun
    Xu, Weiran
    Wang, William Yang
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 8673 - 8680
  • [10] Gradient Matching Generative Networks for Zero-Shot Learning
    Sariyildiz, Mert Bulent
    Cinbis, Ramazan Gokberk
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 2163 - 2173