Semantic-diversity transfer network for generalized zero-shot learning via inner disagreement based OOD detector

被引:9
|
作者
Liu, Bo [1 ,2 ]
Dong, Qiulei [1 ,3 ,4 ]
Hu, Zhanyi [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Automat, Natl Lab Pattern Recognit, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Future Technol, Beijing 100049, Peoples R China
[3] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing 100049, Peoples R China
[4] Chinese Acad Sci, Ctr Excellence Brain Sci & Intelligence Technol, Shanghai 200031, Peoples R China
基金
中国国家自然科学基金;
关键词
Zero-shot learning; Visual-semantic embedding; Out-of-distribution detection;
D O I
10.1016/j.knosys.2021.107337
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Zero-shot learning (ZSL) aims to recognize objects from unseen classes, where the key is to transfer knowledge from seen classes to unseen classes by establishing appropriate mappings between visual and semantic features. Currently, the knowledge transfer in many existing works is rather limited due to various factors: (i) the widely used visual features are global ones, and they are not completely consistent with semantic attributes; (ii) only one mapping is learned, which is not able to effectively model diverse visual-semantic relations; (iii) the bias problem in the generalized ZSL (GZSL) could not be effectively handled. In this paper, we propose two techniques to alleviate these limitations. Firstly, we propose a Semantic-diversity transfer Network (SetNet) addressing the first two limitations, where (1) a multiple-attention architecture and a diversity regularizer are proposed to learn multiple local visual features being more consistent with semantic attributes and (2) a projector ensemble which geometrically takes diverse local features as inputs is proposed to diversify visual-semantic relations. Secondly, we propose an inner disagreement based domain detection module (ID3M) for GZSL to alleviate the bias problem, which picks out unseen-class data before class-level classification. Due to the lack of unseen-class data in the training stage, ID3M employs a novel self-contained training scheme and detects out unseen-class data based on a proposed inner disagreement criterion. Experimental results on three public datasets show that the proposed SetNet with the explored ID3M achieves a significant improvement against many state-of-the-art methods. (C) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Generalized Zero-Shot Learning Based on Manifold Alignment
    Xu, Rui
    Shao, Shuai
    Liu, Baodi
    Liu, Weifeng
    2022 16TH IEEE INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING (ICSP2022), VOL 1, 2022, : 202 - 207
  • [32] Dual Prototype Contrastive Network for Generalized Zero-Shot Learning
    Jiang, Huajie
    Li, Zhengxian
    Hu, Yongli
    Yin, Baocai
    Yang, Jian
    van den Hengel, Anton
    Yang, Ming-Hsuan
    Qi, Yuankai
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (02) : 1111 - 1122
  • [33] Inductive Generalized Zero-Shot Learning with Adversarial Relation Network
    Yang, Guanyu
    Huang, Kaizhu
    Zhang, Rui
    Goulermas, John Y.
    Hussain, Amir
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2020, PT II, 2021, 12458 : 724 - 739
  • [34] Generative Dual Adversarial Network for Generalized Zero-shot Learning
    Huang, He
    Wang, Changhu
    Yu, Philip S.
    Wang, Chang-Dong
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 801 - 810
  • [35] Dual Progressive Prototype Network for Generalized Zero-Shot Learning
    Wang, Chaoqun
    Mina, Shaobo
    Chenl, Xuejin
    Sun, Xiaoyan
    Li, Houqiang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [36] Co-Representation Network for Generalized Zero-Shot Learning
    Zhang, Fei
    Shi, Guangming
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [37] Zero-Shot Visual Recognition via Semantic Attention-Based Compare Network
    Nian, Fudong
    Sheng, Yikun
    Wang, Junfeng
    Li, Teng
    IEEE ACCESS, 2020, 8 : 26002 - 26011
  • [38] Visual-Semantic Aligned Bidirectional Network for Zero-Shot Learning
    Gao, Rui
    Hou, Xingsong
    Qin, Jie
    Shen, Yuming
    Long, Yang
    Liu, Li
    Zhang, Zhao
    Shao, Ling
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 1649 - 1664
  • [39] Semantic-Visual Combination Propagation Network for Zero-Shot Learning
    Song, Wenli
    Zhang, Lei
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2022, 69 (04) : 2341 - 2345
  • [40] Zero-shot learning via visual-semantic aligned autoencoder
    Wei, Tianshu
    Huang, Jinjie
    Jin, Cong
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2023, 20 (08) : 14081 - 14095