Explanatory Object Part Aggregation for Zero-Shot Learning

被引:4
作者
Chen, Xin [1 ]
Deng, Xiaoling [1 ]
Lan, Yubin [1 ]
Long, Yongbing [1 ]
Weng, Jian [2 ]
Liu, Zhiquan [2 ]
Tian, Qi [3 ]
机构
[1] South China Agr Univ, Natl Ctr Int Collaborat Res Precis Agr Aviat Pesti, Coll Elect Engn, Coll Artificial Intelligence,Guangdong Lab Lingnan, Guangzhou 510642, Peoples R China
[2] Jinan Univ, Coll Cyber Secur, Guangzhou 510632, Peoples R China
[3] Huawei Technol, Cloud BU, Shenzhen 518129, Peoples R China
基金
中国国家自然科学基金;
关键词
Zero-shot learning; generalized zero-shot learning; virtual connections; object part discovery; explanatory graph; part aggregation; feature distillation;
D O I
10.1109/TPAMI.2023.3325533
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Zero-shot learning (ZSL) aims to recognize objects from unseen classes only based on labeled images from seen classes. Most existing ZSL methods focus on optimizing feature spaces or generating visual features of unseen classes, both in conventional ZSL and generalized zero-shot learning (GZSL). However, since the learned feature spaces are suboptimal, there exists many virtual connections where visual features and semantic attributes are not corresponding to each other. To reduce virtual connections, in this paper, we propose to discover comprehensive and fine-grained object parts by building explanatory graphs based on convolutional feature maps, then aggregate object parts to train a part-net to obtain prediction results. Since the aggregated object parts contain comprehensive visual features for activating semantic attributes, the virtual connections can be reduced by a large extent. Since part-net aims to extract local fine-grained visual features, some attributes related to global structures are ignored. To take advantage of both local and global visual features, we design a feature distiller to distill local features into a master-net which aims to extract global features. The experimental results on AWA2, CUB, FLO, and SUN dataset demonstrate that our proposed method obviously outperforms the state-of-the-arts in both conventional ZSL and GZSL tasks.
引用
收藏
页码:851 / 868
页数:18
相关论文
共 67 条
  • [1] Multi-Cue Zero-Shot Learning with Strong Supervision
    Akata, Zeynep
    Malinowski, Mateusz
    Fritz, Mario
    Schiele, Bernt
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 59 - 68
  • [2] Akata Z, 2015, PROC CVPR IEEE, P2927, DOI 10.1109/CVPR.2015.7298911
  • [3] Label-Embedding for Attribute-Based Classification
    Akata, Zeynep
    Perronnin, Florent
    Harchaoui, Zaid
    Schmid, Cordelia
    [J]. 2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, : 819 - 826
  • [4] Generating Visual Representations for Zero-Shot Classification
    Bucher, Maxime
    Herbin, Stephane
    Jurie, Frederic
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2017), 2017, : 2666 - 2673
  • [5] Synthesized Classifiers for Zero-Shot Learning
    Changpinyo, Soravit
    Chao, Wei-Lun
    Gong, Boqing
    Sha, Fei
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 5327 - 5336
  • [6] An Empirical Study and Analysis of Generalized Zero-Shot Learning for Object Recognition in the Wild
    Chao, Wei-Lun
    Changpinyo, Soravit
    Gong, Boqing
    Sha, Fei
    [J]. COMPUTER VISION - ECCV 2016, PT II, 2016, 9906 : 52 - 68
  • [7] Zero-Shot Visual Recognition using Semantics-Preserving Adversarial Embedding Networks
    Chen, Long
    Zhang, Hanwang
    Xiao, Jun
    Liu, Wei
    Chang, Shih-Fu
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 1043 - 1052
  • [8] MSDN: Mutually Semantic Distillation Network for Zero-Shot Learning
    Chen, Shiming
    Hong, Ziming
    Xie, Guo-Sen
    Yang, Wenhan
    Peng, Qinmu
    Wang, Kai
    Zhao, Jian
    You, Xinge
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 7602 - 7611
  • [9] Fine-Grained Generalized Zero-Shot Learning via Dense Attribute-Based Attention
    Dat Huynh
    Elhamifar, Ehsan
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 4482 - 4492
  • [10] Link the head to the "beak": Zero Shot Learning from Noisy Text Description at Part Precision
    Elhoseiny, Mohamed
    Zhu, Yizhe
    Zhang, Han
    Elgammal, Ahmed
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 6288 - 6297