Explanatory Object Part Aggregation for Zero-Shot Learning

被引:4
作者
Chen, Xin [1 ]
Deng, Xiaoling [1 ]
Lan, Yubin [1 ]
Long, Yongbing [1 ]
Weng, Jian [2 ]
Liu, Zhiquan [2 ]
Tian, Qi [3 ]
机构
[1] South China Agr Univ, Natl Ctr Int Collaborat Res Precis Agr Aviat Pesti, Coll Elect Engn, Coll Artificial Intelligence,Guangdong Lab Lingnan, Guangzhou 510642, Peoples R China
[2] Jinan Univ, Coll Cyber Secur, Guangzhou 510632, Peoples R China
[3] Huawei Technol, Cloud BU, Shenzhen 518129, Peoples R China
基金
中国国家自然科学基金;
关键词
Zero-shot learning; generalized zero-shot learning; virtual connections; object part discovery; explanatory graph; part aggregation; feature distillation;
D O I
10.1109/TPAMI.2023.3325533
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Zero-shot learning (ZSL) aims to recognize objects from unseen classes only based on labeled images from seen classes. Most existing ZSL methods focus on optimizing feature spaces or generating visual features of unseen classes, both in conventional ZSL and generalized zero-shot learning (GZSL). However, since the learned feature spaces are suboptimal, there exists many virtual connections where visual features and semantic attributes are not corresponding to each other. To reduce virtual connections, in this paper, we propose to discover comprehensive and fine-grained object parts by building explanatory graphs based on convolutional feature maps, then aggregate object parts to train a part-net to obtain prediction results. Since the aggregated object parts contain comprehensive visual features for activating semantic attributes, the virtual connections can be reduced by a large extent. Since part-net aims to extract local fine-grained visual features, some attributes related to global structures are ignored. To take advantage of both local and global visual features, we design a feature distiller to distill local features into a master-net which aims to extract global features. The experimental results on AWA2, CUB, FLO, and SUN dataset demonstrate that our proposed method obviously outperforms the state-of-the-arts in both conventional ZSL and GZSL tasks.
引用
收藏
页码:851 / 868
页数:18
相关论文
共 67 条
[1]   Multi-Cue Zero-Shot Learning with Strong Supervision [J].
Akata, Zeynep ;
Malinowski, Mateusz ;
Fritz, Mario ;
Schiele, Bernt .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :59-68
[2]  
Akata Z, 2015, PROC CVPR IEEE, P2927, DOI 10.1109/CVPR.2015.7298911
[3]   Label-Embedding for Attribute-Based Classification [J].
Akata, Zeynep ;
Perronnin, Florent ;
Harchaoui, Zaid ;
Schmid, Cordelia .
2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, :819-826
[4]  
[Anonymous], 2009, P 23 INT C NEURAL IN
[5]  
[Anonymous], 2013, NeurIPS, DOI DOI 10.48550/ARXIV.1310.4546
[6]   Generating Visual Representations for Zero-Shot Classification [J].
Bucher, Maxime ;
Herbin, Stephane ;
Jurie, Frederic .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2017), 2017, :2666-2673
[7]   Synthesized Classifiers for Zero-Shot Learning [J].
Changpinyo, Soravit ;
Chao, Wei-Lun ;
Gong, Boqing ;
Sha, Fei .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :5327-5336
[8]   An Empirical Study and Analysis of Generalized Zero-Shot Learning for Object Recognition in the Wild [J].
Chao, Wei-Lun ;
Changpinyo, Soravit ;
Gong, Boqing ;
Sha, Fei .
COMPUTER VISION - ECCV 2016, PT II, 2016, 9906 :52-68
[9]   Zero-Shot Visual Recognition using Semantics-Preserving Adversarial Embedding Networks [J].
Chen, Long ;
Zhang, Hanwang ;
Xiao, Jun ;
Liu, Wei ;
Chang, Shih-Fu .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :1043-1052
[10]   MSDN: Mutually Semantic Distillation Network for Zero-Shot Learning [J].
Chen, Shiming ;
Hong, Ziming ;
Xie, Guo-Sen ;
Yang, Wenhan ;
Peng, Qinmu ;
Wang, Kai ;
Zhao, Jian ;
You, Xinge .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, :7602-7611