Dual Part Discovery Network for Zero-Shot Learning

被引:14
作者
Ge, Jiannan [1 ]
Xie, Hongtao [1 ]
Min, Shaobo [2 ]
Li, Pandeng [1 ]
Zhang, Yongdong [1 ]
机构
[1] Univ Sci & Technol China, Hefei, Peoples R China
[2] Tencent Data Platform, Shenzhen, Peoples R China
来源
PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022 | 2022年
关键词
Zero-shot learning; object recognition; joint embedding;
D O I
10.1145/3503161.3547889
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Zero-Shot Learning (ZSL) aims to recognize unseen classes by transferring knowledge from seen classes. Recent methods focus on learning a common semantic space to align visual and attribute information. However, they always over-relied on provided attributes and ignored the category discriminative information that contributes to accurate unseen class recognition, resulting in weak transferability. To this end, we propose a novel Dual Part Discovery Network (DPDN) that considers both attribute and category discriminative information by discovering attribute-guided parts and category-guided parts simultaneously to improve knowledge transfer. Specifically, for attribute-guided parts discovery, DPDN can localize the regions with specific attribute information and significantly bridge the gap between visual and semantic information guided by the given attributes. For category-guided parts discovery, the local parts are explored to discover other important regions that bring latent crucial details ignored by attributes, with the guidance of adaptive category prototypes. To better mine the transferable knowledge, we impose class correlations constraints to regularize the category prototypes. Finally, attribute- and category-guided parts complement each other and provide adequate discriminative subtle information for more accurate unseen class recognition. Extensive experimental results demonstrate that DPDN can discover discriminative parts and outperform state-of-the-art methods on three standard benchmarks.
引用
收藏
页码:3244 / 3252
页数:9
相关论文
共 53 条
[21]  
Jayaraman D, 2014, ADV NEUR IN, V27
[22]   Rethinking Knowledge Graph Propagation for Zero-Shot Learning [J].
Kampffmeyer, Michael ;
Chen, Yinbo ;
Liang, Xiaodan ;
Wang, Hao ;
Zhang, Yujia ;
Xing, Eric P. .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :11479-11488
[23]   Attention Models in Graphs: A Survey [J].
Lee, John Boaz ;
Rossi, Ryan A. ;
Kim, Sungchul ;
Ahmed, Nesreen K. ;
Koh, Eunyee .
ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2019, 13 (06)
[24]   Rethinking Zero-Shot Learning: A Conditional Visual Classification Perspective [J].
Li, Kai ;
Min, Martin Renqiang ;
Fu, Yun .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :3582-3591
[25]   Online Residual Quantization Via Streaming Data Correlation Preserving [J].
Li, Pandeng ;
Xie, Hongtao ;
Min, Shaobo ;
Zha, Zheng-Jun ;
Zhang, Yongdong .
IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 :981-994
[26]  
Li Pandeng, 2022, P AAAI C ART INT
[27]   Discriminative Learning of Latent Features for Zero-Shot Recognition [J].
Li, Yan ;
Zhang, Junge ;
Zhang, Jianguo ;
Huang, Kaiqi .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :7463-7471
[28]   Bidirectional Attention-Recognition Model for Fine-Grained Object Classification [J].
Liu, Chuanbin ;
Xie, Hongtao ;
Zha, Zhengjun ;
Yu, Lingyun ;
Chen, Zhineng ;
Zhang, Yongdong .
IEEE TRANSACTIONS ON MULTIMEDIA, 2020, 22 (07) :1785-1795
[29]  
Liu L, 2020, AAAI CONF ARTIF INTE, V34, P4868
[30]   Goal-Oriented Gaze Estimation for Zero-Shot Learning [J].
Liu, Yang ;
Zhou, Lei ;
Bai, Xiao ;
Huang, Yifei ;
Gu, Lin ;
Zhou, Jun ;
Harada, Tatsuya .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :3793-3802