Distilling Cross-Task Knowledge via Relationship Matching

被引:22
作者
Ye, Han-Jia [1 ]
Lu, Su [1 ]
Zhan, De-Chuan [1 ]
机构
[1] Nanjing Univ, Nanjing, Peoples R China
来源
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020) | 2020年
基金
国家重点研发计划;
关键词
D O I
10.1109/CVPR42600.2020.01241
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The discriminative knowledge from a high-capacity deep neural network (a.k.a. the "teacher") could be distilled to facilitate the learning efficacy of a shallow counterpart (a.k.a. the "student"). This paper deals with a general scenario reusing the knowledge from a cross-task teacher two models are targeting non-overlapping label spaces. We emphasize that the comparison ability between instances acts as an essential factor threading knowledge across domains, and propose the RElationship FacIlitated Local cLassifiEr Distillation (REFILLED) approach, which decomposes the knowledge distillation flow into branches for embedding and the top-layer classifier. In particular, different from reconciling the instance-label confidence between models, REFILLED requires the teacher to reweight the hard triplets push forwarded by the student so that the similarity comparison levels between instances are matched. A local embedding-induced classifier from the teacher further supervises the student's classification confidence. REFILLED demonstrates its effectiveness when reusing cross-task models, and also achieves state-of-the-art performance on the standard knowledge distillation benchmarks. The code of the paper can be accessed at https://github.com/njulus/ReFilled.
引用
收藏
页码:12393 / 12402
页数:10
相关论文
共 70 条
[1]  
Achille A, 2018, J MACH LEARN RES, V19
[2]   Variational Information Distillation for Knowledge Transfer [J].
Ahn, Sungsoo ;
Hu, Shell Xu ;
Damianou, Andreas ;
Lawrence, Neil D. ;
Dai, Zhenwen .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :9155-9163
[3]  
[Anonymous], Knowledge distill via neuron selectivity transfer
[4]  
[Anonymous], 2018, INT C MACHINE LEARNI
[5]  
[Anonymous], 2007, P 24 INT C MACH LEAR, DOI 10.1145/1273496.1273523
[6]  
[Anonymous], 2018, NEURIPS
[7]  
[Anonymous], 2018, CORR
[8]  
[Anonymous], 2018, ICML
[9]  
[Anonymous], 2016, Deep Model Compression: Distilling Knowledge from Noisy Teachers
[10]  
[Anonymous], 2015, INT C MACH LEARN