Self-Referenced Deep Learning

被引:5
作者
Lan, Xu [1 ]
Zhu, Xiatian [2 ]
Gong, Shaogang [1 ]
机构
[1] Queen Mary Univ London, London, England
[2] Vis Semant Ltd, London, England
来源
COMPUTER VISION - ACCV 2018, PT II | 2019年 / 11362卷
基金
“创新英国”项目;
关键词
D O I
10.1007/978-3-030-20890-5_19
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Knowledge distillation is an effective approach to transferring knowledge from a teacher neural network to a student target network for satisfying the low-memory and fast running requirements in practice use. Whilst being able to create stronger target networks compared to the vanilla non-teacher based learning strategy, this scheme needs to train additionally a large teacher model with expensive computational cost. In this work, we present a Self-Referenced Deep Learning (SRDL) strategy. Unlike both vanilla optimisation and existing knowledge distillation, SRDL distils the knowledge discovered by the in-training target model back to itself to regularise the subsequent learning procedure therefore eliminating the need for training a large teacher model. SRDL improves the model generalisation performance compared to vanilla learning and conventional knowledge distillation approaches with negligible extra computational cost. Extensive evaluations show that a variety of deep networks benefit from SRDL resulting in enhanced deployment performance on both coarse-grained object categorisation tasks (CIFAR10, CIFAR100, Tiny ImageNet, and ImageNet) and fine-grained person instance identification tasks (Market-1501).
引用
收藏
页码:284 / 300
页数:17
相关论文
共 52 条
[1]  
[Anonymous], 2014, NIPS
[2]  
[Anonymous], PERSON REIDENTIFICAT
[3]  
[Anonymous], 2014, ARXIV E PRINTS
[4]  
[Anonymous], ARXIV E PRINTS
[5]  
[Anonymous], ARXIV E PRINTS
[6]  
[Anonymous], ARXIV E PRINTS
[7]  
[Anonymous], 2018, CVPR
[8]  
[Anonymous], 2015, ICCV
[9]  
[Anonymous], 2016, CVPR
[10]  
[Anonymous], SIGKDD ACM