Rethinking Feature-based Knowledge Distillation for Face Recognition

被引:18
作者
Li, Jingzhi [1 ]
Guo, Zidong [1 ]
Li, Hui [1 ]
Han, Seungju [2 ]
Baek, Ji-won [2 ]
Yang, Min [1 ]
Yang, Ran [1 ]
Suh, Sungjoo [2 ]
机构
[1] Samsung R&D Inst China Xian SRCX, Xian, Peoples R China
[2] Samsung Adv Inst Technol SAIT, Suwon, South Korea
来源
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2023年
关键词
D O I
10.1109/CVPR52729.2023.01930
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the continual expansion of face datasets, feature-based distillation prevails for large-scale face recognition. In this work, we attempt to remove identity supervision in student training, to spare the GPU memory from saving massive class centers. However, this naive removal leads to inferior distillation result. We carefully inspect the performance degradation from the perspective of intrinsic dimension, and argue that the gap in intrinsic dimension, namely the intrinsic gap, is intimately connected to the infamous capacity gap problem. By constraining the teacher's search space with reverse distillation, we narrow the intrinsic gap and unleash the potential of feature-only distillation. Remarkably, the proposed reverse distillation creates universally student-friendly teacher that demonstrates outstanding student improvement. We further enhance its effectiveness by designing a student proxy to better bridge the intrinsic gap. As a result, the proposed method surpasses state-of-the-art distillation techniques with identity supervision on various face recognition benchmarks, and the improvements are consistent across different teacher-student pairs.
引用
收藏
页码:20156 / 20165
页数:10
相关论文
共 45 条
[1]   Partial FC: Training 10 Million Identities on a Single Machine [J].
An, Xiang ;
Zhu, Xuhan ;
Gao, Yuan ;
Xiao, Yang ;
Zhao, Yongle ;
Feng, Ziyong ;
Wu, Lan ;
Qin, Bin ;
Zhang, Ming ;
Zhang, Debing ;
Fu, Ying .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, :1445-1449
[2]  
[Anonymous], 2020, P IEEE CVF C COMP VI, DOI DOI 10.1109/IGARSS39084.2020.9323876
[3]  
Ansuini A, 2019, ADV NEUR IN, V32
[4]  
Ben-Baruch Emanuel, 2022, ARXIV220106945
[5]   Knowledge distillation: A good teacher is patient and consistent [J].
Beyer, Lucas ;
Zhai, Xiaohua ;
Royer, Amelie ;
Markeeva, Larisa ;
Anil, Rohan ;
Kolesnikov, Alexander .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, :10915-10924
[6]  
Chen S, 2018, CHIN CONT DECIS CONF, P428, DOI 10.1109/CCDC.2018.8407171
[7]  
Chen Yuntao, 2018, P AAAI C ART INT, V32
[8]   On the Efficacy of Knowledge Distillation [J].
Cho, Jang Hyun ;
Hariharan, Bharath .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :4793-4801
[9]   New Insights into Impacts of Toxic Metals in Aquatic Environments [J].
David, Elise ;
Cosio, Claudia .
ENVIRONMENTS, 2021, 8 (01)
[10]   ArcFace: Additive Angular Margin Loss for Deep Face Recognition [J].
Deng, Jiankang ;
Guo, Jia ;
Xue, Niannan ;
Zafeiriou, Stefanos .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4685-4694