Joint weighted knowledge distillation and multi-scale feature distillation for long-tailed recognition

被引:1
作者
He, Yiru [1 ,3 ]
Wang, Shiqian [2 ]
Yu, Junyang [1 ,3 ]
Liu, Chaoyang [4 ]
He, Xin [1 ,3 ]
Li, Han [1 ,3 ]
机构
[1] Henan Univ, Sch Software, Kaifeng 475000, Peoples R China
[2] Henan Prov Elect Power Co Chinas State Grid, Econ & Technol Res Inst, Zhengzhou 450046, Peoples R China
[3] Intelligent Data Proc Engn Res Ctr Henan Prov, Kaifeng, Peoples R China
[4] Henan Univ, Sch Comp & Informat Engn, Kaifeng, Peoples R China
关键词
Long-tailed distribution; Knowledge distillation; Multi-Scale feature extraction; Vision classification;
D O I
10.1007/s13042-023-01988-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Data in the natural open world tends to follow a long-tailed class distribution, leading deep models trained on such datasets to frequently exhibit inferior performance on the tail classes. Although existing approaches improve a model's performance on tail categories through strategies such as class rebalancing, they often sacrifice the deep features that the model has already learned. In this paper, we propose a new joint distillation framework called JWAFD (Joint weighted knowledge distillation and multi-scale feature distillation) to address the long-tailed recognition problem from the perspective of knowledge distillation. The framework comprises two effective modules. Firstly, the weighted knowledge distillation module, which uses a category prior to adjust the weights of each category. By doing so, the training process becomes more balanced across all categories. Then, the multi-scale feature distillation module, which helps to further optimize the feature representation, thus solving the problem of under-learning of features encountered in previous studies. Compared with previous studies, the proposed framework significantly improves the performance of rare classes while maintaining the performance of head classes recognition. Extensive experiments on three benchmark datasets(CIFAR-100-LT, ImageNet-LT and iNaturalist2018) have demonstrated that the proposed novel distillation framework achieves comparable performance to the state-of-the-art long-tailed recognition methods. Our code is available at: https://github.com/xiaohe6/JWAFD.
引用
收藏
页码:1647 / 1661
页数:15
相关论文
共 51 条
[1]  
Cao KD, 2019, ADV NEUR IN, V32
[2]  
Cao Y, 2016, AAAI CONF ARTIF INTE, P3457
[3]   A knowledge-guide hierarchical learning method for long-tailed image classification [J].
Chen, Qiong ;
Liu, Qingfa ;
Lin, Enlu .
NEUROCOMPUTING, 2021, 459 :408-418
[4]   Feature Space Augmentation for Long-Tailed Data [J].
Chu, Peng ;
Bian, Xiao ;
Liu, Shaopeng ;
Ling, Haibin .
COMPUTER VISION - ECCV 2020, PT XXIX, 2020, 12374 :694-710
[5]   Class-Balanced Loss Based on Effective Number of Samples [J].
Cui, Yin ;
Jia, Menglin ;
Lin, Tsung-Yi ;
Song, Yang ;
Belongie, Serge .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :9260-9269
[6]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[7]   The Pascal Visual Object Classes (VOC) Challenge [J].
Everingham, Mark ;
Van Gool, Luc ;
Williams, Christopher K. I. ;
Winn, John ;
Zisserman, Andrew .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2010, 88 (02) :303-338
[8]   Long-tailed visual recognition with deep models: A methodological survey and evaluation [J].
Fu, Yu ;
Xiang, Liuyu ;
Zahid, Yumna ;
Ding, Guiguang ;
Mei, Tao ;
Shen, Qiang ;
Han, Jungong .
NEUROCOMPUTING, 2022, 509 :290-309
[9]  
Han H., 2005, ICIC, P878, DOI DOI 10.1007/11538059_91
[10]   Global-local information based oversampling for multi-class imbalanced data [J].
Han, Mingming ;
Guo, Husheng ;
Li, Jinyan ;
Wang, Wenjian .
INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2023, 14 (06) :2071-2086