A knowledge-guide hierarchical learning method for long-tailed image classification

被引:17
作者
Chen, Qiong [1 ]
Liu, Qingfa [1 ]
Lin, Enlu [1 ]
机构
[1] South China Univ Technol, Sch Comp Sci & Engn, Guangzhou 510006, Peoples R China
关键词
Imbalanced data; Long-tailed distribution; Image classification;
D O I
10.1016/j.neucom.2021.07.008
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep visual recognition methods have achieved excellent performance on artificially constructed image datasets where the data distribution is balanced. However, in real-world scenarios, data distribution is usually extremely imbalanced and exhibit a long-tailed distribution where data in each head class is more than the class in the tail. Many efficient deep learning methods fail to work normally, i.e., they perform well in the head class while poor in the tail class. In this paper, we propose a two-layer HierarchicalLearning Long-Tailed Recognition (HL-LTR) algorithm which transforms the long-tailed problem into a hierarchical classification problem by constructing a hierarchical superclass tree in which each layer corresponds to a recognition task. In the first layer of the tree, the degree of data imbalance is largely decreased. The recognition task of the second layer is the original long-tailed recognition problem. The training of HL-LTR is top-down. The knowledge learned by the first layer transfers to classes of the second layer and guides the feature learning of the second layer by using attention mechanism module and knowledge distillation method. Compared with directly solving the most difficult long-tailed recognition task, HL-LTR achieves better performance due to its progressive learning method from easy to difficult and effective knowledge transfer strategy. (C) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:408 / 418
页数:11
相关论文
共 42 条
[1]   Network of Experts for Large-Scale Image Categorization [J].
Ahmed, Karim ;
Baig, Mohammad Haris ;
Torresani, Lorenzo .
COMPUTER VISION - ECCV 2016, PT VII, 2016, 9911 :516-532
[2]  
[Anonymous], 2011, Modern hierarchical, agglomerative clustering algorithms
[3]  
[Anonymous], 2017, PROCNT
[4]  
Cao KD, 2019, ADV NEUR IN, V32
[5]   Class-Balanced Loss Based on Effective Number of Samples [J].
Cui, Yin ;
Jia, Menglin ;
Lin, Tsung-Yi ;
Song, Yang ;
Belongie, Serge .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :9260-9269
[6]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[7]  
Drummond Chris, 2003, ICML WORKSH, V11, P1
[8]   Dynamic Few-Shot Visual Learning without Forgetting [J].
Gidaris, Spyros ;
Komodakis, Nikos .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :4367-4375
[9]  
Hinton G., 2015, NIPS DEEP LEARN REPR, P38, DOI DOI 10.48550/ARXIV.1503.02531
[10]  
Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/CVPR.2018.00745, 10.1109/TPAMI.2019.2913372]