Long-Tailed Classification Based on Coarse-Grained Leading Forest and Multi-Center Loss

被引:0
作者
Yang, Jinye [1 ]
Xu, Ji [1 ]
Wu, Di [2 ]
Tang, Jianhang [1 ]
Li, Shaobo [1 ]
Wang, Guoyin [3 ]
机构
[1] Guizhou Univ, State Key Lab Publ Big Data, Guiyang 550025, Peoples R China
[2] Southwest Univ, Coll Comp & Informat Sci, Chongqing 400715, Peoples R China
[3] Chongqing Univ Posts & Telecommun, Chongqing Key Lab Computat Intelligence, Chongqing 400065, Peoples R China
来源
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE | 2024年
基金
中国国家自然科学基金;
关键词
Forestry; Tail; Training; Representation learning; Data models; Computational modeling; Computational intelligence; Imbalanced learning; long-tailed learning; coarse-grained leading forest; invariant feature learning; multi-center loss;
D O I
10.1109/TETCI.2024.3445869
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Long-tailed (LT) classification is an unavoidable and challenging problem in the real world. Most existing long-tailed classification methods focus only on solving the class-wise imbalance while ignoring the attribute-wise imbalance. The deviation of a classification model is caused by both class-wise and attribute-wise imbalance. Due to the fact that attributes are implicit in most datasets and the combination of attributes is complex, attribute-wise imbalance is more difficult to handle. For this purpose, we proposed a novel long-tailed classification framework, aiming to build a multi-granularity classification model by means of invariant feature learning. This method first unsupervisedly constructs Coarse-Grained Leading Forest (CLF) to better characterize the distribution of attributes within a class. Depending on the distribution of attributes, one can customize suitable sampling strategies to construct different imbalanced datasets. We then introduce multi-center loss (MCL) that aims to gradually eliminate confusing attributes during feature learning process. The proposed framework does not necessarily couple to a specific LT classification model structure and can be integrated with any existing LT method as an independent component. Extensive experiments show that our approach achieves state-of-the-art performance on both existing benchmarks ImageNet-GLT and MSCOCO-GLT and can improve the performance of existing LT methods.
引用
收藏
页数:15
相关论文
共 45 条
[1]   Deep Over-sampling Framework for Classifying Imbalanced Data [J].
Ando, Shin ;
Huang, Chun Yuan .
MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2017, PT I, 2017, 10534 :770-785
[2]  
Arjovsky Martin, 2020, INVARIANT RISK MINIM
[3]   A systematic study of the class imbalance problem in convolutional neural networks [J].
Buda, Mateusz ;
Maki, Atsuto ;
Mazurowski, Maciej A. .
NEURAL NETWORKS, 2018, 106 :249-259
[4]  
Cao KD, 2019, ADV NEUR IN, V32
[5]   Real-Time Activities of Daily Living Recognition Under Long-Tailed Class Distribution [J].
Chaudhary, Atul ;
Gupta, Hari Prabhat ;
Shukla, K. K. .
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2022, 6 (04) :740-750
[6]   Randaugment: Practical automated data augmentation with a reduced search space [J].
Cubuk, Ekin D. ;
Zoph, Barret ;
Shlens, Jonathon ;
Le, Quoc, V .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, :3008-3017
[7]  
Drummond C., 2003, P ICML WORKSH LEARN, V11, P1
[8]  
ELKAN C, 2001, P 17 INT JOINT C ART, P978
[9]   A multiple resampling method for learning from imbalanced data sets [J].
Estabrooks, A ;
Jo, TH ;
Japkowicz, N .
COMPUTATIONAL INTELLIGENCE, 2004, 20 (01) :18-36
[10]   Co-teaching: Robust Training of Deep Neural Networks with Extremely Noisy Labels [J].
Han, Bo ;
Yao, Quanming ;
Yu, Xingrui ;
Niu, Gang ;
Xu, Miao ;
Hu, Weihua ;
Tsang, Ivor W. ;
Sugiyama, Masashi .
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31