Improving long-tailed classification with PixDyMix: a localized pixel-level mixing method

被引:4
作者
Zeng, Wu [1 ]
Xiao, Zhengying [1 ]
机构
[1] Putian Univ, Engn Training Ctr, Putian 351100, Peoples R China
关键词
Long-tailed classification; Imbalanced learning; Image classification; Data augmentation; Adaptive weight adjustment; Pixel-level dynamic mixing; SMOTE; GAN;
D O I
10.1007/s11760-024-03382-z
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
With the continuous expansion of dataset size, the issue of long-tailed distribution has become increasingly prominent. Traditional approaches often favor head categories while neglecting the importance of tail categories. To address this limitation, this paper innovatively proposes the PDMLT (pixel-level dynamic mixing for long-tailed classification) method, the core of which lies in a pixel-level dynamic mixing image data augmentation technique called PixDyMix (pixel-level dynamic mixing). This technique intelligently adjusts mixing weights based on image cropping area, effectively preventing excessive loss of key pixel information during large-area cropping and improving the quality and label matching of newly generated samples. By generating higher-quality tail category sample images, it effectively increases the number of high-quality tail category samples, thereby enhancing the overall generalization ability of the model. Additionally, to overcome the limitations of existing resampling strategies in category weight allocation, we introduce an adaptive weight function to optimize the sampling process. This function can adaptively adjust the sampling weights of each category based on the degree of imbalance in the dataset, significantly improving the classification accuracy and stability of the model. Through comprehensive experimental validation on three standard long-tailed distribution datasets, our method demonstrates clear advantages and effectiveness.
引用
收藏
页码:7157 / 7170
页数:14
相关论文
共 30 条
[1]   MFC-GAN: Class-imbalanced dataset classification using Multiple Fake Class Generative Adversarial Network [J].
Ali-Gombe, Adamu ;
Elyan, Eyad .
NEUROCOMPUTING, 2019, 361 :212-221
[2]  
Bunkhumpornpat C, 2009, LECT NOTES ARTIF INT, V5476, P475, DOI 10.1007/978-3-642-01307-2_43
[3]  
Byrd J, 2019, PR MACH LEARN RES, V97
[4]  
Cao KD, 2019, ADV NEUR IN, V32
[5]   SMOTE: Synthetic minority over-sampling technique [J].
Chawla, Nitesh V. ;
Bowyer, Kevin W. ;
Hall, Lawrence O. ;
Kegelmeyer, W. Philip .
2002, American Association for Artificial Intelligence (16)
[6]   AREA: Adaptive Reweighting via Effective Area for Long-Tailed Classification [J].
Chen, Xiaohua ;
Zhou, Yucan ;
Wu, Dayan ;
Yang, Chule ;
Li, Bo ;
Hu, Qinghua ;
Wang, Weiping .
2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, :19220-19230
[7]   Class-Balanced Loss Based on Effective Number of Samples [J].
Cui, Yin ;
Jia, Menglin ;
Lin, Tsung-Yi ;
Song, Yang ;
Belongie, Serge .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :9260-9269
[8]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[9]  
DeVries T, 2017, Arxiv, DOI arXiv:1708.04552
[10]   Effective data generation for imbalanced learning using conditional generative adversarial networks [J].
Douzas, Georgios ;
Bacao, Fernando .
EXPERT SYSTEMS WITH APPLICATIONS, 2018, 91 :464-471