Improving Imbalanced Learning by Pre-finetuning with Data Augmentation

被引:0
作者
Shi, Yiwen [1 ]
ValizadehAslani, Taha [2 ]
Wang, Jing [3 ]
Ren, Ping [3 ]
Zhang, Yi [3 ]
Hu, Meng [3 ]
Zhao, Liang [3 ]
Liang, Hualou [4 ]
机构
[1] Drexel Univ, Coll Comp & Informat, Philadelphia, PA 19104 USA
[2] Drexel Univ, Coll Engn, Philadelphia, PA 19104 USA
[3] US FDA, Off Res & Stand, Off Gener Drugs, Ctr Drug Evaluat & Res, Silver Spring, MD USA
[4] Drexel Univ, Sch Biomed Engn Sci & Hlth Syst, Philadelphia, PA 19104 USA
来源
FOURTH INTERNATIONAL WORKSHOP ON LEARNING WITH IMBALANCED DOMAINS: THEORY AND APPLICATIONS, VOL 183 | 2022年 / 183卷
关键词
Finetuning; Data Augmentation; BERT; Natural Language Processing; SMOTE;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Imbalanced data is ubiquitous in the real world, where there is an uneven distribution of classes in the datasets. Such class imbalance poses a major challenge for modern deep learning, even with the typical class-balanced approaches such as re-sampling and re-weighting. In this work, we introduced a simple training strategy, namely pre-finetuning, as a new intermediate training stage in between the pretrained model and finetuning. We leveraged the idea of data augmentation to learn an initial representation that better fits the imbalanced distribution of the domain task during the pre-finetuning stage. We tested our method on manually contrived imbalanced datasets (both two-class and multi-class) and the FDA drug labeling dataset for ADME (i.e., absorption, distribution, metabolism, and excretion) classification. We found that, compared with standard single-stage training (i.e., vanilla finetuning), our method consistently attains improved model performance by large margins. Our work demonstrated that pre-finetuning is a simple, yet effective, learning strategy for imbalanced data.
引用
收藏
页码:68 / 82
页数:15
相关论文
共 50 条
  • [1] A Survey of Predictive Modeling on Im balanced Domains
    Branco, Paula
    Torgo, Luis
    Ribeiro, Rita P.
    [J]. ACM COMPUTING SURVEYS, 2016, 49 (02)
  • [2] Brown T. B., 2020, P 34 INT C NEUR INF
  • [3] A systematic study of the class imbalance problem in convolutional neural networks
    Buda, Mateusz
    Maki, Atsuto
    Mazurowski, Maciej A.
    [J]. NEURAL NETWORKS, 2018, 106 : 249 - 259
  • [4] Byrd J, 2019, PR MACH LEARN RES, V97
  • [5] Cao KD, 2019, ADV NEUR IN, V32
  • [6] SMOTE: Synthetic minority over-sampling technique
    Chawla, Nitesh V.
    Bowyer, Kevin W.
    Hall, Lawrence O.
    Kegelmeyer, W. Philip
    [J]. 2002, American Association for Artificial Intelligence (16)
  • [7] SMOTEBoost: Improving prediction of the minority class in boosting
    Chawla, NV
    Lazarevic, A
    Hall, LO
    Bowyer, KW
    [J]. KNOWLEDGE DISCOVERY IN DATABASES: PKDD 2003, PROCEEDINGS, 2003, 2838 : 107 - 119
  • [8] Chen JA, 2020, Arxiv, DOI arXiv:2004.12239
  • [9] Class-Balanced Loss Based on Effective Number of Samples
    Cui, Yin
    Jia, Menglin
    Lin, Tsung-Yi
    Song, Yang
    Belongie, Serge
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 9260 - 9269
  • [10] Dhole KD, 2022, Arxiv, DOI arXiv:2112.02721