Self-Evolution Learning for Mixup: Enhance Data Augmentation on Few-Shot Text Classification Tasks

被引:0
|
作者
Zheng, Haoqi [1 ]
Zhong, Qihuang [2 ]
Ding, Liang [3 ]
Tian, Zhiliang [1 ]
Niu, Xin [1 ]
Wang, Changjian [1 ]
Li, Dongsheng [1 ]
Tao, Dacheng [4 ]
机构
[1] Natl Univ Def Technol, Coll Comp, Changsha, Peoples R China
[2] Wuhan Univ, Sch Comp Sci, Wuhan, Peoples R China
[3] JD Explore Acad, Beijing, Peoples R China
[4] Univ Sydney, Sydney, NSW, Australia
来源
2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023) | 2023年
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Text classification tasks often encounter fewshot scenarios with limited labeled data, and addressing data scarcity is crucial. Data augmentation with mixup merges sample pairs to generate new pseudos, which can relieve the data deficiency issue in text classification. However, the quality of pseudo-samples generated by mixup exhibits significant variations. Most of the mixup methods fail to consider the varying degree of learning difficulty in different stages of training. And mixup generates new samples with one-hot labels, which encourages the model to produce a high prediction score for the correct class that is much larger than other classes, resulting in the model's over-confidence. In this paper, we propose a self-evolution learning (SE) based mixup approach for data augmentation in text classification, which can generate more adaptive and model-friendly pseudo samples for the model training. SE caters to the growth of the model learning ability and adapts to the ability when generating training samples. To alleviate the model over-confidence, we introduce an instance-specific label smoothing regularization approach, which linearly interpolates the model's output and one-hot labels of the original samples to generate new soft labels for label mixing up. Through experimental analysis, experiments show that our SE brings consistent and significant improvements upon different mixup methods. In-depth analyses demonstrate that SE enhances the model's generalization ability.
引用
收藏
页码:8964 / 8974
页数:11
相关论文
共 50 条
  • [1] MEDA: Meta-Learning with Data Augmentation for Few-Shot Text Classification
    Sun, Pengfei
    Ouyang, Yawen
    Zhang, Wenming
    Dai, Xin-yu
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 3929 - 3935
  • [2] Few-Shot Text Classification with Triplet Networks, Data Augmentation, and Curriculum Learning
    Wei, Jason
    Huang, Chengyu
    Vosoughi, Soroush
    Cheng, Yu
    Xu, Shiqi
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 5493 - 5500
  • [3] Few-shot learning for short text classification
    Leiming Yan
    Yuhui Zheng
    Jie Cao
    Multimedia Tools and Applications, 2018, 77 : 29799 - 29810
  • [4] Few-shot learning for short text classification
    Yan, Leiming
    Zheng, Yuhui
    Cao, Jie
    MULTIMEDIA TOOLS AND APPLICATIONS, 2018, 77 (22) : 29799 - 29810
  • [5] Continual Few-Shot Learning for Text Classification
    Pasunuru, Ramakanth
    Stoyanov, Veselin
    Bansal, Mohit
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 5688 - 5702
  • [6] ALP: Data Augmentation Using Lexicalized PCFGs for Few-Shot Text Classification
    Kim, Hazel H.
    Woo, Daecheol
    Oh, Seong Joon
    Cha, Jeong-Won
    Han, Yo-Sub
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10894 - 10902
  • [7] Few-shot imbalanced classification based on data augmentation
    Chao, Xuewei
    Zhang, Lixin
    MULTIMEDIA SYSTEMS, 2023, 29 (05) : 2843 - 2851
  • [8] Few-shot imbalanced classification based on data augmentation
    Xuewei Chao
    Lixin Zhang
    Multimedia Systems, 2023, 29 : 2843 - 2851
  • [9] A load classification method based on data augmentation and few-shot machine learning
    Liu, Haoran
    Li, Huaqiang
    Yu, Xueying
    Wang, Ziyao
    Chen, Yipeng
    IET RENEWABLE POWER GENERATION, 2024,
  • [10] Few-shot learning through contextual data augmentation
    Arthaud, Farid
    Bawden, Rachel
    Birch, Alexandra
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 1049 - 1062