Reinforced Counterfactual Data Augmentation for Dual Sentiment Classification

被引:0
|
作者
Chen, Hao [1 ]
Xia, Rui [1 ]
Yu, Jianfei [1 ]
机构
[1] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing, Jiangsu, Peoples R China
来源
2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021) | 2021年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Data augmentation and adversarial perturbation approaches have recently achieved promising results in solving the over-fitting problem in many natural language processing (NLP) tasks including sentiment classification. However, existing studies aimed to improve the generalization ability by augmenting the training data with synonymous examples or adding random noises to word embeddings, which cannot address the spurious association problem. In this work, we propose an end-toend reinforcement learning framework, which jointly performs counterfactual data generation and dual sentiment classification. Our approach has three characteristics: 1) the generator automatically generates massive and diverse antonymous sentences; 2) the discriminator contains a original-side sentiment predictor and an antonymous-side sentiment predictor, which jointly evaluate the quality of the generated sample and help the generator iteratively generate higher-quality antonymous samples; 3) the discriminator is directly used as the final sentiment classifier without the need to build an extra one. Extensive experiments show that our approach outperforms strong data augmentation baselines on several benchmark sentiment classification datasets. Further analysis confirms our approach's advantages in generating more diverse training samples and solving the spurious association problem in sentiment classification.
引用
收藏
页码:269 / 278
页数:10
相关论文
共 50 条
  • [41] Interactive Dual Attention Network for Text Sentiment Classification
    Zhu, Yinglin
    Zheng, Wenbin
    Tang, Hong
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2020, 2020
  • [42] DATA AUGMENTATION FOR CHEST PATHOLOGIES CLASSIFICATION
    Sirazitdinov, Ilyas
    Kholiavchenko, Maksym
    Kuleev, Ramil
    Ibragimov, Bulat
    2019 IEEE 16TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI 2019), 2019, : 1216 - 1219
  • [43] Automated Data Augmentation for Audio Classification
    Sun, Yanjie
    Xu, Kele
    Liu, Chaorun
    Dou, Yong
    Wang, Huaimin
    Ding, Bo
    Pan, Qinghua
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 2716 - 2728
  • [44] Graph Data Augmentation for Node Classification
    Wei, Ziyu
    Xiao, Xi
    Zhang, Bin
    Hu, Guangwu
    Li, Qing
    Xia, Shutao
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 4899 - 4905
  • [45] Data Augmentation with Transformers for Text Classification
    Medardo Tapia-Tellez, Jose
    Jair Escalante, Hugo
    ADVANCES IN COMPUTATIONAL INTELLIGENCE, MICAI 2020, PT II, 2020, 12469 : 247 - 259
  • [46] Fish Species Classification with Data Augmentation
    Mujtaba, Dena F.
    Mahapatra, Nihar R.
    2021 INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND COMPUTATIONAL INTELLIGENCE (CSCI 2021), 2021, : 1588 - 1593
  • [47] Data Augmentation for Infant Cry Classification
    Kachhi, Aastha
    Chaturvedi, Shreya
    Patil, Hemant A.
    Singh, Dipesh Kumar
    2022 13TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING (ISCSLP), 2022, : 433 - 437
  • [48] ADAPTIVE DATA AUGMENTATION FOR IMAGE CLASSIFICATION
    Fawzi, Alhussein
    Samulowitz, Horst
    Turaga, Deepak
    Frossard, Pascal
    2016 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2016, : 3688 - 3692
  • [49] Data Augmentation for Blind Signal Classification
    Wang, Peng
    Vindiola, Manuel
    MILCOM 2019 - 2019 IEEE MILITARY COMMUNICATIONS CONFERENCE (MILCOM), 2019,
  • [50] Data Augmentation for Heart Arrhythmia Classification
    Rezaei, Mercedeh J.
    Woodward, John R.
    Ramirez, Julia
    Munroe, Patricia
    2020 IEEE 32ND INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI), 2020, : 929 - 934