Synonym-Based Essay Generation and Augmentation for Robust Automatic Essay Scoring

被引:0
作者
Tashu, Tsegaye Misikir [1 ,2 ]
Horvath, Tomas [1 ,3 ]
机构
[1] Eotvos Lorand Univ, Fac Informat, Dept Data Sci & Engn, Telekom Innovat Labs, Pazmany Peter Setany 1-C, H-1117 Budapest, Hungary
[2] Wollo Univ, Kombolcha Inst Technol, Coll Informat, Kombolcha 208, Ethiopia
[3] Pavol Jozef Safarik Univ, Fac Sci, Inst Comp Sci, Jesenna 5, Kosice 04001, Slovakia
来源
INTELLIGENT DATA ENGINEERING AND AUTOMATED LEARNING - IDEAL 2022 | 2022年 / 13756卷
关键词
Adversarial attack; Data augmentation; Automatic essay scoring;
D O I
10.1007/978-3-031-21753-1_2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Automatic essay scoring (AES) models based on neural networks (NN) have had a lot of success. However, research has shown that NN-based AES models have robustness issues, such that the output of a model changes easily with small changes in the input. We proposed to use keyword-based lexical substitution using BERT that generates new essays (adversarial samples) which are lexically similar to the original essay to evaluate the robustness of AES models trained on the original set. In order to evaluate the proposed approach, we implemented three NN-based scoring approaches and trained the scoring models using two stages. First, we trained each model using the original data and evaluate the performance using the original test and newly generated test set to see the impact of the adversarial sample of the model. Secondly, we trained the models by augmenting the generated adversarial essay with the original data to train a robust model against synonym-based adversarial attacks. The results of our experiments showed that extracting the most important words from the essay and replacing them with lexically similar words, as well as generating adversarial samples for augmentation, can significantly improve the generalization of NN-based AES models. Our experiments also demonstrated that the proposed defense is capable of not only defending against adversarial attacks, but also of improving the performance of NN-based AES models.
引用
收藏
页码:12 / 21
页数:10
相关论文
共 26 条
[1]  
Alikaniotis D, 2016, PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, P715
[2]  
Alshemali B., 2020, P DEEP LEARN INS OUT, P20
[3]   Improving the Reliability of Deep Neural Networks in NLP: A Review [J].
Alshemali, Basemah ;
Kalita, Jugal .
KNOWLEDGE-BASED SYSTEMS, 2020, 191
[4]  
Alzantot M., 2018, P 2018 C EMPIRICAL M, P2890, DOI 10.18653/v1/D18-1316
[5]  
Bayer M., 2021, arXiv
[6]  
Belinkov Yonatan, 2018, INT C LEARNING REPRE
[7]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[8]  
Dong F., 2017, P 21 C COMP NAT LANG, P153
[9]  
Dong Fei, 2016, P 2016 C EMP METH NA, P1072, DOI DOI 10.18653/V1/D16-1115
[10]  
Ebrahimi J, 2018, PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2, P31