Investigating Transformers for Automatic Short Answer Grading

被引:37
作者
Camus, Leon [1 ]
Filighera, Anna [1 ]
机构
[1] Tech Univ Darmstadt, Darmstadt, Germany
来源
ARTIFICIAL INTELLIGENCE IN EDUCATION (AIED 2020), PT II | 2020年 / 12164卷
关键词
Self-attention; Transfer learning; Short answer grading;
D O I
10.1007/978-3-030-52240-7_8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent advancements in the field of deep learning for natural language processing made it possible to use novel deep learning architectures, such as the Transformer, for increasingly complex natural language processing tasks. Combined with novel unsupervised pre-training tasks such as masked language modeling, sentence ordering or next sentence prediction, those natural language processing models became even more accurate. In this work, we experiment with fine-tuning different pretrained Transformer based architectures. We train the newest and most powerful, according to the glue benchmark, transformers on the SemEval2013 dataset. We also explore the impact of transfer learning a model fine-tuned on the MNLI dataset to the SemEval-2013 dataset on generalization and performance. We report up to 13% absolute improvement in macro-average-F1 over state-of-the-art results. We show that models trained with knowledge distillation are feasible for use in short answer grading. Furthermore, we compare multilingual models on a machinetranslated version of the SemEval-2013 dataset.
引用
收藏
页码:43 / 48
页数:6
相关论文
共 23 条
  • [1] Alikaniotis D, 2016, Arxiv, DOI arXiv:1606.04289
  • [2] [Anonymous], 2011, ACL HLT, DOI DOI 10.5555/2002472
  • [3] Barrault L, 2019, FOURTH CONFERENCE ON MACHINE TRANSLATION (WMT 2019), P1
  • [4] Conneau A, 2020, Arxiv, DOI arXiv:1911.02116
  • [5] Devlin J, 2019, Arxiv, DOI arXiv:1810.04805
  • [6] Dzikovska M.O., 2013, Second Joint Conference on Lexical and Computational Semantics (* SEM),, V2, P263
  • [7] Heilman M., 2013, P 7 INT WORKSH SEM E, P275
  • [8] Jimenez S., 2013, P 7 INT WORKSHOP SEM, V2, P280
  • [9] Kumar S, 2017, PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P2046
  • [10] Lample G, 2019, Arxiv, DOI [arXiv:1901.07291, DOI 10.48550/ARXIV.1901.07291]