Automatic Task Requirements Writing Evaluation via Machine Reading Comprehension

被引:0
作者
Xu, Shiting [1 ]
Xu, Guowei [1 ]
Jia, Peilei [1 ]
Ding, Wenbiao [1 ]
Wu, Zhongqin [1 ]
Liu, Zitao [1 ]
机构
[1] TAL Educ Grp, Beijing, Peoples R China
来源
ARTIFICIAL INTELLIGENCE IN EDUCATION (AIED 2021), PT I | 2021年 / 12748卷
基金
国家重点研发计划;
关键词
Task requirements writing; Machine reading comprehension; Pre-training language model; Neural networks;
D O I
10.1007/978-3-030-78292-4_36
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Task requirements (TRs) writing is an important question type in Key English Test and Preliminary English Test. A TR writing question may include multiple requirements and a high-quality essay must respond to each requirement thoroughly and accurately. However, the limited teacher resources prevent students from getting detailed grading instantly. The majority of existing automatic essay scoring systems focus on giving a holistic score but rarely provide reasons to support it. In this paper, we proposed an end-to-end framework based on machine reading comprehension (MRC) to address this problem to some extent. The framework not only detects whether an essay responds to a requirement question, but clearly marks where the essay answers the question. Our framework consists of three modules: question normalization module, ELECTRA based MRC module and response locating module. We extensively explore state-of-the-art MRC methods. Our approach achieves 0.93 accuracy score and 0.85 F1 score on a real-world educational dataset. To encourage reproducible results, we make our code publicly available at https://github.com/aied2021TRMRC/AIED 2021 TRMRC code.
引用
收藏
页码:446 / 458
页数:13
相关论文
共 30 条
  • [1] Carlile W, 2018, PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL), VOL 1, P621
  • [2] ELECTRA: PRE-TRAINING TEXT ENCODERS AS DISCRIMINATORS RATHER THAN GENERATORS
    Clark, Kevin
    Luong, Minh-Thang
    Le, Quoc V.
    Manning, Christopher D.
    [J]. INFORMATION SYSTEMS RESEARCH, 2020,
  • [3] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [4] Gated-Attention Readers for Text Comprehension
    Dhingra, Bhuwan
    Liu, Hanxiao
    Yang, Zhilin
    Cohen, William W.
    Salakhutdinov, Ruslan
    [J]. PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 1832 - 1846
  • [5] Dong F., 2017, P 21 C COMP NAT LANG, P153, DOI DOI 10.18653/V1/K17-1017
  • [6] Hewlett D, 2016, PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, P1535
  • [7] Hochreiter S, 1997, NEURAL COMPUT, V9, P1735, DOI [10.1162/neco.1997.9.1.1, 10.1007/978-3-642-24797-2]
  • [8] JIA R., 2017, P 22 C EMPIRICAL MET, P2021, DOI [DOI 10.18653/V1/D17-1215.URL, 10.18653/v1/D17-1215, DOI 10.18653/V1/D17-1215]
  • [9] TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension
    Joshi, Mandar
    Choi, Eunsol
    Weld, Daniel S.
    Zettlemoyer, Luke
    [J]. PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 1601 - 1611
  • [10] Kadlec R, 2016, PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, P908