An Automated Grader for Chinese Essay Combining Shallow and Deep Semantic Attributes

被引:10
作者
Yang, Yiqin [1 ]
Xia, Li [2 ,3 ]
Zhao, Qianchuan [1 ]
机构
[1] Tsinghua Univ, Ctr Intelligent & Networked Syst, Dept Automat, Beijing 100084, Peoples R China
[2] Sun Yat Sen Univ, Business Sch, Guangzhou 510275, Peoples R China
[3] Sun Yat Sen Univ, Guangdong Prov Key Lab Computat Sci, Guangzhou 510275, Peoples R China
来源
IEEE ACCESS | 2019年 / 7卷
基金
中国国家自然科学基金;
关键词
Automated Chinese essay evaluation; AGCE; natural language processing; semantic attributes; semantic feedback; SYSTEM;
D O I
10.1109/ACCESS.2019.2957582
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Writing is a pivotal part of the language exam, which is considered as a useful tool to accurately reflect students' language competence. As Chinese language tests become popular, manual grading becomes a heavy and expensive task for language test organizers. In the past years, there is a large volume of research about the automated English evaluation systems. Nevertheless, since the Chinese text has more complex grammar and structure, much fewer studies have been investigated on automated Chinese evaluation systems. In this paper, we propose an automated Chinese essay evaluation system called AGCE (Automated Grader for Chinese Essay), which combines shallow and deep semantic attributes of essays. We implement and train our AGCE system on a Chinese essay dataset, which is created by ourselves based on more than 1000 student essays from a Chinese primary school. Experimental results indicate that our AGCE system achieves the quadratic weighted Kappa of 0.7590 on a small dataset, which is of higher grading accuracy compared with other four popular neural network methods trained on large-scale datasets. In addition, our AGCE system can provide constructive feedback about Chinese writing, such as misspelling feedback and grammatical feedback about writers' essays, which is helpful to improve their writing capability.
引用
收藏
页码:176306 / 176316
页数:11
相关论文
共 35 条
  • [1] Attali J., 2006, The Journal of Technology, Learning and Assessment, V4, P1
  • [2] A validity-based approach to quality control and assurance of automated scoring
    Bejar, Isaac
    [J]. ASSESSMENT IN EDUCATION-PRINCIPLES POLICY & PRACTICE, 2011, 18 (03) : 319 - 341
  • [3] Bengio Y, 2001, ADV NEUR IN, V13, P932
  • [4] Burstein J, 2003, AUTOMATED ESSAY SCORING: A CROSS-DISCIPLINARY PERSPECTIVE, P113
  • [5] Burstein J., 2000, Proceedings of the COLING-2000 Workshop on using toolsets and architectures to build NLP systems, P44
  • [6] Chang-Hwan Lee, 2011, Proceedings of the 2011 IEEE 11th International Conference on Data Mining (ICDM 2011), P1146, DOI 10.1109/ICDM.2011.29
  • [7] Dikli S., 2006, The Journal of Technology, Learning and Assessment, V5, P1
  • [8] Elliot S, 2003, AUTOMATED ESSAY SCORING: A CROSS-DISCIPLINARY PERSPECTIVE, P71
  • [9] Latent semantic analysis for text-based research
    Foltz, PW
    [J]. BEHAVIOR RESEARCH METHODS INSTRUMENTS & COMPUTERS, 1996, 28 (02): : 197 - 202
  • [10] Graves A, 2012, STUD COMPUT INTELL, V385, P1, DOI [10.1162/neco.1997.9.1.1, 10.1007/978-3-642-24797-2]