An error-analysis study from an EFL writing context: Human and Automated Essay Scoring Approaches

被引:14
作者
Almusharraf, Norah [1 ]
Alotaibi, Hind [2 ]
机构
[1] Prince Sultan Univ, Riyadh, Saudi Arabia
[2] King Saud Univ, Riyadh, Saudi Arabia
关键词
EFL; Writing; Correlation; Feedback; Automated essay scoring (AES); Human raters; FEEDBACK; RATERS;
D O I
10.1007/s10758-022-09592-z
中图分类号
G40 [教育学];
学科分类号
040101 ; 120403 ;
摘要
Evaluating written texts is believed to be a time-consuming process that can lack consistency and objectivity. Automated essay scoring (AES) can provide solutions to some of the limitations of human scoring. This research aimed to evaluate the performance of one AES system, Grammarly, in comparison to human raters. Both approaches' performances were analyzed quantitatively using Corder's (1974) error analysis approach to categorize the writing errors in a corpus of 197 essays written by English as a foreign language (EFL) learners. Pearson correlation coefficient and paired sample t-tests were conducted to analyze and compare errors detected by both approaches. According to the study's results, a moderate correlation between human raters and AES in terms of the total scores and the number of errors detected. Results also indicated that the total number of errors detected by AES is significantly higher than human raters and that the latter tend to give students higher scores. The findings encourage a more open attitude towards AES systems to support EFL writing teachers in assessing students' work.
引用
收藏
页码:1015 / 1031
页数:17
相关论文
共 54 条
[1]  
Al-Ahdal A., 2020, International Journal of Innovation, Creativity and Change, V12, P418
[2]  
Alrashidi O., 2015, English Language Teaching, V8, P33
[3]  
Alshakhi Abullah., 2019, English Language Teaching, V12, P176
[4]  
Attali Y., 2013, Handbook of automated essay evaluation, P181, DOI DOI 10.4324/9780203122761.CH11
[5]   A comparison of newly-trained and experienced raters on a standardized writing assessment [J].
Attali, Yigal .
LANGUAGE TESTING, 2016, 33 (01) :99-115
[6]   Scoring with the computer: Alternative procedures for improving the reliability of holistic essay scoring [J].
Attali, Yigal ;
Lewis, Will ;
Steier, Michael .
LANGUAGE TESTING, 2013, 30 (01) :125-141
[7]   Do ESL Essay Raters' Evaluation Criteria Change With Experience? A Mixed-Methods, Cross-Sectional Study [J].
Barkaoui, Khaled .
TESOL QUARTERLY, 2010, 44 (01) :31-57
[8]   Variability in ESL Essay Rating Processes: The Role of the Rating Scale and Rater Experience [J].
Barkaoui, Khaled .
LANGUAGE ASSESSMENT QUARTERLY, 2010, 7 (01) :54-74
[9]  
Bennett R.E., 1998, Educational Measurement: Issues and Practice, V17, P9, DOI [10.1111/j.1745-3992.1998.tb00631.x, 10.1002/j.2333-8504.1997.tb01734.x, DOI 10.1002/J.2333-8504.1997.TB01734.X]
[10]  
Burstein J., 2010, OXFORD HDB APPL LING, P529, DOI [10.1093/oxfordhb/9780195384253.013.0036, DOI 10.1093/OXFORDHB/9780195384253.013.0036]