Using generalizability theory to investigate the variability and reliability of EFL composition scores by human raters and e-rater

被引:1
|
作者
Sari, Elif [1 ]
Han, Turgay [2 ]
机构
[1] Karadeniz Tech Univ, Trabzon, Turkey
[2] Ordu Univ, Ordu, Turkey
关键词
EFL writing assessment; generalizability theory; scoring variability; scoring reliability; automated writing evaluation (AWE); AUTOMATED WRITING EVALUATION;
D O I
10.30827/portalin.vi38.18056
中图分类号
G40 [教育学];
学科分类号
040101 ; 120403 ;
摘要
Using the generalizability theory (G-theory) as a theoretical framework, this study aimed at investigating the variability and reliability of holistic scores assigned by human raters and e-rater to the same EFL essays. Eighty argumentative essays written on two different topics by tertiary level Turkish EFL students were scored holistically by e-rater and eight human raters who received a detailed rater training. The results showed that e-rater and human raters assigned significantly different holistic scores to the same EFL essays. G-theory analyses revealed that human raters assigned considerably inconsistent scores to the same EFL essays although they were given a detailed rater training and more reliable ratings were attained when e-rater was integrated in the scoring procedure. Some implications are given for EFL writing assessment practices.
引用
收藏
页码:27 / 45
页数:19
相关论文
共 5 条