Automated Assessment of Non-Native Learner Essays: Investigating the Role of Linguistic Features

被引:55
作者
Vajjala, Sowmya [1 ]
机构
[1] Iowa State Univ, Ames, IA 50011 USA
关键词
Automated writing assessment; Essay scoring; Natural language processing; Text analysis; Linguistic features; Student modeling;
D O I
10.1007/s40593-017-0142-3
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Automatic essay scoring (AES) refers to the process of scoring free text responses to given prompts, considering human grader scores as the gold standard. Writing such essays is an essential component of many language and aptitude exams. Hence, AES became an active and established area of research, and there are many proprietary systems used in real life applications today. However, not much is known about which specific linguistic features are useful for prediction and how much of this is consistent across datasets. This article addresses that by exploring the role of various linguistic features in automatic essay scoring using two publicly available datasets of non-native English essays written in test taking scenarios. The linguistic properties are modeled by encoding lexical, syntactic, discourse and error types of learner language in the feature set. Predictive models are then developed using these features on both datasets and the most predictive features are compared. While the results show that the feature set used results in good predictive models with both datasets, the question "what are the most predictive features?" has a different answer for each dataset.
引用
收藏
页码:79 / 105
页数:27
相关论文
共 55 条
[1]  
ATTALI Y, 2006, J TECHNOLOGY LEARNIN, V4
[2]   Modeling local coherence: An entity-based approach [J].
Barzilay, Regina ;
Lapata, Mirella .
COMPUTATIONAL LINGUISTICS, 2008, 34 (01) :1-34
[3]  
Blanchard D., 2013, ETSRR1324
[4]   The Eras and Trends of Automatic Short Answer Grading [J].
Burrows, Steven ;
Gurevych, Iryna ;
Stein, Benno .
INTERNATIONAL JOURNAL OF ARTIFICIAL INTELLIGENCE IN EDUCATION, 2015, 25 (01) :60-117
[5]  
Burstein J, 2003, E RATER SCORING ENGI
[6]  
Burstein J., 2010, P HUM LANG TECHN 201, P681
[7]  
Chodorow M., 2004, TETS RES REP SER, V8, DOI [10.1002/j.2333-8504.2004.tb01931.x, DOI 10.1002/J.2333-8504.2004.TB01931.X]
[8]  
Council of Europe, 2001, COMMON EUROPEAN FRAM
[9]  
Crossley S., 2013, P 26 INT FLORIDA ART, P208
[10]  
Crossley S. A., 2014, J WRITING ASSESSMENT, V7