A systematic review of item response theory in language assessment: Implications for the dimensionality of language ability

被引:12
作者
Min, Shangchao [1 ]
Aryadoust, Vahid [2 ]
机构
[1] Zhejiang Univ, Inst Appl Linguist, Hangzhou, Peoples R China
[2] Nanyang Technol Univ, Natl Inst Educ, Singapore, Singapore
关键词
Dimensionality; Item response theory (IRT); Language ability; Systematic review; Language assessment; SAMPLE-SIZE; TESTLET; MODEL; FIT; DEPENDENCE; PERFORMANCE; UNIDIMENSIONALITY; COMPREHENSION; INFORMATION; VALIDATION;
D O I
10.1016/j.stueduc.2020.100963
中图分类号
G40 [教育学];
学科分类号
040101 ; 120403 ;
摘要
The present study conducted a systematic review of the item response theory (IRT) literature in language assessment to investigate the conceptualization and operationalization of the dimensionality of language ability. Sixty-two IRT-based studies published between 1985 and 2020 in language assessment and educational measurement journals were first classified into two categories based on a unidimensional and multidimensional research framework, and then reviewed to examine language dimensionality from technical and substantive perspectives. It was found that 12 quantitative techniques were adopted to assess language dimensionality. Exploratory factor analysis was the primary method of dimensionality analysis in papers that had applied unidimensional IRT models, whereas the comparison modeling approach was dominant in the multidimensional framework. In addition, there was converging evidence within the two streams of research supporting the role of a number of factors such as testlets, language skills, subskills, and linguistic elements as sources of multi-dimensionality, while mixed findings were reported for the role of item formats across research streams. The assessment of reading, listening, speaking, and writing skills was grounded within both unidimensional and multidimensional framework. By contrast, vocabulary and grammar knowledge was mainly conceptualized as unidimensional. Directions for continued inquiry and application of IRT in language assessment are provided.
引用
收藏
页数:10
相关论文
共 133 条
[51]  
Engelhard G, 2018, RASCH MODELS RATER M
[52]   Is science really facing a reproducibility crisis, and do we need it to? [J].
Fanelli, Daniele .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2018, 115 (11) :2628-2631
[53]  
Ferrando P J., 2014, Handbook of Item Response Theory Modeling, P128
[54]   Investigation of Specific Learning Disability and Testing Accommodations Based Differential Item Functioning Using a Multilevel Multidimensional Mixture Item Response Theory Model [J].
Finch, W. Holmes ;
Hernandez Finch, Maria E. .
EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT, 2013, 73 (06) :973-993
[55]   Sex Differences in Item Functioning in the Comprehensive Inventory of Basic Skills-II Vocabulary Assessments [J].
French, Brian F. ;
Gotch, Chad M. .
JOURNAL OF PSYCHOEDUCATIONAL ASSESSMENT, 2013, 31 (04) :410-417
[56]   Hypothetical Use of Multidimensional Adaptive Testing for the Assessment of Student Achievement in the Programme for International Student Assessment [J].
Frey, Andreas ;
Seitz, Nicki-Nils .
EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT, 2011, 71 (03) :503-522
[57]   A Bifactor Multidimensional Item Response Theory Model for Differential Item Functioning Analysis on Testlet-Based Items [J].
Fukuhara, Hirotaka ;
Kamata, Akihito .
APPLIED PSYCHOLOGICAL MEASUREMENT, 2011, 35 (08) :604-622
[58]  
Geranpayeh A., 2007, Language Assessment Quarterly, V4, P190, DOI [https://doi.org/10.1080/15434300701375758, DOI 10.1080/15434300701375758]
[59]   FULL-INFORMATION ITEM BIFACTOR ANALYSIS [J].
GIBBONS, RD ;
HEDEKER, DR .
PSYCHOMETRIKA, 1992, 57 (03) :423-436
[60]  
Gibbons RD, 2018, CH CRC STAT SOC BEHA, P47