Two Approaches to Estimation of Classification Accuracy Rate Under Item Response Theory

被引:17
作者
Lathrop, Quinn N. [1 ]
Cheng, Ying [1 ]
机构
[1] Univ Notre Dame, Notre Dame, IN 46556 USA
关键词
classification; classification accuracy; item response theory; cut scores; empirical accuracy; CONSISTENCY; IRT;
D O I
10.1177/0146621612471888
中图分类号
O1 [数学]; C [社会科学总论];
学科分类号
03 ; 0303 ; 0701 ; 070101 ;
摘要
Within the framework of item response theory (IRT), there are two recent lines of work on the estimation of classification accuracy (CA) rate. One approach estimates CA when decisions are made based on total sum scores, the other based on latent trait estimates. The former is referred to as the Lee approach, and the latter, the Rudner approach, each after its representative contributor. In this article, the two approaches are delineated in the same framework to highlight their similarities and differences. In addition, a simulation study manipulating IRT model, sample size, test length, and cut score location was conducted. The study investigated the empirical CA that can be achieved using either the total scores or the latent trait estimates. It also evaluated the performances of the two approaches in estimating their respective empirical CAs. Results on the empirical CA suggest that when the model fits, classifications made with the latent trait estimate shall be equally or more accurate than classifications made with total score. The magnitude of difference was governed by divergence from the one-parameter logistic (1PL) model. Both Lee and Rudner approaches provided good estimates of their respective empirical CAs for every condition that was simulated. Practical implications of the simulation results are discussed.
引用
收藏
页码:226 / 241
页数:16
相关论文
共 25 条
[1]  
[Anonymous], 2011, R: A Language and Environment for Statistical Computing
[2]  
[Anonymous], 2004, TEST EQUATING SCALIN, DOI [DOI 10.1007/978-1-4757-4310-4, DOI 10.1007/978-1-4939-0317-7]
[3]  
Bergeson T., 2007, WASHINGTON LANGUAGE
[4]   THE ASYMPTOTIC POSTERIOR NORMALITY OF THE LATENT TRAIT IN AN IRT MODEL [J].
CHANG, HH ;
STOUT, W .
PSYCHOMETRIKA, 1993, 58 (01) :37-52
[5]   Estimating Classification Consistency and Accuracy for Cognitive Diagnostic Assessment [J].
Cui, Ying ;
Gierl, Mark J. ;
Chang, Hua-Hua .
JOURNAL OF EDUCATIONAL MEASUREMENT, 2012, 49 (01) :19-38
[6]   Factors Affecting the Item Parameter Estimation and Classification Accuracy of the DINA Model [J].
de la Torre, Jimmy ;
Hong, Yuan ;
Deng, Weiling .
JOURNAL OF EDUCATIONAL MEASUREMENT, 2010, 47 (02) :227-249
[7]   Estimating Classification Accuracy for Complex Decision Rules Based on Multiple Scores [J].
Douglas, Karen M. ;
Mislevy, Robert J. .
JOURNAL OF EDUCATIONAL AND BEHAVIORAL STATISTICS, 2010, 35 (03) :280-306
[8]  
Guo F., 2006, RES EVALUAT, V11, P1
[9]  
Hanson B. A., 1994, EXTENSION LORD UNPUB
[10]  
Lathrop Q., 2011, cacIRT: Classification accuracy and consistency under Item Response Theory (Version 1.1)