Using Differential Item Functioning to Analyze the Domain Generality of a Common Scientific Reasoning Test

被引:4
作者
Opitz, Ansgar [1 ]
Heene, Moritz [1 ]
Fischer, Frank [1 ]
机构
[1] Ludwig Maximilians Univ Munchen, Dept Psychol, Leopoldstr 13, D-80802 Munich, Germany
关键词
scientific reasoning; domain generality; Differential Item Functioning; assessment; higher education; MEASUREMENT INVARIANCE; SKILLS; RATIONALE; TREES;
D O I
10.1027/1015-5759/a000662
中图分类号
B849 [应用心理学];
学科分类号
040203 ;
摘要
A significant problem that assessments of scientific reasoning face at the level of higher education is the question of domain generality, that is, whether a test will produce biased results for students from different domains. This study applied three recently developed methods of analyzing differential item functioning (DIF) to evaluate the domain generality assumption of a common scientific reasoning test. Additionally, we evaluated the usefulness of these new, tree- and lasso-based, methods to analyze DIF and compared them with methods based on classical test theory. We gave the scientific reasoning test to 507 university students majoring in physics, biology, or medicine. All three DIF analysis methods indicated a domain bias present in about one-third of the items, mostly benefiting biology students. We did not find this bias by using methods based on classical test theory. Those methods indicated instead that all items were easier for physics students compared to biology students. Thus, the tree- and lasso-based methods provide a clear added value to test evaluation. Taken together, our analyses indicate that the scientific reasoning test is neither entirely domain-general, nor entirely domain-specific. We advise against using it in high-stakes situations involving domain comparisons.
引用
收藏
页码:251 / 260
页数:10
相关论文
共 44 条
[1]   Analyzing the many skills involved in solving complex physics problems [J].
Adams, Wendy K. ;
Wieman, Carl E. .
AMERICAN JOURNAL OF PHYSICS, 2015, 83 (05) :459-467
[2]  
Amthauer R., 2001, Der Intelligenz-Struktur-Test 2000R: Manual [The Intelligence-Structure-Test 2000R: Manual]
[3]  
[Anonymous], DIFTREE ITEM FOCUSED
[4]  
[Anonymous], 2007, PISA 2006: Science competencies for tomorrow's world: Volume 1: Analysis, DOI DOI 10.1787/9789264040014-EN
[5]   PHYSICS Learning and Scientific Reasoning [J].
Bao, Lei ;
Cai, Tianfan ;
Koenig, Kathy ;
Fang, Kai ;
Han, Jing ;
Wang, Jing ;
Liu, Qing ;
Ding, Lin ;
Cui, Lili ;
Luo, Ying ;
Wang, Yufeng ;
Li, Lieming ;
Wu, Nianle .
SCIENCE, 2009, 323 (5914) :586-587
[6]  
Boone W.J., 2014, Rasch analysis in the human sciences
[7]   When does measurement invariance matter? Commentary [J].
Borsboom, Denny .
MEDICAL CARE, 2006, 44 (11) :S176-S181
[8]   A breakdown of reliability coefficients by test type and reliability method, and the clinical implications of low reliability [J].
Charter, RA .
JOURNAL OF GENERAL PSYCHOLOGY, 2003, 130 (03) :290-304
[9]   A Chemistry Concept Reasoning Test [J].
Cloonan, Carrie A. ;
Hutchinson, John S. .
CHEMISTRY EDUCATION RESEARCH AND PRACTICE, 2011, 12 (02) :205-209
[10]   Interpreting FCI scores: Normalized gain, preinstruction scores, and scientific reasoning ability [J].
Coletta, VP ;
Phillips, JA .
AMERICAN JOURNAL OF PHYSICS, 2005, 73 (12) :1172-1182