From byproduct to design factor: on validating the interpretation of process indicators based on log data

被引:28
作者
Goldhammer, Frank [1 ,2 ]
Hahnel, Carolin [1 ,2 ]
Kroehne, Ulf [1 ]
Zehner, Fabian [1 ,2 ]
机构
[1] DIPF Leibniz Inst Res & Informat Educ, Rostocker Str 6, D-60323 Frankfurt, Germany
[2] Ctr Int Student Assessment ZIB, Frankfurt, Germany
关键词
Log data; Low-level feature; Process indicator; Cognitive assessment; Evidence-centered design; Validation strategies; PERFORMANCE; SPEED; TIME; VALIDITY; BEHAVIOR;
D O I
10.1186/s40536-021-00113-5
中图分类号
G40 [教育学];
学科分类号
040101 ; 120403 ;
摘要
International large-scale assessments such as PISA or PIAAC have started to provide public or scientific use files for log data; that is, events, event-related attributes and timestamps of test-takers' interactions with the assessment system. Log data and the process indicators derived from it can be used for many purposes. However, the intended uses and interpretations of process indicators require validation, which here means a theoretical and/or empirical justification that inferences about (latent) attributes of the test-taker's work process are valid. This article reviews and synthesizes measurement concepts from various areas, including the standard assessment paradigm, the continuous assessment approach, the evidence-centered design (ECD) framework, and test validation. Based on this synthesis, we address the questions of how to ensure the valid interpretation of process indicators by means of an evidence-centered design of the task situation, and how to empirically challenge the intended interpretation of process indicators by developing and implementing correlational and/or experimental validation strategies. For this purpose, we explicate the process of reasoning from log data to low-level features and process indicators as the outcome of evidence identification. In this process, contextualizing information from log data is essential in order to reduce interpretative ambiguities regarding the derived process indicators. Finally, we show that empirical validation strategies can be adapted from classical approaches investigating the nomothetic span and construct representation. Two worked examples illustrate possible validation strategies for the design phase of measurements and their empirical evaluation.
引用
收藏
页数:25
相关论文
共 73 条
  • [1] Abele S., 2019, Handbook of diagnostic classification models: Models and model extensions, applications, software packages, P461, DOI [DOI 10.1007/978-3-030-05584-422, DOI 10.1007/978-3-030-05584-4_22]
  • [2] Ajzen I, 1985, ACTION CONTROL COGNI, V2, P11, DOI [10.1007/978-3-642-69746-32, 10.1007/978-3-642-69746-3_2, DOI 10.1007/978-3-642-69746-3_2]
  • [3] Automated, Unobtrusive, Action-by-Action Assessment of Self-Regulation During Learning With an Intelligent Tutoring System
    Aleven, Vincent
    Roll, Ido
    McLaren, Bruce M.
    Koedinger, Kenneth R.
    [J]. EDUCATIONAL PSYCHOLOGIST, 2010, 45 (04) : 224 - 233
  • [4] [Anonymous], 2017, VALIDATION SCORE MEA
  • [5] [Anonymous], 2012, Journal of Educational Data Mining, DOI 10.5281/zenodo.3554643
  • [6] [Anonymous], 2014, STAND ED PSYCH TEST
  • [7] [Anonymous], 2001, EDUC MEAS-ISSUES PRA
  • [8] [Anonymous], 2013, PISA 2012 ASS AN FRA
  • [9] Behrens J.T., 2014, Learning analytics: from research to practice pp, P39
  • [10] Braun H, 2011, TEACH COLL REC, V113, P2309