Validation of educational assessments: a primer for simulation and beyond

被引:196
作者
Cook D.A. [1 ,2 ,3 ]
Hatala R. [4 ]
机构
[1] Mayo Clinic Online Learning, Mayo Clinic College of Medicine, Rochester, MN
[2] Office of Applied Scholarship and Education Science, Mayo Clinic College of Medicine, Rochester, MN
[3] Division of General Internal Medicine, Mayo Clinic College of Medicine, Mayo 17-W, 200 First Street SW, Rochester, 55905, MN
[4] Department of Medicine, University of British Columbia, Vancouver, BC
关键词
Content Evidence; Lumbar Puncture; Validation Framework; Validity Argument; Validity Evidence;
D O I
10.1186/s41077-016-0033-y
中图分类号
学科分类号
摘要
Background: Simulation plays a vital role in health professions assessment. This review provides a primer on assessment validation for educators and education researchers. We focus on simulation-based assessment of health professionals, but the principles apply broadly to other assessment approaches and topics. Key principles: Validation refers to the process of collecting validity evidence to evaluate the appropriateness of the interpretations, uses, and decisions based on assessment results. Contemporary frameworks view validity as a hypothesis, and validity evidence is collected to support or refute the validity hypothesis (i.e., that the proposed interpretations and decisions are defensible). In validation, the educator or researcher defines the proposed interpretations and decisions, identifies and prioritizes the most questionable assumptions in making these interpretations and decisions (the “interpretation-use argument”), empirically tests those assumptions using existing or newly-collected evidence, and then summarizes the evidence as a coherent “validity argument.” A framework proposed by Messick identifies potential evidence sources: content, response process, internal structure, relationships with other variables, and consequences. Another framework proposed by Kane identifies key inferences in generating useful interpretations: scoring, generalization, extrapolation, and implications/decision. We propose an eight-step approach to validation that applies to either framework: Define the construct and proposed interpretation, make explicit the intended decision(s), define the interpretation-use argument and prioritize needed validity evidence, identify candidate instruments and/or create/adapt a new instrument, appraise existing evidence and collect new evidence as needed, keep track of practical issues, formulate the validity argument, and make a judgment: does the evidence support the intended use? Conclusions: Rigorous validation first prioritizes and then empirically evaluates key assumptions in the interpretation and use of assessment scores. Validation science would be improved by more explicit articulation and prioritization of the interpretation-use argument, greater use of formal validation frameworks, and more evidence informing the consequences and implications of assessment. © 2016, The Author(s).
引用
收藏
相关论文
共 59 条
  • [1] Norcini J., Burch V., Workplace-based assessment as an educational tool: AMEE Guide No. 31, Med Teach, 29, pp. 855-871, (2007)
  • [2] Kogan J.R., Holmboe E.S., Hauer K.E., Tools for direct observation and assessment of clinical skills of medical trainees: a systematic review, JAMA, 302, pp. 1316-1326, (2009)
  • [3] Holmboe E.S., Sherbino J., Long D.M., Swing S.R., Frank J.R., The role of assessment in competency-based medical education, Med Teach, 32, pp. 676-682, (2010)
  • [4] Ziv A., Wolpe P.R., Small S.D., Glick S., Simulation-based medical education: an ethical imperative, Acad Med, 78, pp. 783-788, (2003)
  • [5] Schuwirth L.W.T., van der Vleuten C.P.M., The use of clinical simulations in assessment, Med Educ, 37, s1, pp. 65-71, (2003)
  • [6] Boulet J.R., Jeffries P.R., Hatala R.A., Korndorffer J.R., Feinstein D.M., Roche J.P., Research regarding methods of assessing learning outcomes, Simul Healthc, 6, pp. S48-S51, (2011)
  • [7] Issenberg S.B., McGaghie W.C., Hart I.R., Mayer J.W., Felner J.M., Petrusa E.R., Et al., Simulation technology for health care professional skills training and assessment, JAMA, 282, pp. 861-866, (1999)
  • [8] Amin Z., Boulet J.R., Cook D.A., Ellaway R., Fahal A., Kneebone R., Et al., Technology-enabled assessment of health professions education: consensus statement and recommendations from the Ottawa 2010 conference, Med Teach, 33, pp. 364-369, (2011)
  • [9] Cook D.A., Brydges R., Zendejas B., Hamstra S.J., Hatala R., Technology-enhanced simulation to assess health professionals: a systematic review of validity evidence, research methods, and reporting quality, Acad Med, 88, pp. 872-883, (2013)
  • [10] Kane M.T., Validation, Educational measurement, pp. 17-64, (2006)