Applying Kane’s validity framework to a simulation based assessment of clinical competence

被引:0
作者
Walter Tavares
Ryan Brydges
Paul Myre
Jason Prpic
Linda Turner
Richard Yelle
Maud Huiskamp
机构
[1] University of Toronto/University Health Network,The Wilson Centre, Department of Medicine
[2] University of Toronto,Post
[3] Regional Municipality of York,MD Education (Post
[4] Ornge Transport Medicine,Graduate Medical Education/Continued Professional Development)
[5] Base Hospital and Clinical Affairs,Paramedic and Senior Services, Community and Health Services Department
[6] Health Sciences North Base Hospital,undefined
[7] Sunnybrook Base Hospital,undefined
来源
Advances in Health Sciences Education | 2018年 / 23卷
关键词
Assessment; Competence; OSCE; Paramedic; Simulation; Validation; Validity;
D O I
暂无
中图分类号
学科分类号
摘要
Assessment of clinical competence is complex and inference based. Trustworthy and defensible assessment processes must have favourable evidence of validity, particularly where decisions are considered high stakes. We aimed to organize, collect and interpret validity evidence for a high stakes simulation based assessment strategy for certifying paramedics, using Kane’s validity framework, which some report as challenging to implement. We describe our experience using the framework, identifying challenges, decisions points, interpretations and lessons learned. We considered data related to four inferences (scoring, generalization, extrapolation, implications) occurring during assessment and treated validity as a series of assumptions we must evaluate, resulting in several hypotheses and proposed analyses. We then interpreted our findings across the four inferences, judging if the evidence supported or refuted our proposed uses of the assessment data. Data evaluating “Scoring” included: (a) desirable tool characteristics, with acceptable inter-item correlations (b) strong item-total correlations (c) low error variance for items and raters, and (d) strong inter-rater reliability. Data evaluating “Generalizability” included: (a) a robust sampling strategy capturing the majority of relevant medical directives, skills and national competencies, and good overall and inter-station reliability. Data evaluating “Extrapolation” included: low correlations between assessment scores by dimension and clinical errors in practice. Data evaluating “Implications” included low error rates in practice. Interpreting our findings according to Kane’s framework, we suggest the evidence for scoring, generalization and implications supports use of our simulation-based paramedic assessment strategy as a certifying exam; however, the extrapolation evidence was weak, suggesting exam scores did not predict clinical error rates. Our analysis represents a worked example others can follow when using Kane’s validity framework to evaluate, and iteratively develop and refine assessment strategies.
引用
收藏
页码:323 / 338
页数:15
相关论文
共 55 条
[1]  
Brennan BL(2013)Commentary on “validating the interpretations and uses of test scores” Journal of Educational Measurement 50 74-83
[2]  
Clauser BE(2012)Validity considerations in the assessment of professionalism Advances in Health Sciences Education, Theory, and Practice 17 165-181
[3]  
Margolis MJ(2015)A contemporary approach to validity arguments: A practical guide to Kane’s framework Medical Education 49 560-575
[4]  
Holtman MC(2016)Validation of educational assessments: A primer for simulation and beyond Advances in Health Sciences Education Theory and Practice 1 31-250
[5]  
Katsufrakis PJ(2014)What counts as validity evidence? Examples and prevalence in a systematic review of simulation-based assessment Advances in Health Sciences Education Theory and Practice 19 233-171
[6]  
Hawkins RE(1989)Construct validation after thirty years Intelligence Measurement Theory and Public Policy 3 147-1175
[7]  
Cook D(2015)Constructing a validity argument for the objective structured assessment of technical skills (OSATS): A systematic review of validity evidence Advances in Health Sciences Education Theory and Practice 20 1149-732
[8]  
Brydges R(2002)Standard setting: A comparison of case-author and modified borderline-group methods in a small-scale OSCE Academic Medicine 77 729-17
[9]  
Ginsburg S(1999)Validating measures of performance Educational Measurement: Issues and Practice 20 5-17
[10]  
Hatala R(2012)Validating score interpretations and uses Language Testing 29 3-73