Comparison of an aggregate scoring method with a consensus scoring method in a measure of clinical reasoning capacity

被引:42
作者
Charlin, B
Desaulniers, M
Gagnon, R
Blouin, D
van der Vleuten, C
机构
[1] Univ Montreal, Fac Med Direct, URDESS, Montreal, PQ H3C 3J7, Canada
[2] Univ Sherbrooke, Dept Obstet & Gynecol, Sherbrooke, PQ J1K 2R1, Canada
[3] Univ Maastricht, Dept Educ Dev & Res, Maastricht, Netherlands
[4] Univ Montreal, Fac Med, Montreal, PQ H3C 3J7, Canada
关键词
D O I
10.1207/S15328015TLM1403_3
中图分类号
G40 [教育学];
学科分类号
040101 ; 120403 ;
摘要
Background: Diversity of clinical reasoning paths of thought among experts is well known. Nevertheless, in written clinical reasoning assessment, the common practice is to ask experts to reach a consensus on each item and to assess students on a unique 11 good answer" Purposes: To explore the effects of taking the variability of experts answers into account in a method of clinical reasoning assessment based on authentic tasks: the Script Concordance Test. Methods: Two different methods were used to build answer keys. The first incorporated variability among a group of experts (criterion experts) through an aggregate scoring method. The second was made with the consensus obtained from the group of criterion experts for each answer Scores obtained with the two methods by students and another group of experts (tested experts) were compared. The domain of assessment was gynecology-obstetric clinical knowledge. The sample consisted of 150 clerkship students and seven other experts (tested experts). Results: In a context of authentic tasks, experts' answers on items varied substantially. Amazingly, 59% of answers given individually by criterion group experts differed from the answer they provided when they were asked in a group to provide the "good answer" required from students. The aggregate scoring method showed several advantages and was more sensitive to detecting expertise. Conclusions: The findings suggest that, in assessment of complex performance in ill-defined situations, the usual practice of asking experts to reach a consensus on each item reduces and hinders the detection of expertise. If these results are confirmed by other researches, this practice should be reconsidered.
引用
收藏
页码:150 / 156
页数:7
相关论文
共 20 条
  • [1] [Anonymous], 1982, JUDGMENT UNCERTAINTY
  • [2] [Anonymous], 1987, ASSESS EVAL HIGH EDU, DOI DOI 10.1080/0260293870120307
  • [3] BARROWS HS, 1996, WHAT YOUR TUTOR MAY
  • [4] BARROWS HS, 1978, ANAL CLIN METHODS ME
  • [5] Measurement of clinical reflective capacity early in training as a predictor of clinical reasoning performance at the end of residency:: an experimental study on the script concordance test
    Brailovsky, C
    Charlin, B
    Beausoleil, S
    Coté, S
    Van der Vleuten, C
    [J]. MEDICAL EDUCATION, 2001, 35 (05) : 430 - 436
  • [6] CHAMBERLAND M, 1992, ANN COMMUNITY ORIENT, V5, P235
  • [7] The Script Concordance test: A tool to assess the reflective clinician
    Charlin, B
    Roy, L
    Brailovsky, C
    Goulet, F
    van der Vleuten, C
    [J]. TEACHING AND LEARNING IN MEDICINE, 2000, 12 (04) : 189 - 195
  • [8] Script questionnaires: their use for assessment of diagnostic knowledge in radiology
    Charlin, B
    Brailovsky, CA
    Brazeau-Lamontagne, L
    Samson, L
    Leduc, C
    Van der Vleuten, C
    [J]. MEDICAL TEACHER, 1998, 20 (06) : 567 - 571
  • [9] Scripts and medical diagnostic knowledge: Theory and applications for clinical reasoning instruction and research
    Charlin, BD
    Tardif, J
    Boshuizen, HPA
    [J]. ACADEMIC MEDICINE, 2000, 75 (02) : 182 - 190
  • [10] Charter M, 1998, ENVIRONM EDUC COMMUN, V3, P51