Validating the standardized-patient assessment administered to medical students in the New York City Consortium

被引:25
作者
Swartz, MH
Colliver, JA
Bardes, CL
Charon, R
Fried, ED
Moroff, S
机构
[1] SO ILLINOIS UNIV, SCH MED, DIV STAT & RES CONSULTING, SPRINGFIELD, IL 62702 USA
[2] MT SINAI SCH MED, MORCHAND CTR CLIN COMPETENTCE, NEW YORK, NY USA
[3] CORNELL UNIV, COLL MED, NEW YORK, NY USA
[4] COLUMBIA UNIV, COLL PHYS & SURG, NEW YORK, NY USA
[5] SUNY, HLTH SCI CTR, BROOKLYN, NY USA
[6] YESHIVA UNIV, ALBERT EINSTEIN COLL MED, NEW YORK, NY 10033 USA
关键词
D O I
10.1097/00001888-199707000-00014
中图分类号
G40 [教育学];
学科分类号
040101 ; 120403 ;
摘要
Purpose. To test the criterion validity of existing standardized-patient (SP)-examination scores using global ratings by a panel of faculty-physician observers as the gold-standard criterion; to determine whether such ratings can provide a reliable gold-standard criterion to be used for validity-related research; and to encourage the use of these gold standard ratings for validation research and examination development, including scoring and standard setting, and for enhancing understanding of the clinical competence construct. Method. Five faculty physicians independently observed and rated videotaped performances of 44 students from one medical school on the seven SP cases that make up the fourth-year assessment administered at The Morchand Center of Mount Sinai School of Medicine to students in the eight member schools in the New York City Consortium. Results. The validity coefficients showed correlations between scores on the examination and the overall ratings ranging from .60 to .70. The reliability coefficients for ratings of overall examination performance reached the commonly recommended .80 level and were very close at the case lever, with interrater reliabilities generally in the .70 to .80 range. Conclusion. The results are encouraging, with validity coefficients high enough to warrant optimism about the possibility of increasing them to the recommended .80 level, based on further studies to identify those measurable performance characteristics that most reflect the gold-standard ratings. The high interrater reliabilities indicate that faculty-physician ratings of performance on SP cases and examinations can or may be able to provide a reliable gold standard for validating and refining SP assessment.
引用
收藏
页码:619 / 626
页数:8
相关论文
共 17 条
[1]  
[Anonymous], 1987, ASSESS EVAL HIGH EDU, DOI DOI 10.1080/0260293870120307
[2]   Validity of standardized-patient examination scores as an indicator of faculty observer ratings [J].
Bardes, CL ;
Colliver, JA ;
Alonso, DR ;
Swartz, MH .
ACADEMIC MEDICINE, 1996, 71 (01) :S82-S83
[3]  
Brennan R.L., 1983, Elements of generalizability theory
[4]   TECHNICAL ISSUES - TEST APPLICATION [J].
COLLIVER, JA ;
WILLIAMS, RG .
ACADEMIC MEDICINE, 1993, 68 (06) :454-460
[5]   USING A RECEIVER OPERATING CHARACTERISTIC (ROC) ANALYSIS TO SET STANDARDS FOR A STANDARDIZED-PATIENT EXAMINATION OF CLINICAL COMPETENCE [J].
COLLIVER, JA ;
BARNHART, AJ ;
MARCY, ML ;
VERHULST, SJ .
ACADEMIC MEDICINE, 1994, 69 (10) :S37-S39
[6]  
Colliver JA, 1995, ACAD MED, V70, P1062
[7]  
COLLIVER JA, 1989, TEACH LEARN MED, V1, P31
[8]   HOW ACCURATE ARE FACULTY EVALUATIONS OF CLINICAL COMPETENCE [J].
HERBERS, JE ;
NOEL, GL ;
COOPER, GS ;
PANGARO, LN ;
WEAVER, MJ ;
HARVEY, J .
JOURNAL OF GENERAL INTERNAL MEDICINE, 1989, 4 (03) :202-208
[9]   VALIDATING INTERPRETIVE ARGUMENTS FOR LICENSURE AND CERTIFICATION EXAMINATIONS [J].
KANE, MT .
EVALUATION & THE HEALTH PROFESSIONS, 1994, 17 (02) :133-159
[10]   COMPARING CHECKLISTS AND DATABASES WITH PHYSICIANS RATINGS AS MEASURES OF STUDENTS HISTORY AND PHYSICAL-EXAMINATION SKILLS [J].
MACRAE, HM ;
VU, NV ;
GRAHAM, B ;
WORDSIMS, M ;
COLLIVER, JA ;
ROBBS, RS .
ACADEMIC MEDICINE, 1995, 70 (04) :313-317