A SYSTEMATIC REVIEW OF ASSESSMENT IN MEDICAL EDUCATION Christie Palladino, MD, MSc; Rachel Whitaker, ScM; Kirsten Wildermuth*, BS; Ruth-Marie Fincher, MD; Lara Stepleman, PhD Education Discovery Institute Medical College of Georgia at Georgia Health Sciences University, Augusta, GA Background Method, continued Results, continued • Commonly cited barriers to highly-effective medical education research include insufficient use of validated instruments and difficulty accessing and culling information about validated instruments. • We identified assessments as PEARLS (Psychometrically Evidenced, Appearing Repeatedly in a Literature Search) if they met the following criteria: 1. Cited at least twice in the literature and at least once in the medical education literature in the last ten years; and 2. Had reported psychometric data. Objectives • To conduct a systematic review of assessment measures used in medical education to: • Describe characteristics of the assessment measures which are used; and • Identify gaps in current medical education assessment. Method • We performed a two-phased systematic literature search, developed in consultation with a research librarian.† 1. Review of the five most recent years of publications from four key medical education journals (Academic Medicine, Medical Education, Teaching and Learning in Medicine, Medical Teacher); and 2. MeSH (Medical Subject Heading) term and keyword search of articles from the last ten years in PubMed. Results • We identified 1,184 English-language studies reporting on 1,008 explicit assessment tools: • 36.4% of the studies reported psychometric data. • The most common types of assessment tools were surveys (57.7%) & clinical observations (16.0%). • Almost half of the assessment tools were unnamed (45.5%). • 287 of the 1,008 assessment tools (28.5%) met the criteria for PEARLS (Figures 1 & 2) Total Measures n=1,008 *Ms. Wildermuth is pursuing a Masters of Psychology at Augusta State University. †We would like to acknowledge Kathy Davies, MLS for her contributions to this work. PEARLS Single instance with no reported psychometric data n=471 28.5% Single instance with reported psychometric data n=231 • We included English-language studies that reported an explicit tool or method used to assess medical students or residents. • Information abstracted from articles included target population, assessment source, assessment method, reported psychometric data, and country. Single Instance/ No Psychometrics Multiple use with no reported psychometric data n=18 PEARLS n=287 Figure 1 46.8% 1.8% Multiple Use/ No Psychometrics 22.9% Single Instance/ + Psychometrics Figure 2 PEARLS (n=287) Country US Non-US Multi-Country Assessment Method Survey Clinical Observation (e.g. OSCE) Simulation Oral (e.g. interview) Other/unclear Assessor Role Self-assessment Unnamed Measure 92 (32.1%) 194 (67.6%) 1 (0.3%) 162 (56.4%) 45 (15.7%) 3 (1.0%) 6 (2.1%) 71 (24.7%) 172 (59.9%) 70 (24.4%) Conclusions • Our results demonstrate that almost half of published medical education assessment measures are locallydeveloped and without reported psychometric properties. • This limits our ability to use consistent methods of assessment across institutions and makes identification of appropriate assessment tools difficult for both researchers and educators. • Our next steps include: • Continuing to identify medical assessment tools through an expanded PubMed search and a search of . the social science literature; and • Developing a framework to better disseminate valid . assessment tools to the medical education community. Presented at the Southern Group on Educational Affairs Regional Conference, Houston, TX 04/15/2011