CLINICAL EDUCATION Validation of a Scenario-Based Assessment of Critical Thinking Using an Externally Validated Tool Jennifer L. Buur n Peggy Schmidt n Dean Smylie n Kris Irizarry n Carlos Crocker n John Tyler n Margaret Barr ABSTRACT With medical education transitioning from knowledge-based curricula to competency-based curricula, critical thinking skills have emerged as a major competency. While there are validated external instruments for assessing critical thinking, many educators have created their own custom assessments of critical thinking. However, the face validity of these assessments has not been challenged. The purpose of this study was to compare results from a custom assessment of critical thinking with the results from a validated external instrument of critical thinking. Students from the College of Veterinary Medicine at Western University of Health Sciences were administered a custom assessment of critical thinking (ACT) examination and the externally validated instrument, California Critical Thinking Skills Test (CCTST), in the spring of 2011. Total scores and sub-scores from each exam were analyzed for significant correlations using Pearson correlation coefficients. Significant correlations between ACT Blooms 2 and deductive reasoning and total ACT score and deductive reasoning were demonstrated with correlation coefficients of 0.24 and 0.22, respectively. No other statistically significant correlations were found. The lack of significant correlation between the two examinations illustrates the need in medical education to externally validate internal custom assessments. Ultimately, the development and validation of custom assessments of non-knowledge-based competencies will produce higher quality medical professionals. Key words: assessment, validation, critical thinking INTRODUCTION Critical thinking in the health professions has been identified as an essential non-technical competency.1–3 According to the National Council for Excellence in Critical Thinking Instruction, ‘‘critical thinking is the intellectually disciplined process of actively and skillfully conceptualizing, applying, analyzing, synthesizing, or evaluating information gathered from, or generated by, observation, experience, reflection, reasoning, or communication, as a guide to belief and action.’’4 Currently, there are just a few fully validated external instruments for assessing critical thinking including the California Critical Thinking Skills Test (CCTST) and the Watson-Glaser Critical Thinking Appraisal.5,6 These validated external instruments have been used to track the development of critical thinking over time in many health professions. Since critical thinking is considered to be a transferable skill, medical professional programs have developed a variety of custom assessments of critical thinking including written assignments, extended match multiple choice, case vignettes, portfolios, and oral examinations.2,7,8 However, the face validity of these custom assessments has not been challenged. To our knowledge, a study comparing the face validity of a custom assessment in critical thinking to a validated external instrument has not been published to date. Another way of evaluating critical thinking has been to compare outcomes to Bloom’s taxonomy of learning. These levels represent a hierarchy by which learners build upon foundations to increase levels of thinking starting with knowledge and proceeding through comprehension, application, analysis, synthesis, and ending in evaluation. 276 A complete review of Bloom’s taxonomy is beyond the scope of this manuscript and interested readers are referred to the literature.9,10 The purpose of this study was to compare student performance on a custom assessment (assessment of critical thinking exam) with a validated external instrument (CCTST). We hypothesized that there would be a direct correlation between the total score and sub-scores of the assessment of critical thinking (ACT) examination and the total score and sub-scores of the CCTST. METHODOLOGY Student Population Of the Western University of Health Sciences College of Veterinary Medicine Graduating Class of 2013, 106 members were administered both the ACT Exam and the CCTST during March 2011 as a normal part of the curriculum. Students who did not take both exams within 72 hours of each other were excluded from the study. Use of human subjects was waived by the Western University of Health Sciences Human Subject Institutional Review Board. ACT Exam The ACT exam consists of a paper-based novel clinical scenario including physical examination and preliminary diagnostic testing results with an identified problem list. Students are asked to focus on a single identified problem and the corresponding list of differential diagnoses. Students must choose their top two differentials and explain doi:10.3138/jvme.0112-009R JVME 39(3) 6 2012 AAVMC Table 1: Descriptive statistics of second-year veterinary student performance on the ACT exam and the CCTST taken in March 2011 ACT total ACT B1 ACT B2 ACT B3 CCTST total IR DR AI IN EE Average 84.82 20.55 35.54 24.47 20.89 11.45 9.44 5.09 10.06 5.73 Median 84.25 20 35.5 24 21 11 9 5 10 6 2.77 1.26 SD 5.84 2.09 3.89 2.97 4.34 2.15 2.22 2.13 ACT ¼ assessment of critical thinking; CCTST ¼ California Critical Thinking Skills Test; B ¼ Blooms level; IR ¼ inductive reasoning; DR ¼ deductive reasoning; AI ¼ analysis and interpretation; IN ¼ inference; EE ¼ evaluation and explanation; SD ¼ standard deviation the pathophysiology of the specific disease process and how it relates to the case, identify supporting and nonsupporting information presented in the scenario, and then develop a diagnostic plan to confirm their choices. Using a problem-oriented medical approach,11 differential diagnoses are presented mechanistically using the DAMNIT-V (degenerative, anomaly, metabolic, neoplastic/ nutritional, inflammatory/infectious/immune mediated, toxic/traumatic, vascular) system of organization after localization to a major body system. Appendix 1 presents the ACT scenario used in this study as it was presented to students. The rubric for this exam has nine sections including chosen diagnoses, pathophysiology, supporting history, supporting exam data, other supporting data, non-supporting data, further diagnostic plans, terminology, and legibility that contribute to the 100 point total value (Appendix 2). All sections except diagnostic choice, plans, terminology, and legibility are composites of two separate scores that correlate to each specific diagnostic choice. The examination is administered four times throughout the first two years of the curriculum using four different novel clinical scenarios. The key for each exam scenario is written at the level deemed appropriate for students at their current level of competency. The ACT exam constitutes 10% of the student’s final grade. Students in this study had taken this format of exam once during each of the previous three semesters. The key was written for fourth-semester veterinary students as determined by faculty content experts who have board certification in a clinical specialty. Each grader is trained on how to use the rubric and key for the specific section they are grading. This includes mock grading of student exams until sufficient reliability has occurred. Faculty are assigned a single diagnosis to grade. To maintain consistency between students, a single faculty member grades every examination section relating to their assigned diagnosis. A single faculty member grades all plans. For the study, ACT exam scores were composites of three numerical sub-scores that relate to Bloom’s levels of learning. Pathophysiology was considered to be Blooms level 2 (comprehension/understanding) and was valued at 20 points. Identification of supporting and non-supporting data was considered to be Blooms level 3 (application) and was valued at 40 points. Diagnostic choice and diagnostic plans were combined and considered to be Blooms level 4 (analysis) and had a total value of 30 points. Terminology (5 points of total) and legibility (5 JVME 39(3) 6 2012 AAVMC points of total) were not specifically used in the correlations although are retained as part of the total ACT score. California Critical Thinking Skills Test This study used the CCTST Form A which is a standardized, 34-item multiple choice test. The test provides a total critical thinking skills score and sub-scores for five core areas of critical thinking: analysis and interpretation, evaluation and explanation, inference, deductive reasoning, and inductive reasoning. The CCTST was administered once previously at matriculation (August 2009). The CCTST is an external assessment of critical thinking that was validated using the Delphi method. Details of the examination and its validation process are beyond the scope of this article and can be found in the literature.6 All students at Western University of Health Sciences are required to take the CCTST as part of university outcome assessment. Statistical Methods The degree of linear association between total score and sub-score numerical values of the ACT and CCTST was determined using the Pearson correlation coefficient. Significance level was set as p a .05. Statistical analysis was performed using SAS 9.2 (SAS Institute Inc., Cary, NC, USA.) RESULTS Ninety-eight students were included in the study. Student performance on both ACT and CCTST exams are presented in Table 1. As seen in Table 2, ACT sub-scores correlated significantly with the total ACT score. CCTST sub-scores also correlated significantly with total CCTST score. Significant correlations were found between ACT Blooms 2 and deductive reasoning (correlation coefficient 0.24, p ¼ .018) and total ACT score and deductive reasoning (correlation coefficient 0.22, p ¼ .027). No other statistically significant correlations were found. A slight trend was seen between total ACT score and inference (correlation coefficient 0.18, p ¼ .076). DISCUSSION Results from this study demonstrate weak correlations between student performances on the ACT exam and on the CCTST. If student performance on the ACT exam does not correlate strongly with an externally validated tool designed to test critical thinking, then what does the ACT exam actually test? It is possible that the ACT exam 277 Table 2: Pearson correlations between total and sub-scores of ACT exam and total and sub-scores of CCTST administered to second-year veterinary students in March 2011 ACT total ACT B2 ACT B3 ACT B4 ACT total ACT B1 ACT B2 ACT B3 IR DR AI IN 1 0.69 † 0.45 † 0.74 † 0.00 0.22 * 0.14 0.18 0.02 0.14 0.24 * 0.28 † 0.03 0.24 * 0.14 0.15 0.04 0.14 1 0.14 0.05 0.10 0.07 0.07 0.04 0.04 1 0.04 0.11 0.04 0.15 0.01 0.09 1 0.54 † 0.43 † 0.59 † 0.85 † 0.84 † 1 0.51 † 0.89 † 0.63 † 0.91 † 1 0.27 † 0.22 * 0.54 † 1 0.53 † 0.85 † 1 0.83 † 1 IR DR AI IN EE CCTST total EE CCTST total 1 * p < .05 † p < .01 ACT ¼ assessment of critical thinking; CCTST ¼ California Critical Thinking Skills Test; B ¼ Blooms level; IR ¼ inductive reasoning; DR ¼ deductive reasoning; AI ¼ analysis and interpretation; IN ¼ inference; EE ¼ evaluation and explanation provides another format for testing specific veterinary knowledge. Within the class, there is a subset of individuals who earn high marks on the ACT exam while performing poorly on other knowledge-based assessments including multiple choice and laboratory-based assessments (data not shown). It is possible that these students have the ability to recall information on contextual examinations like the ACT rather than in abstract examinations like multiple choice examinations. In addition, these students could also have well-developed writing skills such that they are able to articulate their knowledge better on the ACT than on other assessments. This version of the examination focused on small-animal medicine. It is possible that different clinical scenarios used on the ACT exam could have stronger correlations with the CCTST. With the trend in veterinary medicine toward smallanimal practice,12 it is possible that students used pattern recognition from clinical experience rather than critical thinking to complete the examination. Pattern recognition and critical thinking are both associated with clinical reasoning.13–15 Although it is tempting to postulate that the ACT exam actually assesses clinical reasoning rather than critical thinking, it is hard to see how the subset skill of critical thinking would not be demonstrated in any test of clinical reasoning. Since there are currently no validated assessment tools for clinical reasoning, it is impossible to determine if this is what the ACT is actually assessing. The CCTST has been criticized for not relating to clinically-oriented critical thinking in nursing.16 Given that the ACT examination is a clinical case, it is possible that it is measuring a different application of critical thinking than what is tested on the CCTST. Further studies looking at the correlation between performance on the ACT examination and the Watson-Glaser Critical Thinking Appraisal could be done to further elucidate the origin of the poor correlations. The only significant correlations found in this study were between Blooms level 2 and total ACT score and deductive reasoning on the CCTST. Blooms level 2 is associated 278 with comprehension and the ability to make use of specific knowledge. This level does not imply the full and complete understanding of material and is thought to be equivalent to the thinking ability of advanced highschool students.17 Our results run counter to the published literature that promotes Blooms level 4 or greater as evidence of critical thinking. The CCTST describes deductive reasoning as moving ‘‘from the assumed truth of a set of beliefs or premises to a conclusion which follows of necessity.’’18 Deductive reasoning based on pattern recognition from the provided list of diagnoses could explain the minor correlation found between ACT total score and CCTST deductive reasoning. Alternatively, the assumption that a description of pathophysiology is Blooms level 2 thinking may be inaccurate. While the rubric was designed to try to tease out the effects of pattern recognition from critical thinking, the lack of correlation could be due to the lack of fine resolution between these variables. In this case, the rubric would need to be further refined. Finally, the assumption that higher Blooms levels reflect critical thinking may be false. More research, including correlations to knowledge-based examinations, is required to truly understand what the ACT exam is really testing. The lack of correlation in this study demonstrates a need in medical education to validate custom assessments against validated external instruments such as the CCTST. The pursuit of scholarly teaching requires that educators look at the validity and reliability of assessment strategies to truly understand what assessments are needed to ensure well-trained graduates.19 The CCTST is the most predominant validated external assessment of critical thinking used in the medical professional literature and has been used to look at the development of critical thinking over time in a variety of health professions.20–22 However, to the authors’ knowledge, this is the first publication using this tool to challenge the face validity of a custom assessment strategy. JVME 39(3) 6 2012 AAVMC In conclusion, the ACT exam does not strongly correlate to critical thinking as defined by the CCTST. While this custom assessment could be valid for testing contextual knowledge or clinical reasoning, it is not recommended to implement this examination to assess critical thinking. It does, however, provide a prime example of why medical educators need to subject their custom assessments to rigorous validation using validated external instruments. As medical education shifts from predominately knowledgebased curricula to the development of specific skill sets like critical thinking, it is even more important to have confidence in our assessment techniques. There may also come a time when accreditation bodies will require the use of validated external instruments. Ultimately, having confidence in our assessment strategies will result in better trained graduates. ACKNOWLEDGMENTS The authors wish to thank the faculty of Western University of Health Sciences for their ongoing participation in the Veterinary Basic Medical Science course. REFERENCES 1 Hagan BA. Tools for implementing a competencybased clinical curriculum: the dental school experience. J Vet Med Educ. 2008;35(3):369–74. http://dx.doi.org/ 10.3138/jvme.35.3.369. Medline:19066353 2 Berkow S, Virkstis K, Stewart J, et al. Assessing individual frontline nurse critical thinking. J Nurs Adm. 2011;41(4):168–71. http://dx.doi.org/10.1097/ NNA.0b013e3182118528. Medline:21430465 3 North American Veterinary Medical Education Consortium. Roadmap for veterinary medical education in the 21st century: responsive, collaborative, flexible. Washington, DC: NAVMEC Report and Recommendations; 2011. 4 Paul, R. Nosich GM. A model for the national assessment of higher order thinking [Internet]. Foundation for Critical Thinking; 2011 [cited 2011 Oct 5]. Available from: http://www.criticalthinking.org/pages/ a-model-for-the-national-assessment-of-higher-orderthinking/591. 5 Macpherson K, Owen C. Assessment of critical thinking ability in medical students. Assess Eval High Educ. 2010;35(1):45–58. 6 Facione NC, Facione PA, Sanchez CA. Critical thinking disposition as a measure of competent clinical judgment: the development of the California Critical Thinking Disposition Inventory. J Nurs Educ. 1994;33(8):345–50. Medline:7799093 7 Swinny B. Assessing and developing critical-thinking skills in the intensive care unit. Crit Care Nurs Q. 2010;33(1):2–9. Medline:20019504 8 Connors P. Assessing written evidence of critical thinking using an analytic rubric. J Nutr Educ Behav. 2008;40(3):193–4. http://dx.doi.org/10.1016/ j.jneb.2008.01.014. Medline:18457791 JVME 39(3) 6 2012 AAVMC 9 Gilbert SW. Systematic questioning: taxonomies that develop critical thinking skills. Sci Teach. 1992;59(9):41–6. 10 Richlin L. Blueprint for learning. Sterling, Virginia: Stylus Publishing, LLC; 2006. p. 45–51. 11 Lorenz M. Small animal clinical diagnosis. 3rd ed. Ames, Iowa: Blackwell Publishing; 2009. p. 3–12. 12 Willis NG. The animal health foresight project. Vet Ital. 2007;43(2):247–56. Medline:20411514 13 Tomlin JL, Pead MJ, May SA. Veterinary students’ attitudes toward the assessment of clinical reasoning using extended matching questions. J Vet Med Educ. 2008;35(4):612–21. http://dx.doi.org/10.3138/ jvme.35.4.612. Medline:19228917 14 Tomlin JL, Pead MJ, May SA. Attitudes of veterinary faculty to the assessment of clinical reasoning using extended matching questions. J Vet Med Educ. 2008;35(4):622–30. http://dx.doi.org/10.3138/ jvme.35.4.622. Medline:19228918 15 Carriè€re B, Gagnon R, Charlin B, et al. Assessing clinical reasoning in pediatric emergency medicine: validity evidence for a Script Concordance Test. Ann Emerg Med. 2009;53(5):647–52. http://dx.doi.org/ 10.1016/j.annemergmed.2008.07.024. Medline:18722694 16 Allen GD, Rubenfeld MG, Scheffer BK. Reliability of assessment of critical thinking. J Prof Nurs. 2004;20(1):15–22. http://dx.doi.org/10.1016/ j.profnurs.2003.12.004. Medline:15011189 17 Vuchetich PJ, Hamilton WR, Ahmad SO, et al. Analyzing course objectives: assessing critical thinking in the pharmacy curriculum. J Allied Health. 2006;35(4):e253–75. Medline:19759975 18 Scales of the CCTST [Internet]. Insight Assessment; 2011 [cited 2011 Oct 5]. Available from: http:// www.insightassessment.com/Products/CriticalThinking-Skills-Tests/California-Critical-Thinking-SkillsTest-CCTST/Scales-of-the-CCTST. 19 Downing SM, Haladyna TM. Validity threats: overcoming interference with proposed interpretations of assessment data. Med Educ. 2004;38(3):327–33. http://dx.doi.org/10.1046/j.1365-2923.2004.01777.x. Medline:14996342 20 Bartlett DJ, Cox PD. Measuring change in students’ critical thinking ability: implications for health care education. J Allied Health. 2002;31(2):64–9. Medline:12040999 21 Wong MS. A prospective study on the development of critical thinking skills for student prosthetists and orthotists in Hong Kong. Prosthet Orthot Int. 2007;31(2):138–46. http://dx.doi.org/10.1080/ 03093640600983931. Medline:17520491 22 McCarthy P, Schuster P, Zehr P, et al. Evaluation of critical thinking in a baccalaureate nursing program. J Nurs Educ. 1999;38(3):142–4. Medline:10102514 279 AUTHOR INFORMATION Jennifer L. Buur, DVM, PhD, is Assistant Professor at Western University of Health Sciences, College of Veterinary Medicine, 309 E. Second St., Pomona, CA 91766 USA. E-mail: jbuur@westernu.edu. Peggy Schmidt, DVM, is Associate Professor at Western University of Health Sciences, College of Veterinary Medicine, 309 E. Second St., Pomona, CA 91766 USA. Dean Smylie, DVM, is Assistant Professor at Western University of Health Sciences, College of Veterinary Medicine, 309 E. Second St., Pomona, CA 91766 USA. Kris Irizarry PhD, is Assistant Professor at Western University of Health Sciences, College of Veterinary Medicine, 309 E. Second St., Pomona, CA 91766 USA. Carlos Crocker, PhD is Associate Professor at St Matthew’s University, School of Veterinary Medicine, P.O. Box 30992 Grand Cayman, KY1–1204 Cayman Islands, BWI. John Tyler, DVM, is Associate Professor at Western University of Health Sciences, College of Veterinary Medicine, 309 E. Second St., Pomona, CA 91766 USA. Margaret Barr, DVM, PhD, is Professor at Western University of Health Sciences, College of Veterinary Medicine, 309 E. Second St., Pomona, CA 91766 USA. APPENDIX 1 Assessment of critical thinking (ACT) examination scenario and format given March 2011 to second-year veterinary students Hydration Status Attitude Body Condition Normal BAR 3.5/5 Please list your top two choices for differential diagnoses here before handing in the exam: Mucous Membranes: Pink and moist Differential Diagnosis #1: [Part A1] Capillary Refill Time: 1 sec Differential Diagnosis #2: [Part A2] INDICATE: (N-Normal; A-Abnormal; NE-Not Examined) Presenting Scenario: You are a partner in a busy small-animal practice in Southern California. As you leave one of the exam rooms, your veterinary technician hands you the record for your next patient. ‘‘Roscoe’’ is an 8-year-old male Rottweiler that you have seen regularly since he was a puppy for wellness exams and vaccinations. A quick check of his file reveals that he received his most recent DHP-P and rabies vaccines 18 months ago. Roscoe’s owner, Ms. Taylor, is concerned because Roscoe’s left eye and face ‘‘look odd.’’ The first thing you notice about Roscoe when you enter the exam room is a slightly lopsided (asymmetrical) appearance to his face and head. When Roscoe stands up and trots over to greet you, his gait appears normal. Ms. Taylor reports that Roscoe has not had any problems with vomiting or diarrhea, and he has not been coughing or sneezing. His activity level has decreased over the past few months, and Ms. Taylor believes that Roscoe has gained some weight. His normal diet is a commercial adult dog food (usually whatever is on sale) fed free choice. Although his appetite has been good, he seems to be eating more slowly in addition to dribbling small amounts of food from his mouth occasionally. In response to your question about potential injuries, Ms. Taylor sheepishly admits that Roscoe was riding in the back of her pick-up truck a couple of weeks ago when they were involved in a minor rear-end collision. Roscoe slid forward in the truck and may have bumped his head, but he jumped up immediately and seemed to be okay afterward. Physical Examination: Temperature: 100.6 F Pulse: 68/min 01 HEAD A 05 ORAL CAVITY 02 EYES A 09 MAMMARY N 13 RECTAL GLANDS A 06 CERVICAL N 10 ABDOMEN N 14 SKIN NE N 03 EARS N 07 THORAX N 11 BACK/TAIL N 15 EXTREMITIES N 04 NOSE N 08 HEART N 12 GENITALIA N 16 NEUROLOGIC A 01 Temporalis and masseter muscle atrophy – left side, moderate 02 Diminished palpebral and corneal reflexes OS (menace and pupillary reflexes intact); slight corneal opacity OS 05 Small amount of food is present in left cheek 16 Diminished facial sensation (decreased response to pin pricks) on left side as compared to right side of face; swallow reflex is normal Laboratory data: CBC and Blood Chemistry Panel TEST Patient Normal Range Red blood cells Hemoglobin PCV (hematocrit) White blood cells Neutrophils Lymphocytes Monocytes Eosinophils Platelets 7.3 15 46 11,500 7,475 2,645 460 920 350 5.5–8.5 106/ml 12–18 g/dL 37–55% 6,000–17,000/ml 3,000–11,400/ml 1,000–4,800/ml 150–1,350/ml 100–750/ml 200–900 103/ml BUN ALT Alk Phos Creatine kinase Ca2+ Na+ K+ Total Protein Albumin Globulin 20.4 45 15.5 85 10.2 142 4.63 7.0 3.9 3.1 8.8–26 mg/dL 8.2–57 IU/L 10.6–101 IU/L 14–120 IU/L 8.7–11.8 mg/dL 140–154 meq/L 3.8–5.6 meq/L 6.0–7.5 g/dL 2.7–4.4 g/dL 1.6–3.4 g/dL Respiration: panting Weight: 97.5 lbs 280 JVME 39(3) 6 2012 AAVMC Problem: Asymmetry of head Based on your physical examination and laboratory data, choose your top 2 differential diagnoses from the list below and justify your choices on the following pages: (note – this list is NOT comprehensive) Degenerative e Degenerative myelopathy Metabolic e Hypothyroidism-associated peripheral neuropathy Neoplastic e Trigeminal nerve neoplasia (nerve-sheath tumor) e CNS neoplasia (meningioma, other) Infectious/Inflammatory/Immune-mediated e Viral meningoencephalitis e Masticatory muscle myositis e Otitis media/interna-associated neuropathy Traumatic e Horner’s syndrome secondary to head trauma e Traumatic myopathy Vascular e Cerebrovascular incident (ischemia or hemorrhage) PARTS A1 AND A2 Problem: Asymmetry of head Differential Diagnosis #1 (from list given): [Differential 1 is Part A1. Differential 2 is Part A2. Students provide the pathophysiology and table information for each differential separately.] JVME 39(3) 6 2012 AAVMC How does the PATHOPHYSIOLOGY of this differential diagnosis explain the presenting clinical scenario? [Blooms 1, Rubric Row 2] History Information Supporting: [Blooms 2, Rubric Row 3] Rationale: [Blooms 2] Physical Examination Information Supporting: [Blooms 2, Rubric Row 4] Rationale: [Blooms 2] Other Supporting Information: [Blooms 2, Rubric Row 5] Rationale: [Blooms 2] Non-supportive Information: [Blooms 2, Rubric Row 6] Rationale: [Blooms 2] PART B Problem: Asymmetry of head Differential Diagnosis#1 (from list given): [Blooms 3, Rubric Row 1] Differential Diagnosis#2 (from list given): [Blooms 3, Rubric Row 1] How will you determine which diagnosis is most likely? What can you do to help rule out the other potential diagnoses? Justify your answer. Plan Rationale [Blooms 3, Rubric Row 7] [Blooms 3] 281 APPENDIX 2 Assessment of critical thinking (ACT) rubric used March 2011 for grading of ACT exam delivered to secondyear veterinary students Student ID:___________ Grader ID:___________ No appropriate choices from list of differential diagnoses 4 pts One appropriate choice from list of differential diagnoses consistent with data available 8 pts Two appropriate choices from list of differential diagnoses consistent with data available [Row 1] [Blooms 3] 10 pts Pathophysiologic mechanism is described accurately for the chosen differential diagnosis; demonstrates insight and specific knowledge [Row 2, A1/A2] [Blooms 1] 10/10 pts Minor error in description of pathophysiologic mechanism for the differential diagnosis; moderate evidence of insight or specific knowledge 8/8 pts Multiple minor errors in description of pathophysiologic mechanism; restricted ability to describe pathophysiologic mechanism of differential diagnosis; some evidence of insight or specific knowledge 7/7 pts Major errors in description of mechanism for the differential diagnosis; little evidence of insight or specific knowledge 5/5 pts Pathophysiologic mechanism poorly described for the differential diagnosis; specific knowledge not demonstrated 3/3 pts Correct identification of all supporting history information; appropriate rationale for all information [Row 3, A1/A2] [Blooms 2] 5/5 pts Correct identification of most supporting history information; appropriate rationale for most information 4.5/4.5 pts Correct identification of some supporting history information; appropriate rationale for some information 4/4 pts Correct identification of some supporting history information; rationale not given or incorrect 3/3 pts Incorrect identification of supporting history information; rationale not given or incorrect 1/1 pts Correct identification of all supporting physical exam information; appropriate rationale for all information [Row 4, A1/A2] [Blooms 2] 5/5 pts Correct identification of most supporting physical exam information; appropriate rationale for most information 4.5/4.5 pts Correct identification of some supporting physical exam information; appropriate rationale for some information 4/4 pts Correct identification of some supporting physical exam information; rationale not given or incorrect 3/3 pts Incorrect identification of supporting physical exam information; rationale not given or incorrect 1/1 pts Correct identification of all other supporting information; appropriate rationale for all information [Row 5, A1/A2] [Blooms 2] 5/5 pts Correct identification of most other supporting information; appropriate rationale for most information 4.5/4.5 pts Correct identification of some other supporting information; appropriate rationale for some information 4/4 pts Correct identification of some other supporting information; rationale not given or incorrect 3/3 pts Incorrect identification of other supporting information; rationale not given or incorrect 1/1 pts Correct identification of all nonsupportive information; appropriate rationale for all information [Row 6, A1/A2] [Blooms 2] 5/5 pts Correct identification of most nonsupportive information; appropriate rationale for most information 4.5/4.5 pts Correct identification of some nonsupportive information; appropriate rationale for some information 4/4 pts Correct identification of some nonsupportive information; rationale not given or incorrect 3/3 pts Incorrect identification of nonsupportive information; rationale not given or incorrect 1/1 pts Thorough, detailed investigative plan associated clearly with all defined differentials; appropriate rationale given for each plan [Row 7] [Blooms 3] 20 pts Adequate plan associated clearly with all defined differentials; appropriate rationale given for most plans 16 pts Basic plan; some stated in generalities not clearly associated with defined ideas; rationale not welldefined for some plans 14 pts Generalities not specific and not clearly associated with defined ideas; incorrect rationale given or missing rationale for most plans 10 pts Major action plan of significance to the investigation not defined; little or no rationale given 6 pts Appropriate terminology demonstrating expert use of accurate scientific vocabulary and spelling [Not included] 5 pts Appropriate terminology; usually adequate use of accurate scientific vocabulary 4 pts Some use of lay terminology where scientific terms were expected; some misuse of terms; minor misspellings 3 pts Inadequate use of precise terminology; embarrassing misuse of terms; significant misspellings 2 pts Failed to use appropriate terminology; vocabulary errors would result in miscommunication and injury to the patient 1 pts Notations were orderly, legible and clearly demonstrated the student’s thought process and logic [Not included] 5 pts Notations were effective but present a few problems with order, legibility, thought process and/or logic 4 pts Notations were effective but present multiple problems with order, legibility, thought process and/or logic 3 pts Inadequate or illegible notations; multiple problems with order, legibility, thought process and/or logic 2 pts Failed to communicate most thought process and/or logic 1 pts B 80–89 C 70–79 D 65–69 U 0–64 Scoring Code A % equivalent 90–100 282 JVME 39(3) 6 2012 AAVMC