Overview of Types of Measures Margaret Kasimatis, PhD VP for Academic Planning & Effectiveness Location in Assessment Cycle IDENTIFY SPECIFIC OUTCOMES DETERMINE PRACTICES USED TO ACHIEVE OUTCOMES GATHER EVIDENCE ARTICULATE MISSION/ GOALS RECOMMEND ACTIONS REVIEW & INTERPRET RESULTS Measures Within the Context of Program Assessment • For program assessment, measurement tools should capture student learning that occurs as a result of the program curriculum • Many measurement tools can be used for multiple levels of assessment – But how you use them differs • Some assessment tools are not appropriate for program-level assessment Categories of Measures • Direct Measures – Look at student work products or performances that demonstrate level of learning • Indirect Measures – Capture students’ perceptions of their learning and the educational environment that supports learning Categories of Measures • Direct Measures – Published, standardized tests (e.g., GRE Subject Test, ETS Major Field Test) – Locally developed tests – Systematic evaluation of student work (papers, presentations, creative work, performances) • May or may not be embedded within courses • Usually involves scoring rubrics • Indirect Measures – Published surveys – Locally developed surveys and interviews – Alumni surveys Properties of Good Assessment Techniques • Reliable – internally consistent; consistent across raters • Valid – measures what it is supposed to; appropriate • Actionable – results point reviewers toward challenges to address (and how to address them) • Efficient and cost effective in time and money • Interesting and meaningful – people care about the results and are willing to act on them • Convergence – multiple lines of evidence point to the same conclusion Evaluating Measures: Direct PUBLISHED TESTS PROS • Provide direct evidence of student mastery of content • Some are designed specifically to assess major programs • Generally highly reliable CONS • May be difficult to motivate students to perform at their best level • May not align with program outcomes • Validity established (within a specific context) • Often focus more on content knowledge than higher-order skills • Usually easy to implement and obtain results • Can be expensive • Norms/comparisons available Evaluating Measures: Direct LOCALLY DEVELOPED TESTS PROS • Provide direct evidence of student mastery of content or skills • More flexible in terms of content and format; easier to align with program outcomes • If embedded in courses, student motivation is higher • Faculty are more likely to be interested in and use results CONS • Likely to be less reliable • Validity unknown • Norms/comparisons not available • Can take several iterations (and several years) to work out the “bugs” • Scoring tests and tabulating results can be cumbersome Evaluating Measures: Direct EVALUATION OF STUDENT WORK PROS • Provides direct evidence of student mastery of content or skills • If embedded in course, student motivation is higher • Faculty are more likely to be interested in and use results • Data collection is usually unobtrusive to students CONS • Requires time to develop, conduct training, implement – Creating flexible rubric, at appropriate for program-level assessment can be tricky • Validity unknown; takes time to establish reliability • Requires faculty trust that the program will be assessed, not individual instructors • Norms/comparisons not available Evaluating Measures: Indirect PUBLISHED SURVEYS PROS CONS • Minimal effort to implement and tabulate results • Provides indirect evidence of student learning • Can be administered to a large group of respondents • May not be aligned with program outcomes • Demonstrated reliability and validity • Potential for biased results if sample is not representative • Can address a variety of outcomes • Can be expensive • Norms/comparisons available Evaluating Measures: Indirect LOCALLY DEVELOPED STUDENT SURVEYS & INTERVIEWS PROS CONS • Flexible in terms of content and format; easy to align with program outcomes • Provide indirect evidence of student learning • Usually have face validity • Their validity depends on the quality of questions and response options • Can add open-ended questions that allow you to flesh out quantitative results – More actionable • Can be administered to a large group of respondents • Can address a variety of outcomes • Relatively easy to implement • Potential for biased results if sample is not representative • Can be time-consuming to construct, implement, and tabulate results • Norms/comparisons not available • Open-ended responses can be difficult and time-consuming to analyze Evaluating Measures: Indirect ALUMNI SURVEYS PROS CONS • Same advantages as student surveys • Many of the same disadvantages of student surveys • Can gather more “direct” evidence – It’s particularly difficult to get a good than current student surveys response rate – E.g., employment, enrollment in graduate programs • Can ask questions of alumni that are not appropriate for current students – E.g., the extent to which the program prepared them for their career • The timing can be tricky – Alumni should be far enough out to see an impact of the program on their life/career but not so far out that the program they experienced is very different from current one A Word on Embedded Assessment • Various types of measurement tools can be embedded within courses • Only carefully constructed measures, used in certain types of courses, are appropriate for program-level assessment – Must go beyond individual course content – Some should occur in upper-level courses that are taken only after several other courses in the major QUESTIONS?