Assessment Challenges In Education Jamie Cromack (Chair) Program Manager, ER&P Microsoft Corporation Wilhelmina Savenye Professor – Educational Technology Arizona State University Jane Prey Program Manager, ER&P Microsoft Corporation Agenda Introducing the MSR Assessment Toolkit Background and goals Brief overview of the Toolkit Discussion of assessment challenges Background What is the MSR Assessment Toolkit? The Assessment Toolkit is a unique resource for CS, Engineering, other STEM faculty Though many web-based resources exist, none will do what this MSR tool will A partnership among MSR and leading faculty in assessment and faculty in Computing Science, Science, Science, Technology, Engineering and Mathematics (STEM) disciplines MSR Faculty Summit Packet White Paper Learning about Learning in Computational Science and Science, Technology, Engineering and Mathematics (STEM) Education J. Cromack and W. Savenye, 2007 Background Why was the MSR Assessment Toolkit developed? Faculty must do assessment; few have the time, training, and support to do it as well as they would like STEM faculty are very interested in using the MSR Assessment Toolkit Background Why has assessment become increasingly important? External factors – trends: Accountability/accreditation Funding agency requirements Funding and enrollment issues Changes in pedagogy, instructional design Internal factors – collect data to: Improve educational experience for students Motivate students Microsoft Research AssessmentToolkit Developed by: Willi Savenye and Gamze Ozogul, Arizona State University Jamie Cromack, EP&R, MSR The MSR Assessment Toolkit The Assessment Toolkit consists of a set of assessment resources designed to aid faculty in building stronger measures of the success of their projects, during all phases of their projects’ life cycles Version 1 now deployed We are looking forward to your feedback MSR AssessmentToolkit Components 1. Assessment Planning Overview 2. Assessment FAQs 3. User Grid to help faculty determine which resources might be most useful for their level of assessment experience. 4. Assessment Methods Selection Guide 5. Focused Assessment Resources (an annotated bibliography) 6. Case Studies in Effective Assessment MSR AssessmentToolkit Components 1. Assessment Planning Overview 2. Assessment FAQs 3. The User Grid What level of assessment user are you? BASIC (B) TYPICAL (T) POWER (P) If you : Want to know more about how students learn in or experience your class Have generally administered only the university-constructed end-ofterm student surveys Have not used a rubric to grade an assignment If you have done some of: Have thought about specific learning goals or outcomes for your class Have thought about the overall goals of your curriculum and how it prepares students Supplemented the universityconstructed end-of-term surveys with simple mid-term surveys Done minute papers or fastfeedback sheets Have used pre-designed rubrics with minimal adaptation If you Seek to define and describe in detail student experience with learning goals in your classroom Seek to define and describe in detail student preparation as experienced through your curriculum Have regularly administered pre/post surveys, mid-term surveys and end-of-term surveys Have created from scratch or significantly adapted a rubric for grading Have reported on qualitative or quantitative results of studies of student understanding in peerreviewed literature venues 4. Assessment Methods Selection Guide As faculty begin to select their assessment methods, they should consider making sure that they measure: Student Performance, Satisfaction, and Retention (if desired) Include a balance of both direct and indirect measures, as well as both formative measures (to improve instruction in process) and summative measures (to make evaluative decisions at or near the end of the project) 4. Assessment Methods Selection Guide Direct (observable phenomena) Indirect (self-report data) Summative (accountability, retroactive, evaluative) Exams & quizzes (may be used pre- and postcourse) Projects & assignments (often evaluated using rubrics/checklists) Student performance on case studies Final grades Concept maps/flowcharts Observations (often using rubrics/checklists) Portfolios Capstone projects Course/test embedded assessment Outcomes assessment (often using rubrics/checklists) Demonstration of a skill (often using rubrics/checklists) Exit surveys Exit interviews End of term survey Student ratings Interviews Formative (decision making, proactive, diagnostic) Exams & quizzes (may be used pre- and postcourse) Projects & assignments (often evaluated using rubrics/checklists) Student performance on case studies Mid-course and assignment grades Concept maps/flowcharts Observations (often using rubrics/checklists) Background knowledge probes Minute or fast-feedback papers Student-generated test questions Diagnostic learning logs or journals Muddiest point surveys Satisfaction surveys Attitude surveys Mid-semester feedback surveys Student ratings Interviews Methods For Measuring Student Performance And Retention: (E.g., a short selection) Method User level Class Size Type Description Strengths Limitations Example question(s) Constructed responsemultiple choice items (performan ce, retention) B, T, P La F/S Used to gather data about student performance and level of mastery of the course content. Can be used in pretest and posttest forms administered at the beginning of the course and again at the end. • Can cover broad scope of content • Objective assessment • Easy to score • Reusable • Reliability and validity concerns • Time consuming to build • Answers limited to constructed responses • Which one below is a type of processing? A) Batch B) Series C) Network D) Distributed Short answer items (performan ce/ retention) B, T La F/S Used to gather data about student knowledge about a topic Can be used in pretest and posttest forms administered at the beginning of the course and again at the end. • Easy to prepare • Reusable • Easy to score • Limited number of questions • Limited scope of content What is an algorithm? Grades (mid-course or final grades) (performan ce or retention) B, T, P La/ Sm F/S Collected on projects, assignments, quizzes, etc. mid-course, as well as collected at the end of the course. • Easy to collect, as necessary • Can be tracked over multiple courses over time • May not be finegrained enough to yield differences • Influenced by many factors in addition to course innovations • Grade book entries METHODS FOR MEASURING STUDENT SATISFACTION (Eg. a short selection) Instrument User level Class Size Type Description Strengths Limitations Example question(s) Attitude surveys (satisfaction) B, T,P La/ Sm F/S Provide data about students’ perceptions about a topic/class. Can be used to make formative or summative decisions depending on the time of administration. Generally Likert type questions; may be followed by a few open-ended questions. May be used as a pre/post measure. • Easy to administer • Cover many areas of attitude • Can be used in large groups • Do not provide in- depth information about attitudes • The course projects were difficult for me. • Strongly agree, agree, disagree, strongly disagree Happy sheets (satisfaction) B La F Provide data about students’ perceptions. Generally used to make formative decisions; can be given at the end of any class or seminar. Likert-type or open-ended questions. • Easy to prepare and administer • Provide specific and fast feedback on a class session • Limited number of questions • This session was useful to me: • Agree • Neutral • Disagree Open-ended questionnair es (satisfaction) B, T Sml/S mpl F/S Provide data about students’ perceptions in detail on a specific topic. Open ended questions. • Easy to prepare • Collect detailed information. • Suitable for smaller groups • It takes time to analyze responses to each openended question • Limited number of questions • What did you like the most (and least) about the materials used in this session? 5. Focused Assessment Resources Annotated, responsive to user needs Selected books and articles on assessment focusing on Engineering and Computer Science Education Online “HANDBOOKS” on Assessment and Evaluation Web Sites of Professional Organizations Offering Resources on Assessment and Evaluation Other Useful Evaluation and Assessment Web Sites Selected Books on Learning and Instructional Planning Online Journal on Assessment and Evaluation 6. Case Studies In Effective Assessment 6. Case Studies in Effective Assessment Mark Guzdial Professor of Computer Science Georgia Tech Example of Comprehensive Assessment Projects, adapted at many universities Human Subjects – consent letter Pre-course surveys Post-course surveys Interview protocols Grades Discussion Of Assessment Challenges Introduction Dr. Jane Prey and Dr. Jamie Cromack MSR, ER&P Open Discussion – Participants What are your and your colleagues’ important assessment challenges? MSR Assessment Toolkit http://research.microsoft.com/erp/AssessmentToolkit We welcome your input! © 2007 Microsoft Corporation. All rights reserved. Microsoft, Windows, Windows Vista and other product names are or may be registered trademarks and/or trademarks in the U.S. and/or other countries. The information herein is for informational purposes only and represents the current view of Microsoft Corporation as of the date of this presentation. Because Microsoft must respond to changing market conditions, it should not be interpreted to be a commitment on the part of Microsoft, and Microsoft cannot guarantee the accuracy of any information provided after the date of this presentation. MICROSOFT MAKES NO WARRANTIES, EXPRESS, IMPLIED OR STATUTORY, AS TO THE INFORMATION IN THIS PRESENTATION. Microsoft Research Faculty Summit 2007