MODULE 3 3rd 2nd 1st The Backward Design Learning Objectives • What is the purpose of doing an assessment? • How to determine what kind of evidences to look for? • What kind of methods can be used? When? • How to create assessment tasks and evaluation criteria? • How to make sure the assessment is valid and reliable? Why to assess? The assessment purpose is to measure understanding, not to generate grades!!!! Provide professor with: • Reliable information to infer about student learning • Feedback to improve their teaching methods Provide students with: • Feedback on how well they understand the content • Feedback to improve their learning How to create assessments? 1 Assessment Objectives 2 Evidences of Learning 3 Assessment 4 5 Evaluation Criteria Validity and Reliability 1 Assessment Objectives Worth to be familiar with Superficial knowledge Important to know and do Big Ideas Core Concepts 2 Evidences of Learning Related to the ABILITY of doing something EVIDENCE refers to something that can be DEMONSTRATED! Know Concepts, Definitions Worth to be familiar with Superficial knowledge Important to know and do Micro-descriptive Ability to apply a specified framework to contexts Big Ideas approached in class Core Concepts Micro, domain-specific Ability to transfer knowledge to different contexts Macro, across domains, multi-disciplinary 2 Evidences of Learning Bloom 2001 Judge results of concepts application and make decision about the quality of the application Apply concepts to situations different from the ones approached in class. Create new application or interpretation of the concepts Break concepts into parts and understand their relationship Apply concepts to situations similar to the ones approached in class Summarize ideas, explain concepts Recall Definitions Assessment 3 Assessment Tasks When to assess? Which Method? 3 When to assess? Snapshot Summative vs Photo Album Formative + Summative Formative and Summative Assessment 3 Both are necessary! At least 50% of each! Formative • Summative Objective is give feedback to • More focused on grade students • End of the grading period. • Build learning There is no opportunity to • Students can adjust adjust and show improvement Combination of both leads to a good result! F F S 3 Continuous Assessment Different Moments and Different Methods!! Assessment Tasks 3 Worth to be familiar with Superficial knowledge Important to know and do Big Ideas Traditional Quizzes and Tests • Paper-and-pencil • Multiple-Choice • Constructed response Core Concepts Performance and Task Projects Adapted from “Understanding by Design”, Wiggins and McTighe • Complex • Open-Ended • Authentic Assessment Tasks 3 Bloom 2001 Result of Analysis - Decision Pros vs Cons, Cost vs Benefits, Reflection Authentic Tasks Complex Performance Task Application to new contexts and situations, create artifact or project Analytical Task Experiments, Scenarios Simulation, Cases Simple Performance Task Straightforward application, Exercises Open-Ended Questions Quizzes and Traditional Tests Ask about definition 3 Authentic Task Task that reflects possible real-world challenges It is a performance-based assessment! • Is realistic contextualized • Replicates key challenging real-life situations • Requires judgment and innovation • Students are asked to “do” the subject • Assesses students ability to integrate concepts and ideas • Gives the opportunity to practice and get feedback It is problem-based NOT an exercise! From “Understanding by Design”, Wiggins and McTighe 3 Authentic Task vs. Exercise Authentic Task • Exercise Question is “noisy” and complicated • Various approaches can be used • Integration of concepts and skills • • • Right approach Appropriate solution • Right solution and answer Arguments is what matters • Accuracy is what matters Out of class, summative From “Understanding by Design”, Wiggins and McTighe In Class, Formative 3 How to formulate an Authentic Task? What is the goal of the task? What is the problem that has to be solved? oal What is the student role? What students will be asked to do? ole udience Who is the audience? Who is the client? Who students need to convince? ituation erformance tandards From “Understanding by Design”, Wiggins and McTighe What is the situation or the context? What are the challenges involved? Evaluation Criteria 4 Must… • Provide feedback for students • Be clear • Communicated in advance • Be consisted of independent variables • Focus on the central cause of performance • Focus on the understanding and use of the Big Idea 4 Types of Evaluation Criteria Criteria Check List 4 Check List There are two types of Check Lists 1. List of questions and their correct answers 2. List of individual traits with the maximum points associate to each of them 4 Check List: Questions and answers This type is used to Multiple-choice, True/False, etc. In other words, where there is a correct answer 1. 2. 3. 4. 5. 6. A C D B B D 4 Check List: Traits and their value Performance Trait 1 Weight (%) or points Trait 2 Weight (%) or points Trait ... Weight (%) or points Grade = weighted average or Grade = sum of points Analytic Rubric is better 4 • Provides more detailed feedback for students • Provides students with information about how they will be evaluated • Is clearer • Evaluates independently each characteristic that composes performance On the other hand… • Holistic Rubric is used when it is required only an overall impression 4 Analytic Rubrics How to create them? 4 How to create Analytical Rubrics? Example: a simple rubric to evaluate an essay Levels of achievement Traits Ideas Organization Grammar Excellent Satisfactory Poor 4 It can be created from a Check List! The difference is that each trait is broken down into levels of achievement, which have detailed description! Excellent Trait 1 Weight (%) or points Acceptable Unacceptable Excellent Performance Trait 2 Weight (%) or points Acceptable Unacceptable Excellent Trait ... Weight (%) or points Acceptable Unacceptable How to define traits? 4 It can be defined based on experience or on historical data: 1. Get samples of students’ previous work 2. Classify the sample into different levels (strong, middle, poor…) and write down the reasons 3. Cluster the reasons into traits 4. Write down the definition of each trait 5. Select among the samples the ones that illustrate each trait 6. Continuously refine the traits’ definitions It can also be defined based specific objectives and learning questions From “Understanding by Design”, Wiggins and McTighe 4 How to build Analytic Rubric? The following website is a free tool that helps to create rubrics http://rubistar.4teachers.org/index.php 5 Validity and Reliability 5 Validity and Reliability Target Desired understandings / objectives Shots Assessment Outcomes http://ccnmtl.columbia.edu/projects/qmss/images/target.gif 5 Checking for Validity Self-assess the assessment tasks by asking yourself the following questions: • Is it possible to a student do well on the assessment task, but really not demonstrate the understandings you are after? • Is it possible to a student do poorly, but still have significant understanding of the ideas? Would this student be able to show his understanding in other ways? If yes, the assessment is not valid. It does not provide a good evidence to make any inference (Note: for both questions, consider the task characteristics and the rubrics used for evaluation) Adapted from “Understanding by Design”, Wiggins and McTighe 5 Checking for Validity The previous questions can be broken down into more detailed questions: How likely is that a student could do well on the assessment by: • Making clever guesses based on limited understanding? • Plugging in what was learned, with accurate recall but limited understanding? • Making a good effort, with a lot of hard work, but with limited understanding? • Producing lovely products and performance, but with limited understanding? • Applying natural ability to be articulated and intelligent, but with limited understanding? Next Slide From “Understanding by Design”, Wiggins and McTighe Checking for Validity 5 How likely is that a student could do poorly on the assessment by: • Failing to meet performance goals despite having a deep understanding of the Big Ideas? • Failing to meet the grading criteria despite having a deep understanding of the Big Ideas? Make sure all the answers are “very unlike” !!! From “Understanding by Design”, Wiggins and McTighe Checking for Reliability 5 Assess rubric reliability by asking: • Would different professors grade similarly the same exam? • Would the same professor give the same grade if he grades the test twice, but at different moments? Assess task reliability by asking: • If a student did well (or poorly) in one exam, would he do well (or poorly) in a similar exam? Task reliability can be achieved by applying continuous assessments From “Understanding by Design”, Wiggins and McTighe Summary Observable, Demonstrable Learning Evidences Objectives of Learning Summary Evidences of Learning Time Summative Formative Assessment Tasks Assessment Task • Complexity depends on the desired level of understanding • Clear evaluation criteria (Rubrics) • Task and criteria must provide accurate and consistent judgments Learning Objectives • What is the purpose of doing an assessment? • How to determine what kind of evidences to look for? • What kind of methods can be used? When? • How to create assessment tasks and evaluation criteria? • How to make sure the assessment is valid and reliable? References • The main source of information used in this module is the following book Wiggins, Grant and McTighe, Jay. Understanding by Design. 2nd Edition. ASCD, Virginia, 2005. • Rubrics http://rubistar.4teachers.org/index.php