Mendon-Upton District District Determined Measures Dr. Deborah Brady dbrady3702@msn.com 1. What are DDMs? • Communicating your local expectations • Quelling anxiety • All data stays in the district • Evaluation consequence s are minimal 2. Assessment Development and Quality • • Which tests that we give now can we use or adapt? Defining Quality: • Alignment to CCSS • Rigor 3. K-12 DDMs; 2 DDMs per educator 4. Implementing, Collecting, Organizing and Storing What about “singleton” courses ? What about PE, Art, Music, technology? What about SISPs? Administrators? • Who gives the tests? • Are the teachers’ grades calibrated? • Who grades, stores, and analyzes the assessments? 5. Analyzing scores and determining L, M, H Impact Scores • • • • • What’s high growth? What’s a year’s growth? What’s low growth? Where do we store this data? How do we use the results in the evaluation process? What are your NEXT STEPS? 5. Analyzing scores and determining L, M, H Impact Scores 2. Assessment Development and Quality 1. What are DDMs? 4. Implementing DDMs, Collecting, Organizing and Storing What are your NEXT STEPS? 3. K-12 DDMs; 2 DDMs per educator Agenda 1. 2. 3. 4. 5. DDM Overview Defining Quality DDMS Developing Quality DDMS Examples of DDMS Storing and Scoring • On-line access to materials now at: www.ddmsbrady.wikispaces.com DDM Overview District Determined Measures DEFINITION DDMs are defined as: “Measures of student learning, growth, and achievement related to the Curriculum Frameworks, that are comparable across grade or subject level district-wide” TYPES OF MEASURES • Portfolio assessments • Approved commercial assessments • MCAS ELA 4-8; MCAS Math 4-8 • District developed pre- and postunit and course common assessments • Capstone projects DDM Quick Review Must measure growth, not achievement Growth equals one year’s progress Each educator will have at least 2 DDMs per year and 4 data points (L, M, or H) Teachers’ DDMs will be assessments of student work called a DIRECT Measure Most growth will be based on a pre-test before teaching and a post-test after teaching MCAS SGP for grades 4-8 for ELA and math ONLY can be used (not grade 10) Scores (100%) or Rubrics can be used to measure progress One measure must not be MCAS; it must be a District Determined Measure which can include local assessments, normed assessments (DRA, MAP, SAT). However, self-contained classroom teachers may use both ELA and math SGA if the district makes that decision. Some SISPs, administrators, nurses may have to have 1 INDIRECT Measure (a few who do not work with students may have 2 indirect). INDIRECT Measures are like SMART goals—attendance, graduation rates, MCAS Typical Timeline for DDMs and Impact Ratings Implementation Last Year District-wide training, development of assessments and piloting of 5 required DDMs (Primary ELA or math; Elementary ELA or Math, MS Math, HS Writing to Text plus un-tested area) June 2014 Report: List of DDMs from District plus any waivers granted by DESE. 2014-2015 Year 1 (1st Cohort) Non-waivered; scores are divided into H-M-and Low and stored locally Year 2 (2nd Cohort) Areas waivered by DESE based on June report 2015-2016 Second year data is collected for 1st Cohort. First year of DDM data is collected for the 2nd Cohort. 2016-2017 October 2016: First DDM rating of High Moderate or Low is given to the 1st Cohort. The impact rating linked to the educator’s EPIMS NUMBER is sent to DESE with the October 1 Report based on 2 years of data for two DDMs. A second year of data is collected for 2nd Cohort. Their Impact Rating will be calculated and sent to DESE by October 2017. Performance & Impact Ratings Performance Rating Ratings are obtained through data collected from observations, walkthroughs and artifacts • Exemplary • Proficient • Needs Improvement • Unsatisfactory Impact Rating ( October 1, 2016 for 1st Cohort) Ratings are based on trends and patterns in student learning, growth and achievement over a period of at least 2 years Data gathered from DDM’s and State-wide testing • High • Moderate • Low Performance and Impact Ratings Performance Rating Summative Rating Exemplary 1-yr Self-Directed Growth Plan 2-yr Self-Directed Growth Plan Proficient Needs Improvement Directed Growth Plan Unsatisfactory Improvement Plan Low Moderate High Rating of Impact on Student Learning Impact Massachusetts Department of Elementary and Secondary Education Rating 10 Defining Quality Assessments Assessments • What is needed? • 2 per educator per year • Chart of assessments and general descriptions • Quality Assessment Development • Collection and Approval (Form) – Criteria – Storing, organizing • Pre and Post testing – Collection of data – Analysis of results – Low, Moderate and High Quality Assessments • “Substantive” – Aligned with at least 2 standards of Frameworks – And/or local standards • • • • Rigorous Consistent in substance, alignment, and rigor Consistent with the District’s values, initiatives, expectations Measure growth (to be contrasted with achievement) and shifts the focus of teaching toward the individual learner’s change over time Quality Assessments Almost any kind of assessment can work Must be a “Substantive” assessment (DESE) Aligned with (at least 2) standards of Frameworks or 1 Power Standard And/or local standards Rigorous (appropriately challenging; locally defined) Consistent with K-12 DDMs in substance, alignment, and rigor Consistent with the District’s values, initiatives, expectations Measures growth (to be contrasted with achievement) How will you collect and organize DDMs? Sample Plan Part I SGP and Local Common Assessments* ELA Math Science Social Studies 12 CA/CA CA/CA CA/CA CA/CA 11 CA/CA CA/CA CA/CA CA/CA 10 CA/CA CA/CA CA/CA CA/CA 9 CA/CA CA/CA CA/CA CA/CA 8 MCAS SGP/CA MCAS SGP/CA CA/CA CA/CA 7 MCAS SGP/CA MCAS SGP/CA CA/CA CA/CA 6 MCAS SGP/CA MCAS SGP/CA CA/CA CA/CA 5 MCAS SGP/CA MCAS SGP/CA 4 MCAS SGP/CA MCAS SGP/CA 3 CA/CA CA/CA 2 CA/CA CA/CA 1 CA/CA CA/CA K CA CA Common Assessments or Locally Determined Singleton, Art, Music, Technology, PE by Grade or grade-spans Special Education Specialists, Co-Teachers, substantially separate Indirect Measures (IM) Central Office, Psychologist Administrators Principals, Assistant Principals, Superintendent, Coordinators Inclusion Co-Teachers Can “share” scores with General Ed Teachers Adaptation of the SMART goal process MCAS SGP Either ELA or Math 4-10 12 CA/CA 11 CA/CA 10 CA/CA 9 CA/CA 8 CA/CA Or 7 CA/CA 6 CA/CA 5 CA/CA 4 CA/CA If their students’ goals are substantially different, the assessments can be modified or can focus on the goal of inclusion 3 CA/CA 2 CA/CA 1 CA/CA K CA/CA Specialists Measure Goals Can develop K-12 rubric Speech PE Measure goal that has an impact on student growth Attendance College Applications Technology Growth Plus other direct measures PLUS Indirect Measures Planning and Organizing Handouts • DDM Educator Allocation by School or Department • DDM Cover Sheet for Each DDM • DDM Quality Checklist • MURSD Assessment Checklist…how to connect the work you are already doing? Other • Development of Data Teams • Self-assessment of assessment literacy skills Table Talk on Planning Developing Quality Assessments Five Requirements of DDMs (DESE) 1. Must measure growth 2. Employ a common administration procedure 3. Use a common scoring process 4. Translate these assessments to an Impact Rating (High-Moderate-Low) 5. Assure comparability of assessments within the school (alignment, rigor, validity). 20 Measures Growth:Student Growth Percentiles (SGP Calculation for individual students) 288 to 244/ 25 SGP 4503699 230 to 230/ 35 SGP 214 to 225/ 92 SGP Measures Growth: MCAS SGP ELA/Math for Grades 4-8 (not 10) Small Classes, Co-Taught Classes • The median SGP must be used when a teacher has 20 or more students (altogether) in a content area (ELA or mathematics 3-8) • Median SGPs for 8-19 students have validity and may be used if the district determines this is appropriate • More than one educator (a classroom teacher, a support teacher, a curriculum supervisor, and a supervising administrator) may be considered responsible for a content-area SGP. Different students may be in their respective rosters. Measures Growth: “Cut Scores” for MCAS SGP Typical growth One year’s growth Higher growth Lower growth 0 50 100 65 35 Classroom 40 Whole Grade Percent of students 60 Whole Grade Classroom Rigor • Bloom (next slide) • Hess (next slide) • DESE’s “Curriculum Embedded Performance Assessment” Rubric in the Model Curriculum Units DOK Hess: Rigor Relevance Matrix Cognitive Level 3 Complexity Depth + thinking Level 1 Recall & Reproduction Remember -Recall, locate basic facts, details, events Understand -Select appropriate words to use when intended meaning is clearly evident B Analyze M Not appropriate at this level -Specify or explain relationships -summarize -identify central idea -Explain, generalize, or connect ideas using supporting evidence (quote, example…) -Explain how concepts or ideas specifically relate to other content domains or concepts -Use context to identify meaning of word -Obtain and interpret information using text features -Use concepts to solve non-routine problems -Devise an approach among many alternatives to research a novel problem -Compare literary elements, terms, facts, events -analyze format, organization, & text structures -Analyze or interpret author’s craft (literary devices, viewpoint, or potential bias) to critique a text -Analyze multiple sources -Analyze complex/abstract themes -Cite evidence and develop a logical -Evaluate relevancy, accuracy, & completeness of information argument for conjectures PARCC -Brainstorm ideas about a topic Create Strategic Thinking/ Reasoning MCAS -Use language structure L (pre/suffix) or word relationships Apply (synonym/antonym) to determine O meaning -Identify whether information is O contained in a graph, table, etc. Evaluate Level 2 Skills & Concepts Level 4 Extended Thinking -Generate conjectures based on observations or prior knowledge -Synthesize information within one source or text -Synthesize information across multiple sources or texts GENERIC Rubric for CEPAs in Mass 1 Topic development 2 Little topic/idea development, organization, and/or details Little or no awareness of audience and/or task Evidence and Content Accuracy Little or no evidence is included and/or 3 4 5 Limited or weak topic/idea Rudimentary topic/idea Moderate topic/idea development, organization, development and/or development and and/or details organization organization Limited awareness of audience and/or task Basic supporting details Adequate, relevant details Simplistic language Some variety in language Use of evidence and content Use of evidence and Use of evidence and knowledge is limited or content is included but is accurate content is weak basic and simplistic relevant and adequate Full topic/idea development Rich topic/idea development Logical organization Careful and/or subtle organization Strong details Appropriate use of language Use of evidence and accurate content is logical and appropriate content is inaccurate Use of Visuals/Media Visuals and/or media are missing or do not contribute to the quality of the submission Visuals and/or media demonstrate a limited connection to the submission 1 Standards for English Conventions Errors seriously interfere with communication and Little control of sentence structure, grammar and usage, and mechanics Visuals and/or media are basically connected to the submission and contribute to its quality Visuals and/or media are connected to the submission and contribute to its quality 6 Visuals and/or media contribute to the quality of the submission in a logical and appropriate way Effective/rich use of language A sophisticated selection of and inclusion of evidence and accurate content contribute to an outstanding submission Visuals and/or media are carefully and strategically selected to enhance the content of the submission 2 3 4 Errors interfere somewhat with communication and/or Too many errors relative to the length of the submission or complexity of sentence structure, grammar and usage, and mechanics Errors do not interfere with communication and/or Few errors relative to length of submission or complexity of sentence structure, grammar and usage, and mechanics Control of sentence structure, grammar and usage, and mechanics (length and complexity of submission provide opportunity for student to show control of standard English conventions) Common Administration and Scoring: Calibrating • Equal Opportunity for All students: Floor and Ceiling Effects • Scoring Guides and Directions • Calibrating Scorers with Rubrics Rubric Cautions: Holistic Rubric Show Progress across a Scale, Continuum, Descriptors 1 3 4 No improvement in the level of detail. Modest improvement in the level of detail Considerable Improvement in the level of detail Outstanding Improvement in the level of detail One is true One is true All are true All are true * No new details across versions * There are a few details included across all versions * There are many examples of added details across all versions, * On average there are multiple details added across every version * There are many added details are included, but they are not included consistently, or none are improved or elaborated upon. * At least one example of a detail that is improved or elaborated in future versions * There are multiple examples of details that build and elaborate on previous versions *Details are consistently included in future versions * The added details reflect the most relevant and meaningful additions * New details are added, but not included in future versions. * A few new details are added that are not relevant, accurate or meaningful Details 2 * There are many added details, but several are not relevant, accurate or meaningful *The added details reflect relevant and meaningful additions Example taken from Austin, a first grader from Answer Charter School in Boise, Idaho. Used with permission from Expeditionary Learning. Learn more about this and other examples at http://elschools.org/student-work/butterfly-drafts 28 2 2 2 Preconventional Emerging Developing Beginning Expanding Ages 3-5 Ages 4-6 Ages 5-7 Ages 6-8 Ages 7-9 Relies primarily on pictures to convey meaning. Begins to label and add “words” to pictures. Writes first name. Uses pictures and print to convey meaning. Writes words to describe or support pictures. Copies signs, labels, names, and words (environmental print). Writes 1-2 sentences about a topic. Writes names and familiar words. . 2 2 2 Writes several sentences about a topic. Writes about observations and experiences. Writes short nonfiction pieces (simple facts about a topic) with guidance. Bridging Fluent Proficient Connecting Ages 8-10 Ages 9-11 Ages 10-13 Ages 11-14 2 Writes about feelings and opinions. Writes fiction with clear beginning, middle, and end. Writes poetry using carefully chosen language with guidance. 2 Writes organized nonfiction pieces (e.g., reports, letters, and lists) with guidance. 2 Begins to use paragraphs to organize ideas. 2 Uses strong verbs, interesting language, and dialogue with 1 guidance. 1 1 Begins to write organized fiction and nonfiction (e.g., reports, letters, biographies, and autobiographies). Develops stories with plots that include problems and solutions with guidance. Creates characters in stories with guidance. Writes poetry using carefully chosen language. Begins to experiment with sentence length and complex sentence structure. Varies leads and endings with guidance. Uses description, details, and similes with guidance. 1 Uses dialogue with guidance. Writes persuasively about ideas, feelings, and opinions. Creates plots with problems and solutions. Begins to develop the main characters and describe detailed settings. Begins to write organized and fluent nonfiction, including simple bibliographies. Writes cohesive paragraphs including reasons and examples with guidance. Uses transitional sentences to connect paragraphs. Varies sentence structure, leads, and endings. Begins to use descriptive language, details, and similes. Uses voice to evoke emotional response from readers. Begins to integrate information on a topic from a variety of sources. 2 2 2 2 1 1 1 1 Writes in a variety of genres and forms for different audiences and purposes independently. Writes short fiction and poetry with guidance. Writes a variety of short nonfiction pieces (e.g., facts about a topic, letters, lists) with guidance. Writes with a central idea. Writes using complete sentences. Independent 2 Writes organized, fluent, accurate, and in-depth nonfiction, including references with correct bibliographic format. 2 Writes cohesive, fluent, and effective poetry and fiction. Uses a clear sequence of paragraphs with effective transitions. Begins to incorporate literary devices (e.g., imagery, metaphors, personification, and foreshadowing). Weaves dialogue effectively into stories. Develops plots, characters, setting, and mood (literary elements) effectively. Begins to develop personal voice and style of writing. . Creates plots with a climax. 1 Creates detailed, believable settings and characters in stories. Writes organized, fluent, and 1 detailed nonfiction independently, including bibliographies with correct format. Writes cohesive paragraphs 1 including supportive reasons and examples. 1 Uses descriptive language, details, similes, and imagery to enhance ideas independently. 1 Begins to use dialogue to enhance character development. " Incorporates personal voice in writing with increasing frequency. ? J AP Rubric of Rubrics Prose Analysis (9 levels give students room to improve Holistic) 9-8 Answers all parts of the question completely. Using specific evidence from the work and showing how that evidence is relevant to the point being made. Fashions a convincing thesis and guides reader through the intricacies of argument with sophisticated transitions. Demonstrates clear understanding of the work and recognizes complexities of attitude/tone. Demonstrates stylistic maturity by an effective command of sentence structure, diction, and organization. Need not be without flaws, but must reveal an ability to choose from and control a wide range of the elements of effective writing. 7-6 Also accurately answers all parts of the question, but does so less fully or effectively than essays in the top range. Fashions a sound thesis. Discussion will be less thorough and less specific, not so responsive to the rich suggestiveness of the passage or precise in discussing its impact. Well written in an appropriate style, but with less maturity than the top papers. Some lapses in diction or syntax may appear, but demonstrates sufficient control over the elements of composition to present the writer’s ideas clearly. Confirms the writer’s ability to read literary texts with comprehension and to write with organization and control. 5 Discusses the question, but may be simplistic or imprecise. Constructs a reasonable if reductive thesis. May attempt to discuss techniques or evidence in the passage, but may be overly general or vague. Adequately written, but may demonstrate inconsistent control over the elements of composition. Organization is attempted, but may not be fully realized or particularly effective. 4-3 Attempts to answer the question, but does so either inaccurately or without the support of specific evidence. May confuse the attitude / tone of the passage or may overlook tone shift(s) or otherwise misrepresent the passage. Discussion of illustrations / techniques / necessary parts of the prompt may be omitted or inaccurate. Writing may convey the writer’s ideas, but reveals weak control over diction, syntax, or organization. May contain many spelling or grammatical errors. Essays scored three are even less able and may not refer to illustrations / techniques at all. 2-1 Fails to respond adequately to the question. May misunderstand the question or the passage. May fail to discuss techniques / evidence used or otherwise fail to respond adequately to the question. Unacceptably brief or poorly written on several counts. Writing reveals consistent weakness in grammar or other basic elements of composition. Although may make some attempt to answer the question, response has little clarity and only slight, if any, evidence in its support. Although the writer may have made some attempt to answer the prompt, the views presented have little clarity or coherence; significant problems with reading comprehension seem evident. Essays that are especially inexact, vacuous, and /or mechanically unsound should be scored 1. 0 A blank paper or one that makes no attempt to deal with the question receives no credit. Rubric from Sharon Kingston Criterion Referenced Rubric and Raw Scores or % of 100 4(25)= 100 x 4(22)= 88 x 4(18)= 72 4(15)= 25 60 x + 18 + 22 + x 15 = 80% Comparable Assessments Comparable across schools and within grades – Example: Teachers with the same job (e.g., all 5th grade teachers or all teachers of English 8) give the same assessment – Where possible, measures are identical • Easier to compare identical measures • But the district can determine whether or not these identical measures provide meaningful information about all students – Exceptions: When might assessments not be identical? • Different content (different sections of Algebra I) • Differences in untested skills (reading and writing on math test for ELL students) • Other accommodations (fewer questions to students who need more time) • NOTE: MCAS modifications and accommodations based on IEPs and 504s are appropriate 32 Comparability: Co-Taught Classes and Teacher of Record • In some co-taught classes or when teachers share in the results of the assessment because of additional time spent in a pull-out class, districts may need to determine the teacher or teachers responsible for content covered by statewide testing • Co-teachers may share equally if both co-teach all students or the district may determine one teacher as the teacher of record. Comparability: Roster Verification When DDM Results Have Been Calculated Educators confirm the accuracy of their rosters • Student must be on roster by October 1, and the student must remain on roster through last day of DDM testing. • Student must be present for 90% of instructional time. – DESE recommendation in Model Contract Language Direct Measures Direct measures of student learning, growth, or achievement provide information about student growth from student work. Most classroom teachers will have 2 direct measures More Examples CCSS Math Practices: A HS math department’s use of PARCC examples that require writing asking students to “justify your answer” SS Focus on DBQs and/or PARCC-like writing to Text: A social studies created PARCC exam using as the primary sources. Another social studies department used “mini-DBQs” in freshman and sophomore courses Common Criteria Rubrics for Grade Spans: Art (color, design, mastery of medium), Speech (developmental levels) Some GOOD Common Exam Examples • A Valued Process: PORTFOLIO: 9-12 ELA portfolio measured by a locally developed rubric that assesses progress throughout the four years of high school • K-12 Writing or Writing to Text: A district that required that at least one DDM was “writing to text” based on CCSS appropriate text complexity • Focus on Data that is Important: A HS science department assessment of lab report conclusions for each course (focus on conclusions) Indirect Measures for Guidance, Some Caseload Educators, Some Administrators Indirect measures of student learning, growth, or achievement provide information about students from means other than student work. These measures may include student record information (e.g., grades, attendance or tardiness records, or other data related to student growth or achievement such as high school graduation or college enrollment rates). ESE recommends that at least one of the measures used to determine each educator’s student impact rating be a direct measure and MCAS SGP if available and appropriate. The Process for Developing Indirect Measures • Consider the teams that you are a part of, for example, what many schools call the “child study team” in which many specialists participate, or all of your job-alike colleagues. • Discuss the goals that you have as a district group or as a school-based group. For example, you may be working on increasing the attendance of a group of students who are frequently out or tardy. Or, you may be working to return students to class quickly. Or, you may be focusing on working with a few families or students to support students with, for example, school anxiety. • Note that the measures can focus on a small group and do not have to include all students. For example, students with emotional problems that prevent them from participating fully can be your subgroup. For those educators with a small caseload, supporting your caseload to return to class and to participate fully in the class may be your goal. • Select a goal on something that is already an initiative or is recognized as something worthwhile for students. This goal needs to connect indirectly to student growth. Making sure students are in class and are able to participate fully is an appropriate goal. • Social-emotional growth is an appropriate direct measure. • Develop a method for measuring this goal. For example, nurses in one district realized that the software that they used could also track their goal of quick returns of students to classes. In another district, a schoolbased team decided that their “Student Support Team” had been reviewing numbers of classroom incidents and absences, but had not put aside time to study the most at risk students fully. Their goals was to change their weekly process so that time was set aside to discuss a student more thoroughly and to develop an action plan for that student. Their record was their weekly agenda, and they counted the increased number of action plans and tracked the progress of the most at risk students. • See DESE’s Short Brief on SISP DDMs Indirect Measures More Specifics: • Librarians: Increase the number of classes that work on research and projects in the library. Increase the number of teachers with whom you work to support specific units with materials. • Behavior specialists, speech pathologists, social workers, school psychologists: Increase the number of students who participate in class more fully. You can look at a sub-set of your students or caseload. For example, a behavior specialist was going to count the number of responses to a “non-classroom adult” that students made as they walked to gym or to lunch. This is a DIRECT MEASURE. • A group of speech pathologists used a developmental scale (like a holistic rubric) to measure the improvement of students’ speaking skills. They assessed these skills in their classroom and in general education classrooms. This is a DIRECT MEASURE. • If IEP goals measure student academic or social growth, attainment of success in working toward these goals can be used as DIRECT MEASURES. Thus, achieving IEP growth goals for my caseload can be a goal. Please note that DESE is still looking at subgroup sizes under 7 as being too small for growth measurement. • Guidance counselors set increasing the number of applications for college as a goal. Their comprehensive initiative included visiting classrooms and encouraging more students to take the PSAT, to work with the administration to provide SAT workshops, and to offer workshops through the school in writing college application letters. Indirect Measures – High school SST team example (Frequent Absentees) – Child Study Team example (Universal Process) – School Psychologists (Did not follow procedure for referral) – School Psychologists (subgroup of students studied) – High school guidance example (PSAT, SAT, College Applications) – IEP goals can be used as long as they are measuring growth (academic or social-emotional) Examples of DDMs The GOOD The BAD and the Ugly Quality Assessments, Developed Locally, Adapted, or Adopted Dr. Deborah Brady dbrady3702@msn.com History-Social Studies Read Like an Historian Table Vote Thumbs UP? Halfway? DOWN Morale after Fredericksburg This assessment asks students to source and corroborate a letter from a Union soldier describing low morale among Union soldiers after the Battle of Fredericksburg. Question 1 asks students to evaluate whether the source provides sufficient evidence to demonstrate the morale of the entire Union Army. To answer this question, students must source the document to determine whether the author represents all Union troops. Question 2 asks students to evaluate whether additional documents would corroborate the letter. https://beyondthebubble.stanford.edu/assessments http://sheg.stanford.edu/rlh Stanford University Reading History Like a Historian H-SS: Mini Q: What Caused the Dust Bowl? Overview: In the 1930s America was hit by very bad times. These were the years of the Great Depression. In cities and in small towns across the land, banks were failing, businesses were closing, and workers were being fired. But in some ways farmers were hit hardest of all, and few farmers were hit harder than those in the Southern Great Plains. This region has come to be known as the Dust Bowl. The question for this Mini-Q is what caused the Dust Bowl tragedy. The Documents: Document A: Dusters Document B: Grass Document C: Fred Folkers and his tractor (with photograph) Document D: Acreage Under Plow (chart) Document E: Rainfall on the Plains in the 1930s Table Vote Thumbs UP? Halfway? DOWN Demonstrating Growth (Galileo Example) Billy Bob’s work is shown below. He has made a mistake In the space to the right, solve the problem on your own on the right. Then find Billy Bob’s mistake, circle it and explain how to fix it. Billy Bob’s work ½ X -10 = -2.5 +10 = +10 Your work Thumbs UP? Halfway? DOWN Find the mistake provides students with model. Requires understanding. Requires writing in math _____________________________________________ ½ X +0 = +12.5 (2/1)(1/2)X =12.5 (2) X=25 Table Vote Explain the changes that should be made in Billy Bob’s Work Table Vote Science Lab Rubric—One Focus Thumbs UP? Halfway? DOWN Advanced • Assess student growth in writing conclusions for labs in all of the sciences Conclusions 1. Summarizes data used to draw conclusions 2. Conclusions follow data (not wild guesses or leaps of logic), 3. Discusses applications or real world connections 4. Hypothesis is rejected or accepted based on the data. Proficient 3 of 4 of the "excellent" conditions is met NI Warning 2 of the 4 1 of the 4 excellent excellent conditions met conditions met Learning Skills Criteria (Special Education) • Individual Goals; measured weekly; permanent folder – – – – – – – – – – Notes Planner Work/Action Plan Flexible when Necessary Prepared for Class (materials, work) Revises work Follows instructions Uses time well Gets to work Asks for help when needed – Advocates for self – Moving toward independence – Works collaboratively Table Vote Thumbs UP? Halfway? DOWN Table Vote Thumbs UP? Halfway? DOWN? Are numbers good or a problem? Scoring Guides from Text A scoring guide from a textbook for building a Lou Vee Air Car. Is it good enough to ensure interrater reliability? • Lou Vee Air Car built to specs (50 points) • Propeller Spins Freely (60 points) • Distance car travels – – – – 1m 70 2m 80 3m 90 4m 100 • Best distance (10,8,5) • Best car(10,8,5) • Best all time distance all classes (+5) Table Vote Thumbs UP? Halfway? DOWN? • 235 points total Example: Generic SISP Rubric for Team Process Indirect Criteria Improving assessment process for pK-2 referrals for Special Education (for quality/process improvement) Present Low Improvement Many students are referred to be tested for SPED without first providing classroom-based interventions. The process varies among all of the primary schools Moderate Improvement (what is hoped for) High Improvement (more improvement than expected) All Student Support Teams All referrals result only after will use a consistent process Level I interventions have been tried and assessed for The specialists and SPED at least 6 weeks. staff will provide consulting or workshops support the classroom RTI process Table Vote Thumbs UP? Halfway? DOWN PE Rubric in Progress. Grade 2 for overhand throw and catching. Table Vote Thumbs UP? Halfway? DOWN? Music: Teacher and Student Instructions Table Vote Thumbs UP? Halfway? DOWN? The Ugly Comply with regulations Bring about no change or understanding The Best • Build on what is in the District, school or department • A small step or a larger step in cognitive complexity • Use the results to learn about students’ needs and how to address these needs • Use time to look at student work, to collaboratively plan to improve Scoring and Storing Scoring and DDM Design Considerations • Districts will need to determine fair, efficient and accurate methods for scoring students’ work. • DDMs can be scored by the educators themselves, groups of teachers within the district, external raters, or commercial vendors. • For districts concerned about the quality of scoring when educators score their own student’s work, processes such as randomly re-scoring a selection of student work to ensure proper calibration or using teams of educators to score together, can improve the quality of the results. • When an educator plays a large role in scoring his/her own work, a supervisor may also choose to include the scoring process into making a determination of a Student Impact. • NOTE: Teacher Instructions may be necessary Sample Cut Score Determination (for 100 students) Student Scores Pre-test Post test Difference 20 35 15 5 25 30 5 15 30 50 20 20 35 60 25 25 35 60 25 40 70 40 Sorted low to high Teacher score is based on the MEDIAN Score of her class for each DDM Cut score LOW Growth Lowest ___% 25 median teacher score 35 25 median Teacher score 65 25 25 50 75 25 30 50 80 30 35 50 85 35 35 Top 20% Cut score HIGH GROWTH Highest ___? Measuring Growth Example: Fountas and Pinnell P to Q; N to P; D to K Fountas and Pinnell Growth for Each Student Is Based on 10 Months of Growth Second Grade Student Level Achievement Level End of Year Levels from beginning to the end of the year Pre-Post F&P Levels Growth HIGH, MODERATE, OR LOW GROWTH (10 MONTHS=YEAR) Q Above Benchmark PQ 7 MONTHS GROWTH LOW GROWTH P At Benchmark NOP 10 MONTHS OF GROWTH MODERATE GROWTH K Below Benchmark DEFGHIJK 17 MONTHS OF GROWTH HIGH GROWTH Teacher A Teacher B Teacher C Teacher D Teacher E Teacher F 103 Third Graders All Classes 1 3 3 3 5 7.5 6.5 3.5 9 3.5 5.5 6 6.5 6.5 5.5 6.5 6.5 6.5 6.5 7 10 12 1 3 3 6 6.5 7.5 8.5 10 10 10 10 10 6.5 7.5 8.5 9 10 10 10 7 7 7 7 10 10 11.5 10 10 10 10 12 12.5 12.5 6.5 10 10 12.5 12 6.5 6.5 6.5 6.5 6.5 6.5 6.5 6.5 Cut Score lowest 15% 7 7 7 7 7 7 7 7 7.5 7.5 8.5 8.5 9 9 9 9 9 9.5 10 10 10 None 7 7 12 12 10 12 12 16 13 13 9 9 10 10 12 12 12 12 13 13 13 16 17 17 12 13.5 13.5 13.5 17 16 16 16 16 16.5 16.5 16.5 13.5 13.5 13.5 13.6 15.5 19 10 10 10 10.7 10.9 11.0 11.2 11.4 11.5 11.7 11.9 12.0 12.2 12.5 12.6 12.7 12.9 13.1 13.2 13.4 12.0 Median for whole Grade 3 DDM 13.6 13.8 13.9 14.1 14.3 14.4 14.6 14.8 15.0 15.1 15.3 12 16 Median 6.5 Median 9 15.5 17 17 Median 12 Median 10 Median 16 15.6 12.5 12.5 12.5 13 13 Median 12 13 13 13 13.5 13.5 13.5 13.5 Below 6.5 Between 6.5 Between 6.5 Between 6.5 and 16 and 16 and 16 Between 6.5 and 16 Between 6.5 and 16 13.5 13.5 13.6 15.5 16 16 16 16 16 16 16 16.5 16.5 16.5 LOW Moderate Moderate Moderate High Moderate 16.5 16.5 17 17 17 19 26 16 cut score highest 15% Most Impact Ratings are M’s (K Writing) Teacher A Teacher B Teacher C Teacher D Teacher E 2 0 2 1 1 2 2 2 2 2 3 2 3 2 2 3 2 3 2 2 3 2 3 2 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 Median 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 4 3 3 3 3 4 3 3 3 3 4 3 3 4 4 4 4 4 4 4 4 4 4 Post Test Only Use District’s History to predict scores on AP Calculus Exam, for example Previous Grade’s Math Score Low Growth Moderate Growth High Growth A 3 4 5 B 2 3 4 C 1 2 3 1 2 D Considerations for Scoring Student Work • Districts will need to determine fair, efficient and accurate methods for scoring students’ work. (Use consistent directions for teachers.) • DDMs can be scored by the educators themselves, groups of teachers within the district, external raters, or commercial vendors. • For districts concerned about the quality of scoring when educators score their own student’s work, processes such as randomly re-scoring a selection of student work to ensure proper calibration or using teams of educators to score together, can improve the quality of the results. • When an educator plays a large role in scoring his/her own work, a supervisor may also choose to include the scoring process into making a determination of a Student Impact. Summary DDM Process for Determining L, M, H for Every Teacher Whole Grade Level or Course • Score the entire grade level or course or take the MCAS Growth Scores for all students • Rank the scores from highest to lowest (post minus pre or MCAS SGP) • Identify the median score for the entire group • Determine the “cut” scores for local assessments; MCAS 35 and 65 for classrooms Individual Teacher • Select students for each teacher • Rank the scores from highest to lowest • Identify the Median score • Is the median below or above the “cut” score? Is it in the middle? • Don’t forget Roster Verification might change the specific scores and, therefore, change the Median • Distribute scores to teachers for each DDM Next Steps in Storing • Gather data/tests themselves for analysis – MCAS SGP on line – Local Assessments – Transfer to Excel • Students • Pre and Post or MCAS SGP or Post test only • Cut scores • Determine whole group’s median • Determine teacher’s median • Roster Verification • Determine Low, Medium, or High Growth score for each teacher for each DDM