Rater Reliability

advertisement
Washington State
Teacher and Principal
Evaluation Project
Preparing Educators for Rater Agreement and
Feedback: Planning for Professional Learning
TPEP Sessions for 2014-15

Face-to-Face Sessions




A Virtual Presentation

2
Student Growth 2.0
Rater Agreement Practices
TPEP/ Washington State Learning Standards
Connections
Sharing Electronic Resources
Group Norms
Pausing
Paraphrasing
Posing Questions
Putting Ideas on the Table
Providing Data
Paying Attention to Self and Others
Presuming Positive Intentions







3
TPEP Core Principles
Professional learning is the underpinning of the
evaluation system.
4
Learning Targets:
I can describe the OSPI working definition of rater
agreement and the stages for development.
I can identify best-practices and strategies to maximize
rater agreement.
I can explain what makes a high quality professional
learning process for evaluators and helps administrators
develop strong skills in providing feedback.
I can describe our district’s plan to maximize rater
agreement.




5
Activity 1: District Self-Assessment
Discuss: Using the survey tool, briefly discuss with your
district team to determine the current level of practice in
maximizing rater agreement.
Share: What is your districts identified area of focus?


6
Learning
I can describe the OSPI working definition of rater agreement
and the stages for development.
7
Rater Agreement
Turn and talk to a partner about the following topics:



8
I think rater agreement is important because…
My role in the rater agreement process is…
OSPI Definition of Rater Agreement
 The
extent to which the scores between
the raters have consistency and
accuracy against predetermined
standards. The predetermined standards
are the instructional and leadership
frameworks and rubrics that define the
basis for summative criterion-level scores.
9
Unwrapping the Definition…

Consistency: A measure
of observer data quality
indicating the extent to
which an observer is
assigning scores that agree
with scores assigned to
the same observation of
practice by another typical
observer.
10

Accuracy: A measure of
observer data quality
indicating the extent to
which an observer is
assigning scores that agree
with scores assigned to
the same observation by
an expert rater; the extent
to which rater’s scores
agree with the true or
“correct” score for the
performance.
Take “Five”

Discuss with your team:



What systems are in place for your district to focus on
accuracy?
Are these systems dependent on each other or could a district
start with agreement and then work towards accuracy?
Share: What systems have you identified?
11
Learning Content 1: OSPI Stages of
Rater Agreement
12
OSPI Stages of Rater Agreement


Read “Rater Agreement in Washington State’s Evaluation
System”
Discuss with your team:




How has your district communicated the vision of Rater
Agreement?
What structures are in place for staff to learn about the
Instructional Framework?
How do teachers and building principals use the language of
the frameworks to improve practice?
Share: One strategy that you have identified.
13
Continuous Improvement Process:




Once we reach Stage 3, we do not “graduate.”
Rater agreement is NOT ensured by a single training or
certification test.
We must revisit and review key learnings from Stage 1
and Stage 2 to avoid drift.
Provide ongoing professional development including
ongoing calibration conversations involving real-life or
video-based observation.
14
Rater Drift will naturally occur unless
evaluators have:



Ongoing opportunities to demonstrate agreement
Access to practice videos for difficult-to-score
domains/components
Expectations that their ratings will be monitored
15
How is Rater Agreement Assessed?

Rating through observations requires evaluators to do a lot of
things well. They must:





16
Be objective: Record evidence that is free of “bias, opinion, and
subjectivity” (Bell et al., 2013)
Understand the rubric: Thoroughly and deeply understand each
component and indicator on the district rubric instrument (Bell et
al., 2013)
Interpret evidence: Correctly align evidence to framework criteria
that reflect the context of the evidence (Bell et al., 2013)
Document evidence: Gather, sort, and record a preponderance of
evidence as it happens in the classroom or school (McClellan,
Atkinson, & Danielson, 2012).
Be reliable: Reliability requires that they be able to do all of these
things well over-time, in different classrooms and schools, and for
different educators.
Professional Learning Takeaways
Ensure that observers have opportunities to learn the
following:

Use the rubric language to explain their decisions.

Consistently take notes that gather useful evidence.

Avoid making scoring decisions during note-taking.

Resort back to the scoring criteria when uncertain.

Practice using videos and live classroom observations.
(Bell et al., 2013)
17
THINK – PAIR - SHARE


What did you learn that confirmed your understanding of
Rater Agreement in Washington State’s Evaluation
System?
What did you learn that challenged your understanding of
Rater Agreement in Washington State’s Evaluation
System?
18
Learning
Understand best-practices and strategies to maximize rater
agreement.
19
Why Rater Agreement?

An educator’s observation
scores should be the same
regardless of the observer.



20
Educators can trust the new
evaluation system.
The new system is legally
defensible for personnel
decisions.
Educators receive relevant,
useful feedback for
professional growth.
Two Research-based Rater Agreement
Activities
1.
2.
21
Address issues of rater bias,
interpretation, and objectivity (Jilliam,
Kosa,Tierney, & Tocci, 2014; MET, 2012).
Rater agreement activities should be
provided at least three times a year (Jilliam,
Kosa, Tierney, & Tocci, 2014), but monthly is
ideal (Jerald, 2011).
What Affects Agreement in Observation?
Observer Bias: What are
some of the various “lenses”
that might bias
 A teacher evaluator?
 A principal evaluator?
2222
Turnkey Activity: Common
Sources of Bias
Handout 3: Common Sources of Bias Match-Up
 Step 1. At your table, work as a group to match each
common rater error with a possible strategy evaluators can
use to avoid the error.
 Step 2. After each match, discuss other possible strategies
you have seen used or that you think might be effective.
2323
Debrief
Answer Key
1=D
2=G
3 =A
4=F
5=B
6=E
7=H
8=C
2424
Two Research-based Rater Agreement
Activities
1.
2.
25
Address issues of rater bias, interpretation,
and objectivity (Jilliam, Kosa, Tierney, & Tocci,
2014; MET, 2012).
Rater agreement activities should be
provided at least three times a year
(Jilliam, Kosa,Tierney, & Tocci, 2014),
but monthly is ideal (Jerald, 2011).
Rater Agreement Activities





Rater agreement activities can be as long as several days
of training and as short as 60-minute sessions. It’s
important to note that frequency is the key to conducting
the activities. Short and frequent could be just as
effective as lengthier activities done only a few times per
year. 60-minute rater agreement activities can be
embedded in activities that already exist:
Monthly PLCs
Walk-throughs
Instructional rounds
Professional development
26
Selecting Observation Focus for
Data Collection

Scaffolding Observations:




27
Group1: Management
Group 2: Environment
Group 3: Engagement
Count off table’ by “3’s” – be ready to gather evidence
on the video based on your number
Video 1: Collect Data



Ms. Warburton’s 8th grade math lesson: Sorting and
classifying equations
CCSS Math 8.EE.C.7a
Collect evidence statements
https://www.teachingchannel.org/videos/sortingclassifying-equations-overview
28
Video 1:Share Data
29
Data: Align with Framework

Let’s all look at evidence that aligns with:
Group
Danielson
Marzano CEL
1
2
3

At tables: Based on the evidence we have compiled, talk
about the level of performance that should be assigned.
30
Rater Agreement Activities

All opportunities should be grounded in common
protocols or structures that assist in developing and
extending the common language of the observation tools
and other measures of educator practice.
(Jilliam, Kosa, Tierney, & Tocci, 2014
31
Research-based Rater Agreement
Activities

Discuss with your team:



How has your district included ongoing discussions to address
bias in the evaluation?
What job-embedded strategies are in place for evaluators to
practice specific skills?
Share: One strategy that you have identified.
32
Learning
High quality professional learning process for evaluators in
providing feedback
33
“The post-conference cannot be treated
as a bureaucratic formality; it is one of
the most critical features of an effective
teacher evaluation system if the goal is
not just to measure the quality of
teaching, but also to improve it.”
~Jerald and Van Hook,
2011, p. 23
34
Avoid Dominating
the Conversation


•
A study of evaluation implementation in Chicago found that principals
generally dominated the conversation by speaking 75 percent of the
time in postobservation conferences (Sartain, Stoelinga, & Brown, 2011).
Encourage a balanced conversation (50/50) by asking reflective and follow-up
questions.
Ensure that teachers are prepared to participate and establish this as an
expectation through educator orientation for the new evaluation system.
35
Focus on Evidence
Reduces three big dangers in postobservation
conferences:
 Loose interpretation. Evidence-based feedback separates
observations and interpretations.
 Subjectivity. Drawing upon evidence during feedback
conversations can decrease subjectivity (Sartain et al., 2011).
 Emotion. Evidence-based feedback can also “remove some of
the emotion from the evaluation process” (Sartain et al., 2011,
p. 23).
36
Use Rubric Language and Descriptors
Incorporating rubric language when discussing evidence
helps in the following:
 To build and reinforce a shared understanding of
good instruction
 To ensure the rubric remains the objective point of
reference in the conversation
37
High-Level Questioning
In Chicago, researchers found that only 10 percent of questions asked by
evaluators during postobservation conferences were high level and
promoted discussions about instruction.
(Sartain et al., 2011)
38
Rubric
Example
Low
The evaluator’s question
• Requires limited teacher response—often a single word—rather than
discussion
• Is generally focused on simple affirmation of principal perception
“I think this was basic
because of the evidence I
collected. Do you agree?”
Medium
The evaluator’s question
• Requires a short teacher response
• Is generally focused on completion of tasks and requirements
“Which goals did you not
meet?”
High
High-Level Questioning
Rubric
The evaluator’s question
• Requires extensive teacher response
• Reflects high expectations and requires deep reflection about
instructional practice
• Often prompts the teacher and evaluator to push each other’s
interpretations
“How did student
engagement change in
your class after you that
strategy? Why do you
think that happened?”
(Modified from Sartain et al., 2011, p. 24)
39
Ends With Actions and Supports
Professional Growth Planning
 Connect feedback to professional
growth plan.
 Identify goals, timelines, and
benchmarks for areas for growth.
Action Strategies, Practice, and
Modeling
 Ensure the conversation culminates in
small, specific changes a teacher can
implement in the classroom immediately.
 Have the teacher practice or model the
practice.
 Suggest observing a colleague who strong
in the area
 Direct the teachers to additional
resources (online, print, or other
colleagues).
(Hill & Grossman, 2013)
40
Activity: A Good Conference or Not?
http://vimeo.com/89454466
41
Helpful Resources for
Coaching and Feedback




Learning-focused Supervision: Developing
Professional Expertise in Standards-Driven Systems
(Lipton & Wellman, 2013)
Principal Evaluator’s Toolkit for the Instructional
Feedback Observation (American Institutes for
Research, 2012)
Leveraging Leadership: A Practical Guide to Building
Exceptional Schools (Bambrick-Santoyo, 2012)
The Art of Coaching: Effective Strategies for School
Coaching (Aguilar, 2013)
42
Research-based Coaching & Feedback
Activities

Discuss with your team:



What strategies have you found successful in building a culture
that embraces feedback and reflective conversations about
teaching and learning?
What job-embedded strategies are in place for evaluators to
practice specific skills?
Share: One strategy that you have identified.
43
Resources
.
44
Framework Specific Tools
45
“Video” Protocol Example

North Mason School District



Viewed Video as group and collect Data/Evidence
Sorted and Placed Data/Evidence on framework
Determined Level of Performance



First Individually
With a Partner
Whole Group of Administrators (5-6)
North Mason utilized this process for two years and then in year
three they did classroom observations. They utilize “TeachScape”
Videos.
46
“Instructional Reviews”
Protocol Example

Central Kitsap School District






47
Observational tool created to focus the observation on the
framework: Purpose, Student Engagement and Academic Rigor
Teams entered one school and divided themselves
Observed in 15 minute time blocks
Met and talked about their ratings
Combined ratings into one common rating sheet
Checked for range and variations
eVal

A collection of 93 videos* for:



48
Exploring your instructional framework in the context of
classroom instruction…
Discussing the characteristics of quality instruction as
described by your framework…
Engaging groups in interpretation, analysis and discussion of
classroom practice with the use of tools and rubrics.
Swivel Video Camera

Tool to create internal video for use:
49
Additional Resources:

The URL for eVal Training Video: Debbie Tschirgi ESD112
http://vimeo.com/101444062

The URL for the statewide Ed Tech web site where you
can get the logins and URL for the eVAL Sandbox:
http://www.edtech.wednet.edu/eVALTraining/
50
Implementing
Develop long-range professional development based on the
district’s current level of practice.
51
Plan for Rater Agreement




How will new evaluators receive training in the
framework prior to evaluating staff?
How will new evaluators learn to apply the framework in
a formative process during their first year?
How can existing meetings or professional learning
communities be used to provide opportunities to
maximize rater agreement?
What opportunities will be provided to trained evaluators
to continuously hone their skills?
52
Rater Agreement: District Planning Tool



Initiating
Refining
Sustaining
53
District Planning Tool
Each stage of Rater Agreement Includes:




Assessment
Planning
Implementing
Reflection on Measuring
54
Learning Activity: Monitoring and
Maintaining Rater Agreement
 Read “High Fidelity:
Investing in Evaluation
Training” (15 minutes)


55
How are states
providing training for
new teacher and
principal evaluation
systems?
How is your district
providing support to
monitor and maintain
rater agreement?

Choose one strategy to
add to your District
Planning Tool.
Reflecting
56
Revisiting the Learning Targets

With a Partner:




57
Describe the OSPI working definition of rater agreement and
the stages
Identify a best practice or strategy that could be used to
maximize rater agreement.
Explain a strategy that would help administrators develop skills
in providing feedback.
Discuss how your district plan includes elements listed above.
What’s Next

Homework options:



58
I. Pilot with Administrative Team
II. Develop a Guidebook
III. Developing a Training Protocol
Thank you!
Presenter Name
XXX-XXX-XXXX
xxxxxxxxxxx@xxx.xxx
1234 Street Address
City, State 12345-1234
800-123-1234
59
Download