Telecharger un fichier ppt gratuit : Understanding Value-Added

advertisement
Understanding Value-Added
Lesson 2: How Value-Added Works
Office of Accountability
Recap: What is Value-Added?
 Value-Added is the District’s measure of elementary school and teacher
growth.
 Value-Added is a nationally recognized way of measuring growth.
Academic Growth = Student Learning
2012
2013
 Emphasizes continual student improvement
 Provides information to understand what drives continual
improvement
Office of Accountability
2
Measuring Growth, Not Attainment
In this school, the percent meeting state
standards is 25% in both Year 1 and Year 2.
Attainment is unchanged – but are students
learning?
210
220
230
240
250
Scale
Score
200
210
220
Office of Accountability
230
240
250
Meets State Standards
200
Analyzing growth provides this
information
(Year 2)
260
270
280
290
300
(Year 1)
260
270
280
290
3
300
Accounting for Student Populations
 Student academic growth varies by grade, prior performance, and
demographics.
 The goal of the Value-Added metric is to measure the school or teacher’s
impact on student learning independent of student demographic factors.
 Value-Added accounts for the following student factors:
Prior Reading Score
Low-Income Status
Prior Math Score
ELL Status
Grade Level
IEP Status
Gender
Homelessness
Race/Ethnicity
Mobility
 Controlling for the factors above gives proper credit for growth to low
attainment schools and schools that serve unique populations.
Office of Accountability
4
How it Works
 Value-Added is not a comparison to similar schools.
‐ We do not look for a comparison group of schools that match each other
on all 10 student factors…such a group might not exist.
 Rather, Value-Added compares growth of students in each school
to growth of students across the District, controlling for the list of
student factors.
 To do this, we utilize a regression methodology, developed in
collaboration between CPS and academic experts from the
University of Wisconsin.
Office of Accountability
5
What is Regression?
 By measuring the impact of each student factor, the regression
model isolates the impact of the teacher on student growth.
 In other words, some growth is explained by external factors. We
can measure the average impact of these external factors on
growth at the District level and subtract that impact from the
teacher’s absolute growth.
 The growth that is left over after removing the impact of these
factors is attributed to the teacher. This is the value added by the
teacher.
Office of Accountability
6
For More on Regression…
 Two other presentations on this topic are available at
http://cps.edu/Pages/valueadded.aspx
 For an illustrative example of regression, view the “Oak Tree
Analogy” presentation. This presentation illustrates the ValueAdded model by using an analogy of two gardeners tending to oak
trees.
 For technical details, view “Lesson 3: Technical Specifications of the
Value-Added Regression Model”
Office of Accountability
7
Some Things to Know

Tested Students
•

Mobile Students
•
•

Mobile students count towards the Value-Added score in each school they attended, but are
weighted in the analysis by the amount of time they were in the school during the year.
At the teacher-level, mobile students count towards the Value-Added score for each teacher
that provided instruction to that student, but are weighted in the analysis by the time they
were in the school and the amount of instruction provided by each teacher.
English Language Learners
•
•

All students making normal grade progression who took ISAT or NWEA MAP in both the
previous year and current year are included in analysis.
For ISAT: ELL students in Program Years 0 through 5 are excluded from the analysis.
For NWEA MAP: Students with an ACCESS literacy score below 3.5 are excluded.
Students with Disabilities
•
•
IEP status is differentiated by type of IEP.
For example, the impact of a severe and profound disability is considered separately from the
impact of a speech and language disability.
Office of Accountability
8
Value-Added Scores
Value-Added measures the difference between the growth of
students for whom a school or teacher provided instruction and the
growth of similar students across the District.
A positive score indicates a school or teacher whose
students are growing at a faster pace than similar
students.
Zero (0) is the District average. A score near zero
indicates a school or teacher whose students are
growing at about the same pace as similar students.
A negative score indicates a school or teacher whose
students are growing at a slower pace than similar
students.
Office of Accountability
9
Standardization of Scores
 Growth is measured in scale score points (for NWEA, these are
called “RIT” scores).
Student A “grew” by 35 scale score
points
200
210
220
230
240
 However, one scale score point of growth is more difficult to obtain
in some grade levels than others.
 As a result, standardization is used to ensure that all Value-Added
scores are on the same scale.
Office of Accountability
10
Standardization of Scores
 Standardization is a common statistical process. In this case, it is
used to convert scale score points to a standard scale.
 The unit of measure is the “standard deviation” which is a measure
of distance from the mean.
‐ i.e., how much does School A’s score deviate from the mean?
 This places all scores on the same scale, allowing for more precise
comparisons between scores at different grade levels.
Office of Accountability
11
The Standard Scale
Features of the Standard Scale
 Zero (0) is the District average.
 About 68% of scores fall between -1 and 1.
 About 95% of scores fall between -2 and 2.
 About 99% of scores fall between -3 and 3.
 Only about 1% of scores are less than -3 or more than 3.
34%
2.5%
Office of Accountability
13.5%
34%
13.5%
2.5%
12
Reading the Value-Added Reports
(School-Level Report)
Value-Added Score
Number of Students in the
calculation
Office of Accountability
Percentile: This is the percent of
scores that fall below this score.
Percentiles range from 0th to
99th
Confidence Interval: This is
explained in the next set of slides.
13
Confidence Intervals
 The Value-Added model controls for factors that CPS can measure,
but there are some factors that cannot be measured, such as:
‐ Motivation to learn
‐ Family circumstances
‐ Health
 In addition, the Value-Added model is a statistical estimation of the
school or teacher’s impact on student learning and therefore
contains a certain amount of random error.
 For these reasons, the Value-Added model includes confidence
intervals.
Office of Accountability
14
Real World Example: Political Polling
A Political Polling company surveys a representative random sample of
1,000 community households about for whom they are going to vote on
Election Day. The question they pose is:
If the election were held today, for whom would you cast your ballot?
The percentages of responses breakdown as follows:
 Candidate Jones would receive 54% of the vote
 Candidate Smith would receive 46% of the vote
 There is a +/- 3% margin of error
Office of Accountability
15
Confidence Intervals in Political
Polling
With the margin of error of +/- 3%, the range of the percentage of people who plan on
voting for each candidates is as follows:
Candidate Jones would receive between 51% and 57% of the vote.
43%
44%
45%
46%
47%
48%
49%
50%
51%
52%
53%
54%
55%
56%
57%
Candidate Smith would receive between 43% and 49% of the vote.
The confidence intervals do not overlap. Therefore the race is NOT “too close to call.” We can predict
with a high degree of confidence that Candidate Jones will win the race.
Office of Accountability
16
Confidence Intervals in Value-Added
 A confidence interval is a range of scores around the Value-Added estimate.
The Value-Added estimate is 1.0.
 The confidence interval is ± 0.3.
 The confidence interval range is from 0.7 to 1.3.
 The district average (0) is not in the confidence interval,
so we are 95% confident that the school’s effectiveness
is different than the average (above average in this
example).

Example:
0.7
1.0
1.3
 We are 95% confident that the true Value-Added score falls within the
confidence interval range.
 The confidence interval is “n” dependent, meaning larger samples yield smaller
confidence intervals.
‐ This is because in larger samples, a score that is different from the average
is less likely to be due to random error alone.
Office of Accountability
17
Statistical Significance
 If the confidence interval does not include zero, we say that the score is
statistically significant, meaning we are 95% confident that the score is different
from zero.
 A color is associated with each score based on the statistical significance:
Green
• Entire confidence
interval is above
zero.
• Score is positive
and statistically
significant at the
95% confidence
level.
Office of Accountability
Yellow
• Confidence
interval includes
zero.
• Score is not
statistically
different from
zero at the 95%
confidence level.
Red
• Entire confidence
interval is below
zero.
• Score is negative
and statistically
significant at the
95% confidence
level.
18
How Confidence Intervals are
Reported
This is how Value-Added scores are displayed in the school-level
This school has a Value-Added score of
reports.
-0.5 in reading
(the score is ½ of a standard deviation
below the mean)
The confidence interval ranges from -1.9 to 0.8
Because the confidence interval includes zero, we say that this school is not
statistically different from zero at the 95% confidence level.
For that reason, the bubble is yellow.
Office of Accountability
19
For More Information
 More lessons and other resources for understanding ValueAdded are available at:
http://cps.edu/Pages/valueadded.aspx
 Lesson 2 (Part 2): Oak Tree Analogy
 Lesson 3: Technical Specifications of the Value-Added Regression Model
Office of Accountability
20
Download