The Performance of Albany High School on the California Academic

advertisement
The Performance of Albany High School on the California
Academic Performance Index, 2002-2009
By Ira Sharenow
November 1, 2011
Executive Summary...…………………………………………………………………..2
Introduction and Background Information…………………………………………...3
Performance of Albany High School on the API……………………………………...4
Statistical Analysis of the Data…………………………………………………………6
Conclusion……………………………………………………………………………….9
Summary Tables……………………………………………………………………….10
Bibliography……………………………………………………………………………11
Appendix A: List of Analyses Performed…………………………………………….12
1
Executive Summary
This report demonstrates that Albany High School (AHS) is underperforming, given the
high quality of its students. Moreover, the performance of the high school relative to the
rest of the state and relative to peer schools declined during the period of study, 20022009. Scores on the base Academic Performance Index (API) as computed by the
California Department of Education (CDE) are used as the measure of performance. The
quality of students is measured using demographic data. In particular, parental education
is the primary predictor.
For more technical readers, I performed an exploratory data analysis, including creating
multiple graphs, computing means, medians, Pearson correlation coefficients, and
Spearman correlation coefficients. Then I performed various linear regressions using the
R statistical programming language. Standard diagnostic tests were performed; outliers
and influential points were analyzed; Box-Cox transformations were performed; and
alternative models were considered. Mixed effects models were also analyzed. Then a
peer subgroup was selected and the analyses were performed anew. The results were
consistent across all models.
2
Introduction and Background Information
Each year California public school students take a battery of tests. The California
Department of Education (CDE), then weights the test scores, producing a composite
score for each school called the Academic Performance Index (API). This study uses the
base API scores as the measure of school performance. There are also “growth” scores,
but there is little difference between the two.
CDE converts the API scores using two distinct decile rankings. One is an overall
ranking. The second, and the one which is more relevant to this study, is a peer ranking
based on demographic data. In 2002, Albany High School (AHS) ranked in the top 10%
overall (decile 10) and in the top 30% (decile 8) in the CDE peer group. By 2009, AHS
fell to the top 20% overall (decile 9) and to the bottom 10% in its peer group (decile 1).
See Table 3 below.
This report takes an in depth look at the declining AHS performance, both with respect to
a subset of all schools and to a peer group based on schools which had similar
demographics in 2002.
In 2002, Albany voters elected several new school board members and they along with
the superintendent embarked on a policy of admitting large numbers of non-residents
from lower performing districts such as the West Contra Costa County Unified School
District.
The Albany Unified School District (AUSD) administration declined to provide separate
data for residents and non-residents, so I can only note that the decline in performance
occurred at the same time as non-district enrollment increased. Further research based on
more detailed data will be needed in order to determine if there is a cause and effect
relationship.
The next section describes the declining performance of Albany High School and uses
some graphs to clearly present the data’s message. Then a statistical analysis section
includes a more technical analysis and that section is followed by some example R code.
3
Performance of Albany High School on the API
There are several ways to see that Albany High School (AHS) has had a declining
performance and that its recent scores are well below the level they should be.
From 2002 to 2009, overall state scores increased 96 points, from 645 to 741, and peer
group scores increased 71 points, from 790 to 861. In contrast Albany High School’s
scores decreased 14 points, going from 820 to 806. See Chart 1.
A second way to look at the data is via ranks. Out of 677 schools statewide, in 2002 AHS
ranked 21 on the API but in 2009 AHS fell to a rank of 135. In the peer group of 50
schools, AHS fell from 12th place to 47th place. See Chart 2.
A third approach, and the main emphasis of the statistical analysis section, compared the
actual AHS score to its predicted score. In the first few years, AHS scored better on the
API than it was predicted to do and did so by as much as 30 points. Unfortunately by
2009, it scored 62 points below its predicted score, scoring 806, even though its predicted
score was 868 when parental education was the predictor and no adjustments were made
to the model. See Chart 3.
4
A more sophisticated technique, mixed effects for longitudinal data, provided further
evidence of the decline.
In conclusion, the analyses produced consistent results; Albany High School scored
below its potential and there has been a decline in recent years. The decline is associated
with a steep increase in transfers into the district from poorly performing districts such as
the West Contra Costa Unified School District. The Albany Unified School District needs
to produce data that will enable researchers to determine if the transfer policy is the cause
of the decline in Albany’s scores.
5
Statistical Analysis of the Data
1. Overview
Based on informal observations of the performance of Albany High School (AHS) on
California’s Academic Performance Index (API) combined with the knowledge that after
the November 2002 elections the Albany School Board embarked on an aggressive
campaign to admit large numbers of students from very poorly performing districts, I
hypothesized that there was a statistically significant and meaningful decline in the AHS
performance. Albany is a highly educated community that has a strong focus on student
education, so it is not surprising that AHS students continue to score well above the state
average on the API. However, as demonstrated by comparisons with peer groups,
conditional on the AHS demographics, in recent years AHS scored well below
expectations. Comparing the trend of AHS scores to the state’s and to the peer group’s
shows that the AHS trend is also negative. See Charts 1, 2, and 3 above.
The intent of this paper is to provide information that is useful to the general public and
which is then supported by statistical analyses. Consequently, most results use average
parental education (education level of the student’s most educated parent) as the single
predictor. By state law, California uses numerous predictors which are then summarized
by a School Characteristics Index (SCI). The state then uses the SCI to select 100 similar
schools and uses SCI as its main predictor.
Since the SCI is not an intuitive explanatory variable, CDE has performed some analyses
based on the three main predictors – average parental education (AVGED), percentage of
students on a government funded reduced price meals program (MEALS) and the
percentage of students who are English Learners (EL).
I did analyses using just AVGED and also using the three main predictors that CDE used.
While in some years the standard ANOVA for comparing a model and a subset of that
model produced a statistically significant difference in the models, for the purpose of
presenting findings to the public and for public policy purposes, the predicted scores for
AHS differed by an inconsequential amount. Another analysis was performed using SCI
as the predictor.
Moreover, Albany is atypical in that many English learners are children of very highly
educated people, including graduate students. Furthermore some recent studies question
the accuracy of the English Learners classification methods. A further examination of
Albany’s demographics is beyond the scope of this paper, but it does seem that based on
these subject matter considerations just using average education is a reasonable approach.
2. Methodology overview
For each year, 2002-2009, the standard normal regression modeling techniques were
used. The goal was to find a simple model that was capable of giving reasonably accurate
6
predictions for schools in the upper quarter. Various diagnostic tests were performed and
the Box-Cox transformation was utilized.
Finally, mixed effects models for longitudinal data were analyzed.
A subset of 50 peer schools was chosen based on schools with approximately the same
AVGED as AHS had in 2002. Then the analyses were repeated.
The results were consistent across all analyses – AHS performed poorly given its
demographics and its performance declined in recent years. Please see the table below.
3. The data
The data was downloaded from the CDE API database web site. The documents were
read into R and the data was merged into one large data frame. Only the 677 schools that
had at least 300 scores on the 2002 API (AHS had 576) and were not flagged for an
irregularity and which had an AVGED greater than 1.3 for all years, 2002-2009, were
included. AHS AVGED ~ 4.1.
AVGED is computed as follows. Each student’s most highly educated parent’s education
is scored as: Less than HS graduate = 1; HS graduate = 2; some college = 3; college
degree = 4; graduate degree = 5. Then the average is taken for the school.
My selection process eliminated many of the lowest performing schools but AHS is in the
top quarter of schools, so the selection process had little impact on the analysis. See
Table 3 for a comparison of the report’s descriptive data for the full dataset.
On numerous occasions in 2010, including API demographic data, the Albany Unified
School District submitted inaccurate data to CDE, the details of which are beyond the
scope of this report. Consequently 2010 API results were not included in this report.
No student level data was available to me. School level data was used.
4. The main normal model analyses
My starting point was the standard one predictor linear model with API the response
variable and average parental education (AVGED) the predictor. Separate analyses were
done for each year, 2002-2009. Scatter plots and other diagnostic tools produced
similarly shaped distributions of data. Therefore, year 2009 data will be the main
example data.
The scatter plot indicated that a straight line model was a reasonable starting point.
However, plots of residuals versus predicted values and qqplots indicated that there were
some very large residuals. I used several strategies to transform the data, including
7
omitting the most influential points or omitting the largest outliers. Additionally a BoxCox transformation was performed.
I also analyzed the three predictor model with average parental education (AVGED),
percentage of students on the government subsidized meals program (MEALS), and the
percent of English Learners (EL) as the three predictors.
Finally, I used the School Characteristics Index (SCI) as my explanatory variable.
The AHS predicted scores were similar across all methodologies and are summarized
below. See Table 1.
5. A closer look at the 2009 data
See Tables 1, 2, and 3 for summaries of the results.
The standard linear models and summary functions were run. There were several outliers
and they were sequentially removed. However, the AHS predictions for 2002 and 2009
did not change.
Then the analysis was redone with leverage as the statistic of interest. Cook’s distances
were computed. In this case I eliminated the points with the 14 largest distances, but the
2002 and 2009 AHS predicted scores only decreased by five and two points, respectively.
The Box-Cox transformations reduced the 2002 and 2009 AHS predicted scores by zero
and six points, respectively.
8
The three predictor model was also analyzed. Predictions were similar and the standard
F-test to compare a model to a sub-model was not significant at the 5% level for 2002
and was not significant at the 1% level for 2009.
Finally mixed effects models were analyzed and similar results obtained. The selected
model had AVGED as a fixed effects predictor and an interaction term, AVGED*Years.
Then I selected a peer group of 50 schools, which had similar AVGED scores in 2002. I
redid the analysis. The results were quite similar to the results for the full data set.
Conclusion
Multiple statistical methods demonstrate that Albany High School is underperforming on
the Academic Performance Index and in recent years its performance relative to all state
schools and to a peer group has declined.
9
Summary Tables
Table 1
All Data: AHS score versus predicted scores
Year
AHS
API
Predicted
(AVGED)
Predicted
(AVGED +
MEALS +
EL)
2002
817
800
796
779
795
800
800
801
2009
806
868
860
859
866
868
862
869
Predicted
(SCI)
Predicted
w/o
influential
Predicted w/o
outliers
Predicted BoxCox
Predicted
Mixed
Effects
Table 2
Peer Data: AHS score versus
predicted scores
Year
2002
2009
AHS
API
817
806
Predicted
(AVGED
Predicted + MEALS
(AVGED)
+ EL)
806
802
865
850
Table 3
AHS Deciles: For this report and as reported by CDE
2002
2009
CDE Report
Decile Decile
10
10
9
9
CDE
Peer
Decile
8
1
Report
Peer
Decile
8
1
10
CDE
Median
API Score
585
676
Report
Median API
Score
645
741
Bibliography
Bates, Douglas. (2010). lme4: Mixed-effects Modeling with R. http://lme4.r-forge.rproject.org/book/. Madison, Wisconsin.
Bates, Douglas. (2011). Short Course: Mixed models in R using the lme4 package.
http://lme4.r-forge.r-project.org/slides/2011-01-11-Madison/. Madison, Wisconsin.
California Department of Education API site. http://www.cde.ca.gov/ta/ac/ap/
Conover, W.J. (1999). Practical Nonparametric Statistics, Third Edition. John Wiley &
Sons. Hoboken, NJ.
Dalgaard, Peter. (2008). Introductory Statistics with R. Springer. New York, NY.
Faraway, Julian. (2005). Linear Models with R. Chapman & Hall, Boca Raton, FL.
Guo, S, Liu, R, Bernstein, R. (2009). Comparing Similar Schools Rank Changes:
Evaluating a Three Individual Variable Method. California Educational Research
Association Conference Presentation, San Francisco, CA.
Iman, Ronald and Conover, W.J. (1979). The Use of the Rank Transform in Regression.
Technometrics, vol 21, no 4.
Kutner, M, Nachtsheim, C, Neter, J, Li, W. (2004). Applied Linear Statistical Models,
Fifth Edition. McGraw-Hill Irwin, New York, NY.
Spector, Phil. (2008). Data Manipulation with R. Springer. New York, NY.
11
Appendix A: List of Analyses Performed
All Data
lm(API ~ AVGED)
lm(API ~ AVGED + MEALS + EL)
lm(API~ SCI)
Mixed Effects library(lme4):
lmer(API ~ 1 + AVGED + AVGED*Years + (1 | Years))
Peer Data
lm(API ~ AVGED)
lm(API ~ AVGED + MEALS + EL)
lm(API~ SCI)
Diagnostic and remedial techniques
Plots of residuals versus predicted
Plots of abs(residuals) versus predicted
qqplots
Analysis of outliers
Analysis of leverage points
Tests for normality of residuals
Tests for constant variance
F tests for model reduction
Box-Cox transformations
12
Download