Microsoft PowerPoint - the NCRM EPrints Repository

advertisement

Information for parents on the effectiveness of

English secondary schools for different types of pupil

Lorraine Dearden, John Micklewright and

Anna Vignoles

Institute of Education, University of London

Motivation

• DfE provides information on schools and pupil achievement in a number of ways, including raw scores

• DCSF also measures school performance with a contextualised value added model, which takes account of the different pupil intakes of schools (Ray,

2006)

– better guide to school effectiveness than raw GCSE scores, which capture differences in school intake characteristics

• But evidence that parents look more at raw scores than

CVA (Hansen and Machin, 2010)

• Our first objective is to try to find a simple measure which is easy to understand for parents

Motivation

• Also assumes an average CVA score of a school is meaningful as a summary statistic of the performance of a school

• Yet the literature has shown schools to be differentially effective

– Jesson and Gray, 1991; Teddlie and Reynolds, 2000;

Thomas et al. 1997; Wilson and Piebalga 2009

• Our second objective is to try to provide a simple measure which allows for differential effectiveness

Research aims

• If schools are differentially effective then parents need to know the value added by a school for children with similar prior attainment to their own child

• We propose a measure that would do this

• Abstracts from issues of sorting into schools and mobility

Key research questions

• To what extent do summary measures of school performance, such as CVA, hide differential performance of schools for different types of children?

• Are simple descriptive measures of the differential effectiveness of a school good enough approximations?

Literature

• We contribute to the following literatures:

– technical limitations of published school performance measures (Goldstein and Spiegelhalter 1996, Leckie and Goldstein 2009)

– measurement of differentially effective schools

(Jesson and Gray, 1991; Teddlie and Reynolds, 2000;

Thomas et al. 1997; Wilson and Piebalga 2009).

– incentives for schools when using performance measures to improve school accountability (Ladd and

Walsh, 2000)

Methodology

• Divide pupils into prior attainment groups on the basis of KS2 scores (parents are only given group information)

• Calculate various measures of individual performance at GCSE for pupils in each of the prior attainment groups at KS2

• For each school we average across the values for its pupils in each prior attainment group - 8 summary statistics of pupil performance.

• If these group averages vary significantly - school is differentially effective.

Data

• Integrated National Pupil Database

(NPD)/Pupil Level School Census (PLASC)

• Two cohorts of pupils in year 11 (age 16) in

2006/7 and 2007/8.

• State school pupils for whom we have KS2 test scores

Prior attainment groups

• Key Stage 2’ (KS2) English and mathematics attainment (age of 10/11 year 6)

• Expected level of achievement is 4

• 5 x 5 combinations of mathematics and

English into 8 groups

• Eight groups are below level 3; level 3-3; level

4-3; level 3-4; level 4-4; level 4-5; level 5-4 and level 5-5.

KS2 prior attainment groups for year 11 children in state secondary schools in 2006/7 and 2007/8

KS2 group

Below level 3

33

34

43

44

45

54

55

Frequency

73,922

102,591

73,063

96,762

339,519

119,474

113,325

198,326

6.5

8.7

30.4

%

6.6

9.1

10.7

10.2

17.8

(cumul.)

6.6

15.7

22.2

30.9

61.3

72.0

82.2

100.0

Total 1,116,982 100.0

Outcomes

• Capped GCSE scores

• Based on pupil’s 8 best GCSE scores

• Points achieved in English and mathematics GCSE added to capped score

• Ensures that essential academic skills in mathematics and English are included

– If already present in the capped score, this implies that maths and English enter our measure twice

• This augmented capped score has recently been adopted in official CVA model

Adjusted raw score measure

• Individual’s KS4 score minus the mean of other individuals in the

KS2 prior attainment group

• Similar to the value-added (VA) measure used by DCSF 2002-5

– We use the mean group score rather than the median

– We use prior attainment groups rather than a univariate score

– We do not include science

– Our KS4 measure is the capped 8 score augmented by English and maths rather than the straight capped 8 score.

• DCSF summarised school performance by taking the average of these individual-level differences across all pupils in the school.

• We calculate 8 separate averages for each school, one for each prior attainment group.

VA and Adjusted VA measures

• VA measure then allows fully for prior attainment by estimating the following equation by group to predict expected KS4

• KS4 ig

= a g

+ b g

.KS2

ig

+ u ig g = 1..8 groups

• CVA measure then allows for contextual factors by adding controls

– gender, month of birth, IDACI, FSM, EAL, SEN, ethnicity

Group adjusted raw score

(crudely allows for prior attainment group)

Absolute

1.

diff

KS4

= KS4-KS4 mean metric: KS4 points

Relative

2.

Z

KS4

= [KS4-KS4 mean

]/KS4

SD metric: group KS4 SDs

VA (value added controlling for prior KS2 score)

3.

residual of regression of KS4 on KS2 metric: KS4 points

4.

residual of regression of Z

KS4 on Z

KS2

, where latter defined analogously [equivalent to measure 3 divided by KS4

SD

] metric: group KS4 SDs

Adjusted VA (value added with covariates)

5.

as for measure 3 but with controls in regression metric: KS4 points

6.

as for measure 4 but with controls in regression metric: group KS4 SDs

Groups

Whole School - All

Group 22

Group 33

Group 34

Group 43

Group 44

Group 45

Group 54

Group 55

Group adjusted raw score

15.8

34.0

[12.270]

19.5

[12.695]

28.4

[14.580]

33.0

[12.970]

21.4

[ 4.621]

14.0

[ 9.064]

15.4

[ 7.055]

-11.6

[ 5.783]

P-value (Groups same) 0.005

[11.968]

21.1

[ 4.409]

14.2

[ 8.544]

13.9

[ 6.491]

-7.8

[ 5.231]

VA

15.8

28.5

[12.352]

Covariate

Adjusted

VA

13.2

35.2

[11.981]

20.1

[12.603]

14.8

[11.899]

27.6

20.9

[14.360] [13.312]

31.4

25.3

[11.721]

17.5

[ 4.267]

10.6

[ 8.238]

12.2

[ 6.778]

-5.8

[ 4.857]

No.

Obs.

% total

666

46 6.9

62

34

48

225

75

78

98

9.3

5.1

7.2

33.8

11.3

11.7

14.7

0.02

0.039

Group adjusted

Groups raw score VA

Whole School - All 5.287

4.285

Group 22

Group 33

Group 34

Group 43

Group 44

Group 45

Group 54

Group 55

Groups av

P value

Covariate adjusted

VA

0.163

[ 4.478]

45.606

[ 3.517] [ 3.363]

42.111

33.712

[22.419] [22.059] [19.209]

2.72

4.538

-1.695

[15.359] [14.887] [13.988]

-24.973

-25.092

-25.65

[19.419] [20.035] [19.396]

23.643

[ 8.902]

15.457

[ 5.652]

3.189

27.316

24.274

[ 8.012] [ 8.131]

15.315

9.323

[ 5.256] [ 4.933]

5.476

3.945

[ 9.956]

-11.597

[ 9.043] [ 8.297]

-12.362

-15.72

[11.686] [11.157] [11.487]

-16.698

-11.367

-13.98

[ 6.374]

4.808

[ 3.612]

0.001

[ 6.241] [ 6.066]

6.058

1.924

[ 3.452] [ 3.307]

0.001

0.002

No.

Obs

540

23

46

36

60

174

62

66

73

540

How common is differential effectiveness?

This slide shows the % of schools that are differentially effective, as measured by a significant difference (at the 5% level) in the means of the measures across the prior attainment groups.

Dependent Variable Absolute Relative

Raw score

VA

40.0%

37.9%

35.2%

31.7%

CVA

Number schools

31.7% 25.0%

3096

Differential effectiveness and selective schools

This slide shows the % of schools that are differentially effective including and excluding selective schools.

Incl selective Excl selective

Dependent Variable Absolute Absolute

Raw score 40.0% 37.0%

VA

CVA

Number schools

37.9%

31.7%

3096

35.2%

29.8%

2932

Robustness Test

This slide shows the % of schools that are differentially effective as measured by a significant difference at both the 5% level and the 1% level in the means of the measures across the prior attainment groups.

5% significance 1% significance

Dependent Variable

Raw score

Absolute

37.0%

Absolute

23.4%

VA

CVA

Number schools

35.2%

29.8%

2932

21.6%

17.0%

Rank correlations within group

Group 22

Group 33

Group 34

Group 43

Group 44

Group 45

Group 54

Group 55

Raw/VA Raw/CVA VA/CVA

0.99

0.92

0.93

0.99

0.99

0.91

0.88

0.92

0.89

0.99

0.99

0.91

0.92

0.87

0.89

0.98

0.98

0.97

0.86

0.89

0.89

0.91

0.86

0.9

Value Added rank correlations excluding selective schools

Group 22

Group 33

Group 22 Group 33 Group 44 Group 55

1.00

0.68

1.00

Group 44

Group 55

Average

0.58

0.37

0.70

0.71

0.49

0.82

1.00

0.71

0.94

1.00

0.74

Robustness checks

• Sample size issues so re-estimated results where n>10 in each prior attainment group in each school

• Robustness to missing data problems – using teacher predictions

Things to do....

• Multiple comparisons with the best/ comparison statistics

• Noise in rank correlations

Conclusions

• Schools are differentially effective but estimates are sensitive to how this is measured

– 30-40% of schools are differentially effective at 5% level of significance

– 20% of schools are differentially effective at 1% level of significance

– estimates vary somewhat across measures (raw scores, VA, adjusted VA) though there is high correlation between measures 0.86-0.99

• Even the most conservative estimate suggests one in six schools are differentially effective

Conclusions

• For school league tables (and hence parents) this differential effectiveness would seem to matter

– the rank of schools varies substantially for different prior attainment groups (correlation across groups

0.3-0.7)

– this of course abstracts from the statistical significance of the differences

• But the results suggest that for a non trivial proportion of schools parents need information on value added by school for a particular prior attainment group

Implications

• Simple measures also suggest significant amounts of differential effectiveness but as estimates do vary by measure we need to specify preferred measure

• Results indicate different rankings of schools for different ability groups but further work needed on multiple comparisons and identifying significant differences in rank correlations

Implications for policy: a sizeable minority of schools add different value for pupils with different prior attainment and there are simple measures that can communicate this to parents.

References

• Goldstein, H. and Spiegelhalter, D. J. (1996) League tables and their limitations: statistical issues in comparisons of institutional performance. Journal of the Royal Statistical Society: Series A, 159, 385-443.

• Goldstein H, Rasbash J, Yang M, Woodhouse, G, Pan H, Nuttall, D, and Thomas, S (1993) ‘A multilevel analysis of school examination results’ Oxford Review of Education, 19: 425-33.

• Gorard, S. (2010) All evidence is equal: the flaw in statistical reasoning, Oxford Review of Education, (forthcoming).

• Jesson, D and Gray J (1991). Slants on Slopes: Using Multi-level Models to Investigate Differential School

Effectiveness and its Impact on Pupils’ Examination Results. School Effectiveness and School Improvement: An

International Journal of Research, Policy and Practice. 2(3):230-247.

• Ladd and Walsh (2000) ‘Implementing value-added measures of school effectiveness: getting the incentives right’,

Economics of Education Review, vol. 2 part 1 pp. 1–17.

• Leckie, G. and Goldstein, H. (2009) The limitations of using school league tables to inform school choice. Journal of

the Royal Statistical Society: Series A. vol. 127 part 4, pp835-52.

• Ray, A. (2006) School Value Added Measures in England. Paper for the OECD Project on the Development of Value-

Added Models in Education Systems. London, Department for Education and Skills http://www.dcsf.gov.uk/research/data/uploadfiles/RW85.pdf

.

• Teddlie, C. and Reynolds, D. (2000) The International Handbook of School Effectiveness Research, Reynolds, Falmer

Press, London and New York.

• Thomas, S, Sammons, P, Mortimore, P and Smees, R, (1997) ‘Differential secondary school effectiveness : examining the size, extent and consistency of school and departmental effects on GCSE outcomes for different groups of students over three years’, British Educational Research Journal, no. 23, part 4, p.451-469.

• Wilson D and Piebalga A (2008) ‘Performance measures, ranking and parental choice: an analysis of the English school league tables’ International Public Management Journal, 11: 233-66

Download