Prior Value Added

advertisement
Introduction to CEM
Secondary Information Systems
Dr Robert Clark
ALIS Project Manager
The Projects
The Projects
A / AS / Btec / IB etc
Year 13
Year 12
Alis
GCSE
Computer Adaptive Baseline Test
(paper test also available)
GCSE
Year 11
Yellis
Year 10
Computer Adaptive Baseline Test
INSIGHT
Combines curriculum tests with
developed ability
(paper test also available)
Year 9
Year 8
(+ additional)
Year 7
MidYIS
Computer Adaptive Baseline Test
(paper test also available)
Typical Timeline
Typical Timeline
Early Testing &
Predictions
START:
Measure Baseline
Autumn Term
Feedback
Reports
A ugust
END:
September
Collect and Send Results
Value-Added
Feedback
Testing is available from June at end
of previous academic year
i.e. Y6 / Y9 / Y11
The baseline score is measured as
early as possible and the information
sent to the CEM Centre
The institution then receives feedback
from the CEM Centre: baseline test
results and predictions
The public examination results are
sent to the CEM Centre when received
by the institution
The value added data is available
from September via the Institution’s
website at the CEM Centre
Predictions
.
How CEM ‘Predictions’ are made…
Subject
SubjectXX
Outcome
140
A / B120
100
C 80
60
40
20
0
04
25
46
67
8
Baseline
Regression Line (…Trend Line, Line of Best Fit)
Outcome = gradient x baseline + intercept
Correlation Coefficient (~ 0.7)
Some Subjects are More Equal than Others….
140
A*
A
Grade
120
B
100
>1 grade
Photography
Sociology
English Lit
Psychology
Maths
Physics
Latin
80
C
60
D
E
40
5
C
6
B
A7
Average GCSE
A*8
Some Subjects are More Equal than Others….
GCSE (MidYIS or Yellis)
A*
A
GCSE Grades
B
1 grade
C
D
E
F
Test Score
Art & Design
Biology
Chemistry
Economics
English
French
Geography
German
History
Ict
Mathematics
Media Studies
Music
Physical Education
Physics
Religious Studies
Science (Double)
Spanish
Baseline Measurement
Problems with Key Stage Baselines
1.
Not all students have KS baselines
No KS3
Foreign Students
Vocational Students
Adult Learners
2.
KS exams do not always represent ‘Start of Course’ ability
Post-16 : Year(s) out or intermediate years
3.
Prior Value-Added
Can you add value at every Key Stage ?
Under achievement leading to under expectation
One teacher’s output = another teacher’s input
4.
Teaching to the test
Does performance represent ability, effort or exam technique ?
5.
Aptitude & fluency vs Achievement & knowledge
6.
Transfer of Data
Although Key Stage baselines can be a very
good indicator of potential attainment, by
themselves they are not sufficient.
Key Stage baselines are confounded by the
effects of prior treatment.
The Computer Adaptive Test
•
Test performed online – results automatically transmitted to CEM.
•
Minimal installation / setup required - if any.
•
Adaptive – difficulty of questions changes in relation to ability of student.
•
Efficient – no time wasted answering questions that are far too easy or
difficult.
•
Wider range of ability
•
Less stressful on students – more enjoyable experience than paper test.
•
Less demanding invigilation.
•
Cheaper !
In 2010/ 2011 over 200,000 students across yrs 7-13 sat this test
Try it yourself at www.intuproject.org/demos
Baseline Feedback
Reports, Graphs
& Predictions
IPRs
(Individual Pupil Record Sheets)
Look for sections that are
inconsistent
Also available based on MidYIS,
Alis, SOSCA & INSIGHT scores
Intake Profiles
Also available based on MidYIS, Yellis, SOSCA and INSIGHT scores
Intake Profiles (Historical)
Predictions – MidYIS example
Similar spreadsheets available from Yellis, SOSCA, INSIGHT
Predictions - Alis example
Chances Graphs
English Language - band D
English Language - band C
50
40
31
30
24
23
20
10
Percent
Percent
50
10
9
2
1
20
20
0
0
29
30
10
0
35
40
1
2
U
G
7
6
U
G
F
E
D
C
B
A
A*
F
E
Grade
0
50
Percent
21
20
7
10
0
0
5
2
0
0
U
G
F
E
D
C
Grade
B
A
A*
50
40
24
C
English Language - band A
40
30
D
Grade
English Language - band B
Percent
1
0
B
A
A*
35
40
26
30
23
20
10
8
0
0
0
1
U
G
F
E
7
0
D
C
Grade
B
A
A*
Predictions
vs
Targets
What is a ‘Prediction’ ?
 NOT a forecast of the grade the student will get
 An indication of the grade (points score) achieved on average by students
of similar ability in the previous year
Targets ?
•
Minimum Targets – Round CEM prediction down ?
•
Realistic Targets – Use nearest CEM grade ?
•
Challenging Targets – 75th percentile ?
Prior Value Added ?
Arbitrary grade fraction ?
75th Percentile Predictions
•
•
•
•
If all students attain 75th percentile predictions, School VA will be at top 25%
Provided by all Projects in addition to 50th percentile
Alis : Excel spreadsheet - ‘Predictions – Spreadsheet (75th Percentile)’
Pre-16 : Adjust button on predictions spreadsheet
Prior Value Added
•
•
•
•
•
Only where prior VA is positive ?
1 year or 3 ?
Reasonable to use raw residual figures as this is an approximate measure and
raw residuals give grade fractions
Alis : Can be calculated using PARIS software
Pre-16 : Adjust button on predictions spreadsheet
Data used to inform, not replace, professional judgement
75th Percentile
Prior Value Added
75th
Percentile
Step 1
Prior
Value-Added
Value-Added Feedback
Reports & Graphs
.
Measuring
Value-Added – Terminology
Subject X
100
Outcome
Exam grade
120
80
-ve VA
+ve VA
60
40
20
4
5
6 SCORE
BASELINE
Baseline
Trend Line/Regression Line
7
8
Raw Residual
Measuring Value-Added – An Example
National Trend
‘Average’ Student
A*
B
Result
C
D
-ve (- 2 grades)
Alf
Bob
Subject A
+ve (+ 2 grades)
E
Chris
Subject B
F
G
U
Low Ability
Average Ability
High Ability
Baseline Score
The position of the national trend line is of
critical importance
Standardisation of Residuals
•
(Raw) Residuals can be used to examine an individual’s performance
•
Standardised Residuals are used to compare performance of groups
•
Standardised Residuals are independent of year or qualification type
•
For a class, subject, department or whole institution the Average
Standardised Residual is the ‘Value-Added Score’
•
Standardised Residual = Residual / Standard Deviation (National Sample)
•
When using Standardised Residuals then for an individual subject
Standard Error

1
N
where N = number of results in the group
(for combinations of subjects consult the relevant project)
•
95% Confidence Limit = 2.0 x Standard Error
•
99% Confidence Limit = 2.6 x Standard Error
•
99.7% Confidence Limit = 3.0 x Standard Error
Burning Question :
What is my Value-Added Score ?
Better Question :
Is it Important ?
Value Added Feedback…
Statistical Process Control (SPC) Chart
2000
2001
2002
2003
2004 2005
Year
2006
2007 2008
2009
2010
-4.0
-0.7
Spanish
0.1
Religious Studies
-0.5
Physics
Music
Media Studies
-0.7
Maths
-0.4
Information Technology
-0.4
Home Economics
-0.2
History
German
-0.7
Geography
-0.6
French
-0.1
English Literature
0.1
English Language
-1.0
Drama
Design and Technology
-0.2
Chemistry
-0.2
Biology
Art
Average Standardised Residual
Subject Summary
Standardised Residual Graph
4.0
3.0
2.0
2.0
1.0
0.1
0.4
0.0
-0.2
-0.5
-2.0
-2.2
-3.0
Grade Points Equivalent
The Scatter Plot
Look for Patterns…
General Underachievement / over
achievement ?
Baseline Score
Do any groups of students stand out ?
– high ability vs low ability ?
– male vs female ?
Other things to look for…
Why did these students do so badly ?
Why did this student do so well ?
How did they do in their other subjects ?
PARIS Software
PARIS is …..
•
Software to install and use in school
•
An interactive reporting tool
•
Included free with Alis / Yellis / MidYIS
PARIS analyses …..
•
Potential Performance
•
Intermediate Performance
•
Actual Performance
PARIS provides …..
•
Student level reports
•
Subject Level reports
•
Institution Level Reports
Attitudes
There is more to school / college than exams….
•
•
•
•
•
•
•
Student attitudes
Student Welfare & Safety
Non-academic activities
Support
Social and personal development
Parental Survey
Induction Survey
Self Evaluation (Every Child Matters)
Attitudinal MidYIS
INSIGHT
Attitudinal Yellis
Try it yourself at www.intuproject.org/demos
Full ALIS
Other Issues
Points to mull over…
•
•
•
•
•
•
•
•
•
•
•
•
Independence – no agendas
Transparency of Analysis
Self Evaluation
Straightforward and standardised administration
Prompt Feedback
Full working hours phone / email support
Student focus
Replacement for KS3
Innovative online adaptive testing available – student experience
Longitudinal analysis with appropriate error backgrounds
Non-curriculum embedded baselines available
Attitudinal surveys available – Every Child Matters…
Dr Robert Clark
Alis Project Manager
robert.clark@cem.dur.ac.uk
0191 33 44 193
Download