USING CEM SECONDARY DATA PRACTICAL APPLICATIONS IN SCHOOL APRIL 2011 Bristol Conference Geoff Davies CEM Secondary data includes: • • • • • • Baseline test data (developed ability) ‘Predictive’ data including chances graphs Value-added data Attitudinal data Curriculum assessments (Insight/Sosca) PARIS software programmes The use of this data needs to allow us to do our best to help every pupil to at least achieve if not exceed their potential. It may challenge • • • • The culture of ‘my’ school Accountability policy Expectations Staff training in use of data and ability to cope with data (data overload) • Integrating the data into school procedures, storage, retrieval, distribution and access • Roles and Responsibilities Carol Fitz-Gibbon 2001 British Psychological Society It gradually dawned on me that providing the data to schools was the most important outcome of the effort, far more important than writing research papers….. The provision of data to practitioners meant that they participated in the research. Indeed they were the only ones who knew the surrounding circumstances for their classrooms, their department, each pupil, each family, etc. They were the major players: the ones who could interpret and learn from the detailed data. …..there is a need for teacher researcher posts on the Senior Management team with a brief to develop research that is useful. Given time in the timetable thousands of teachers could become active researchers…. Educational research should be a practical reality contacting scientific enlightenment not a mathematical weight lifting exercise The sense and usefulness of what we are doing induces creative thoughtfulness. QUESTION IS THIS STIFLED IN THE PRESENT EDUCATIONAL CLIMATE? creative thoughtfulness PRIORITISING Can school leaders find the time? Important but not urgent Is it more like? 65%-80% or 15% Urgent but not important Is it more like? I keep six honest serving men Who taught me all I know Their names are What and Why and When And How and Where and Who Rudyard Kipling The Elephant’s Child 15% or 60% QUESTIONS INSPIRED BY CEM CENTRE DATA • • • • • • • • • • • • • • • • • • • • How do we improve our professional judgement? Do MIDYIS scores tell us more than we think? Which ability range of pupils gets the best deal from our school? Which ability range of pupils is served best by a particular subject department? Can we do action research in our school using historical data? Can standardised residuals tell us more about a department than just the + or -? Does learning support work? What do attitudinal surveys over time give us? Can we compare standardised scores with comments made by teachers on reports? Drilling down from SPC charts. Which pupils are we succeeding with? Which pupils are we not succeeding with? What can be done about it? Is there a pattern? What are the gender issues? What are the prior learning issues? What can we learn from plotting standardised scores over time? Can SOSCA help with boy/girl issues? Is key stage 3 teacher assessment sufficiently rigorous? How can ALIS be used to inform teaching and learning? WHAT information do I want? WHY do I want it? WHEN do I need it? HOW do I collect it? WHERE can I find it? From WHO do I get it? The CEM centre ticks many of these boxes for schools A small selection of questions we will look at 1. Do Midyis scores tell us more than we think? 2. Which ability of students do most or least well in our school (using YELLIS data)? 3. Can we review post 16 pedagogy (using ALIS data)? 4. What did SOSCA teach us about Key Stage 3 assessment? Do Midyis scores tell us more than we think? Using Midyis baseline data Using MidYIS IPRs Booklet.pdf What do the sections of the test measure? Vocabulary Score The Vocabulary component of the test is generally an important element for most subjects. For English, History and some Foreign Languages it is the best. However the Vocabulary score is perhaps the most culturally linked of all the scores. Those who have not been exposed to vocabulary-rich talk or a wide variety of reading material or whose first language is not English are unlikely to have developed as high a vocabulary score as they would have developed in a different environment. Maths Score The Maths score is well correlated with most subjects but is particularly important when predicting Maths, Statistics, ICT, Design Technology and Economics. The Maths section has been designed with the emphasis on speed and fluency, rather than knowledge of Maths. Like the Vocabulary score, the Maths score is a good predictor of later academic performance. Non-Verbal Score The Non-Verbal score is composed of the three sub tests: CrossSections, Block Counting and Pictures. The Non-verbal score is important when predicting Maths, Science, Design Technology Geography, Art and Drama. It provides a measure of the pupil’s ability in 3-D visualisation, spatial aptitude, pattern recognition and logical thinking. It can give an insight in to the developed ability for pupils for whom English is a second language Skills Score In the Proof Reading section pupils are asked to spot mistakes in the spelling, punctuation and grammar of a passage of text. eg mis-spelling of words like ‘there’ and ‘their’. The PSA (Perceptual speed and accuracy) section asks pupils to look for matches between a sequence of symbols on the left and a number of possible choices on the right. Given enough time most pupils would probably get the answers correct but we are measuring how quickly pupils can find a correct match. The PSA section allows speed to be demonstrated free from the demands of memory. The Proof Reading and PSA tests are tests for the modern world, and are designed to measure fluency and speed. They rely on a pupil’s scanning and skimming skills, skills that are desirable in examination situations. Some pupils will display an IPR pattern with significant differences between one or two components of the MidYIS Test. These can be the most interesting and possibly the most challenging pupils for mainstream classroom teachers. Scenarios and anecdotal findings It is when the IPR is placed in the hands of a teacher who knows that pupil that it becomes a powerful tool. Confidence Limits The pupil scored 114 on the Vocabulary section. The error bars range from about 105 to 123, about 9 points either side of the pupil’s score. If this pupil was to take this test afresh 100 times, we would expect that 95 of those times the pupil’s score would fall within the range denoted by the error bars Comparing Maths and Vocabulary Scores Are the scores significant? Relative to the Average Performance? The error bars for Vocabulary and Maths do not cross the line at 100 ( av. performance). Performance in Vocabulary is significantly better than average performance and Maths performance is significantly below average. The error bars for the Non-verbal, Skills and Overall MidYIS scores do cross the line at 100 and hence the pupil cannot be considered to have performed significantly different to the average pupil overall. A SELECTION OF MIDYIS SCORES FOR ‘WATERLOO ROAD’ !! Why would this be a very challenging class to teach? Vocabulary Surname Maths Non Verbal Skills MidYIS Score St. Score Band St. Score Band St. Score Band St. Score Band St. Score Band Sex A F 81 D 110 B 108 B 112 A 94 C B F 128 A 107 B 105 B 94 C 120 A C M 106 B 121 A 103 B 90 D 114 A D F 107 B 84 D 96 C 107 B 96 C E M 96 C 90 D 130 A 91 C 92 C F F 86 D 86 D 120 A 74 D 84 D G F 100 B 115 A 80 D 103 B 108 B H F 121 A 96 C 114 A 86 D 111 A I M 92 C 100 C 96 C 123 A 95 C J M 100 C 105 B 100 C 99 C 102 B K M 128 A 132 A 114 A 131 A 133 A L M 76 D 70 D 74 D 73 D 71 D What do I need to know/do to teach this (difficult) class of twelve pupils These are real anonymous scores from a number of schools around the UK Vocabulary scores significantly lower than other component scores Second language? Deprived areas? Difficulty accessing curriculum.? Targeted help does work. Seen in nearly all schools. Worth further diagnosis Vocabulary scores significantly higher than other component scores Good communicators. Get on. Put Maths problems in words? Mathematics significantly higher than other scores From Far East? Done entrance tests? Primary experience? Mathematics significantly lower than other scores Primary experience. Use words and diagrams? Sometimes difficult to change attitude.. Low Mathematics scores with High Non-verbal Scores Use diagrams. Confidence building often needed Pupils with non-verbal scores different from others – High Non-verbal Scores Frustration? Behaviour problems? Don’t do as well as good communicators or numerate pupils? Pupils with non verbal scores different from others – Low Non-verbal Scores Peak at GCSE? A level ? Pupils with low Skills scores Exams a difficulty after good coursework? High Skills Scores Do well in exams compared with classwork? The Average Pupil They do exist! High scores throughout Above a score of 130 puts the pupil in the top 2% nationally Low scores throughout Below a score of 70 puts the puil in the bottom 2% nationally Sharing the MidYIS Information within School Once you have received your MidYIS feedback you need to decide who will be privy to which information. Some schools decide to keep the data within the senior management team, others with Heads of Department and/or Heads of Year, some share with all staff and what about pupils and their parents? Use you MIS systems to put the data where it matters MidYIS data can be useful: •to indicate reasons for student learning difficulties and may go some way to explain lack of progress, flag up causes for underachievement and even behaviour problems. •for all teachers and support staff. It can help to support professional judgement and give a better understanding of the progress students make at school and their potential later performance. •to refer to for pupil reviews, writing reports, meeting parents, monitoring progress and interim assessments. 2. Which ability of students do most or least well in our school (using YELLIS data) Which ability range of pupils gets the best deal from our school? Which ability range of pupils is served best by a particular subject department? Can standardised residuals tell us more about a department than just the + or -? Using standardised residuals in a different way 0.1 0.4 Voc Health & Social Care -0.5 Voc Applied Science -1.1 SC Religious Studies -0.5 SC History -0.3 SC Geography 1.0 SC Art & Design -0.1 Science: GCSE Additional -0.1 Science: GCSE 0.1 Religious Studies -1.3 Music 0.1 Media Studies -0.3 Mathematics 0.2 History -0.2 Geography -0.5 French English Literature -0.6 English -1.0 Design & Technology Art & Design Average Standardised Residual CONTRAST THIS 4.0 3.0 2.0 0.3 0.0 -0.9 -0.5 -2.0 -3.0 -4.0 -1.0 -2.0 -3.0 -4.0 0.0 -0.2 0.8 0.4 SC ICT 0.3 Welsh 0.2 Double Science -0.1 Religious Studies -0.1 Physical Education 0.6 Music 0.5 Maths 0.9 ICT 0.3 Home Economics History 0.4 German 0.2 Geography 0.1 French 0.4 English Literature 0.0 English 1.0 Drama Design & Technology 0.0 Business Studies Art & Design Average Standardised Residual WITH THIS 4.0 3.0 2.0 1.4 0.9 As on the last two slides CEM provides value added charts using average standardised residuals for departments We are going to show how standardised residuals can be used in a different way in your school yellis exercise.doc 1. Review of post 16 pedagogy (using ALIS data) SUBJECT A Why the improvement? 2008 PEDAGOGY….ALIS surveys SUBJECT A 2005 SUBJECT A 2007 SUBJECT A 2008 DO NOT GET TOO EXCITED! We have made a comparison of perceived teaching methods as analysed by ALIS in 2004-5 with those in 2007-8. Some subject areas have appeared to change their methods radically. Others have not. Though the samples are small it is an interesting exercise to try to correlate it with the departments statistical process charts over that period. One would like to say that changes in the variety of teaching methods result in improvement but the evidence is a little tenuous so far. e E x x a e m r p c a i p s e e r s s SUBJECT e s s a y s R e a d i n g d i s c u s s i o n s n o t e s d i c t a t e d o w n n o t e s h a n d o u t s P r a c t i c a l w o r k U s i n g A V I T R e s e a r c h i n g P r e s e n t i n g h e l p i n g s t u d e n t s o r i g i n a l w o r k 2005 to 2008 ART BIOLOGY BUSINESS CHEMISTRY COMPUTING DRAMA FOOD PRODUCT DESIGN ENGLISH FRENCH GEOGRAPHY HISTORY ICT MATHS MUSIC PE PHYSICS RE WELSH YES NO MORE (M) LESS (L) SIGNIFICANT CHANGE YES NO YES NO YES NO YES YES NO YES YES NO YES NO YES YES YES NO YES 11 7 54 29 L L M L L L M L M L L L L L M M M L L L M L L M M L M M M L M L L M M M M L M M M M M M M M M M M M M L M M M M M M M M M 4 3 M M M M L L L 5 3 L L 1 2 L 3 6 4 3 5 1 6 2 5 1 M M M M M M M M M L M 2 2 3 1 2 0 3 0 2 2 6 3 3 0 4. What have we learned from SOSCA? GRAPH 1 AVERAGE STANDARDISED RESIDUALS BOYS/GIRLS MIDYIS TO KEY STAGE THREE TEACHER ASSESSMENT GENDER ISSUES DYFFRYN TAF KEY STAGE 3 AVERAGE STANDARDISED RESIDUALS MIDYIS TO KS3 0.8 0.6 GIRLS GIRLS 0.4 OVERALL BOYS 0.2 OVERALL GIRLS 0 BOYS English Maths -0.2 BOYS -0.4 SUBJECT Science BOYS 95%CONFIDENCE 95% CONFIDENCE GIRLS 95% CONFIDENCE 95% CONFIDENCE OVERALL 95% CONFIDENCE 95% CONFIDENCE GRAPH 2 AVERAGE STANDARDISED RESIDUALS BOYS/GIRLS MIDYIS TO SOSCA GENDER ISSUES DYFFRYN TAF SOSCA 0.4 AVERAGE STANDARDISED RESIDUALS MIDYIS TO SOSCA BOYS 0.3 GIRLS OVERALL OVERALL BOYS 0.2 GIRLS 0.1 BOYS 0 Reading Maths OVERALL Science -0.1 GIRLS -0.2 95% CONFIDENCE -0.3 SUBJECT BOYS 95%CONFIDENCE 95% CONFIDENCE GIRLS 95% CONFIDENCE 95% CONFIDENCE OVERALL 95% CONFIDENCE 95% CONFIDENCE GRAPH 1 AVERAGE STANDARDISED RESIDUALS BOYS/GIRLS MIDYIS TO KEY STAGE THREE TEACHER ASSESSMENT GENDER ISSUES DYFFRYN TAF KEY STAGE 3 GIRLS 0.6 GIRLS 0.4 OVERALL BOYS 0.2 OVERALL GIRLS 0 BOYS English Maths BOYS 95%CONFIDENCE 95% CONFIDENCE GIRLS 95% CONFIDENCE 95% CONFIDENCE OVERALL 95% CONFIDENCE 95% CONFIDENCE Science -0.2 BOYS -0.4 SUBJECT GRAPH 2 AVERAGE STANDARDISED RESIDUALS BOYS/GIRLS MIDYIS TO SOSCA GENDER ISSUES DYFFRYN TAF SOSCA 0.4 BOYS AVERAGE STANDARDISED RESIDUALS MIDYIS TO SOSCA AVERAGE STANDARDISED RESIDUALS MIDYIS TO KS3 0.8 0.3 GIRLS OVERALL OVERALL BOYS 0.2 GIRLS 0.1 BOYS 0 Reading Maths OVERALL Science -0.1 GIRLS -0.2 95% CONFIDENCE -0.3 SUBJECT BOYS 95%CONFIDENCE 95% CONFIDENCE GIRLS 95% CONFIDENCE 95% CONFIDENCE OVERALL 95% CONFIDENCE 95% CONFIDENCE Using MIDYIS and SOSCA puts the school in a strong position to improve its professional judgment of teacher assessments at Key stage 3. Statutory testing disappeared in Wales some five years ago. Comparing the value added for MIDYIS to SOSCA and MIDYIS to KS 3 teacher assessment shows up some interesting data Schools who depend on teacher assessment data only to measure value added from Key stage 3 to Key stage 4 need to be aware of the pitfalls. The use of SOSCA data in this exercise highlights that see http://www.tes.co.uk/article.aspx?storycode=6062337 Subjective views of pupils - as well as pressure from parents - make model unreliable, warns Professor Tymms. The differences appear to relate to the types of assessment used in the various subject areas. English and Welsh use extended writing for teacher assessment which is more likely to have subjective judgments. What we have learnt from this is that despite a moderation process built on portfolios of work for teacher assessment, it is not sufficient in isolation. Computer adaptive tests such as SOSCA and the resulting value added information from MIDYIS are more informative in a diagnostic sense than levels produced by teachers for statutory assessment. SOSCA has also been used to investigate any changes in reading from baseline testing. A high correlation was found between the London Reading score given to pupils on entry the MIDYIS score and the SOSCA reading test. PITFALLS 1. Tracking developed ability measures over time. 2. Looking at average standardised residuals for teaching sets. 3. Effect of one result in a small group of students REGRESSION TOWARDS THE MEAN DIFFERENCE SOSCA-MIDYIS MATHS 20 10 0 SOSCA-MIDYIS DIFFERENCE 50 70 90 110 130 150 170 -10 DIFFERENCE -20 -30 -40 -50 MIDYIS MATHS STANDARDISED Pupils with high MidYIS scores tend to have high SOSCA scores but not quite as high. Similarly pupils with low MidYIS scores tend to have low SOSCA scores, but not quite as low. It is a phenomenon seen in any matched dataset of correlated and normally-distributed scores, the classic example is a comparison of fathers' and sons' heights. Regression lines reflect this phenomenon - if you look at the predictions used in the SOSCA value-added you can see that for pupils with high MidYIS scores their predicted SOSCA scores are lower than their MidYIS scores, whereas for pupils with low MidYIS scores their predicted SOSCA scores are higher than their MidYIS scores. Marksheet Name : SUBJECT REVIEW Marksheet Group : 11S1 A B GCSE Standard Residual Ma WJEC/GCSE 018403 ResGF WJEC/GCSE 018402 ResGF WJEC/GCSE 018401 ResGF YELLIS GCSE PREDICTION MA a 90019 6 b 90090 7 c 90045 6 63 A B B B B 0.10 d 90063 7 64 A A B B B 0.10 e 90166 6 48 B B B C C 0.40 f 90123 7 70 A A A A B -0.40 g 90129 6 47 C C B C C 0.50 h 90146 6 59 B B A B A 1.40 I 90047 7 62 A A B B B 0.20 j 90115 7 67 A A * A A* 1.70 k 90004 6 46 C B B C B 1.50 l 90164 7 65 A A A B A* 1.90 m 90099 7 70 A A A A A* 1.50 n 90011 7 61 A A A B A 1.30 o 90112 7 66 A A B A A 0.80 p 90058 6 70 A A B A A 0.50 q 90150 7 72 A A A A A 0.40 r 90127 6 52 B B B C B 1.00 s 90030 6 58 B B B B B 0.50 t 90050 7 71 A A A A A 0.40 u 90016 6 69 A A B A B -0.40 v 90174 7 74 A A A A A 0.20 w 91165 6 62 A B B B x 90109 7 63 A B B B B 0.10 y 90138 7 47 C B B C B 1.40 z 90122 7 60 A A * B A 1.30 ab 90009 7 60 A A A B A 1.30 ac 90169 7 79 A A * A* A -0.20 ad 90153 6 56 B B B B ae 90010 7 64 A B B B A 1.00 af 90154 7 61 A C B B B 0.30 Total 62 A PREDICTION MATHS YELLIS MATHS BAND Yellis Band Yellis Score Admission No. Students CLASS REVIEW : 04/10/2005 Maths Test K3 Wa Export Date B A B 1.20 B B 0.20 B 0.70 1323 1868 109 105 201 156 12 190 31 30 30 30 31 30 2 29 30 Mean 42.68 62.27 3.63 3.5 6.48 5.2 6 6.55 0.70 Mean Grade 6.00 Num ber of Results B B B B B B 20.90 BEWARE PITFALLS INTERPRETATION SUBJECT M Sex M M F M F F M M F F F M Score (Band) 53 (B) 38 (C) 36 (D) 48 (C) 52 (B) 65 (A) 70 (A) 38 (C) 40 (C) 70 (A) 44 (C) 56 (B) Predicted Grade 5.4 (B/C) 4.5 (C/D) 4.4 (C/D) 5.1 (C) 5.3 (B/C) 6.1 (B) 6.4 (A/B) 4.5 (C/D) 4.6 (C/D) 6.4 (A/B) 4.8 (C) 5.6 (B/C) 5.3 (B/C) Achieved Grade 6 (B) 3 (E) 3 (E) 5 (C) 6 (B) 7 (A) 3 (E) 4 (D) 5 (C) 7 (A) 6 (B) 5 (C) 5.0 (C) Raw Residual 0.6 -1.5 -1.4 -0.1 0.7 0.9 -3.4 -0.5 0.4 0.6 1.2 -0.6 -0.3 Standard ised Residual 0.5 -1.1 -1.0 -0.1 0.5 0.7 -2.5 -0.4 0.3 0.4 0.9 -0.4 -0.2 REVISED 0.5 -1.1 -1.0 -0.1 0.5 0.7 -0.4 0.3 0.4 0.9 -0.4 0.0 PLEA DON’T LET THE SYSTEM DESTROY THIS creative thoughtfulness USING CEM SECONDARY DATA PRACTICAL APPLICATIONS IN SCHOOL APRIL 2011 Bristol Conference Geoff Davies