Using Student Assessment Results to Improve Teacher

advertisement
Using Student Assessment Results to
Improve Teacher Knowledge and
Practice
NESA FLC
24- 27 October2013
MARTINA BOVELL, B.A. Dip. Ed. Grad dip. Arts
(UWA)
Senior Research Fellow
www.acer.edu.au
80 years experience
Independent
Not for profit
Over 300 staff in seven offices
Melbourne, Sydney, Brisbane, Adelaide, Perth, India (New Delhi), UAE (Dubai)
Four goals:
Learners and their needs
every learner engaged in challenging learning opportunities appropriate to their readiness and
needs
The Learning Profession
every learning professional highly skilled, knowledgeable and engaged in excellent practice
Places of learning
every learning community well resourced and passionately committed to improving outcomes for all
learners
A Learning Society
a society in which every learner experiences success and has an opportunity to achieve their
potential
ACER and international assessment
Member of the International Association for
the Evaluation of Educational Achievement
(IEA).
Consortium leader for the OECD’s
Programme for International Student
Assessment (PISA) 1998 – 2012.
Purposes of assessment
• clarify educational standards
• monitor trends over time
• evaluate the effectiveness of
educational initiatives and programs
• ensure that all students achieve
essential skills and knowledge
International Schools’ Assesssment
http://www.acer.edu.au/tests/isa
Over 64 000 students from 312 schools participated in the ISA in
October 2012 and Feb 2013.
The ISA
Grades 3 – 10
•
•
•
•
Mathematics
Reading
Expository writing
Narrative writing
Grades 8 – 10
• Science online
• New in 2013-2014 testing cycle
How ISA results are reported
• All students who participate in ISA tests have their performance
measured against a single scale.
• There is a separate scale for each of the five domains being
assessed (Reading, Mathematics, Narrative Writing, Expository
Writing, and Science).
• The ISA scale score for a student is different from the “raw”
score that the student would get by adding up the number of
correctly answered questions on a test.
• The ISA scale allows meaningful comparisons of results between
different grade levels and between different calendar years even
though the tests administered are not the same.
• The ISA scales are based on those developed for the
Organisation for Economic Co-operation and Development’s
(OECD’s) Program for International Student Assessment (PISA).
Types of information
Whole grade mean and distribution
– all students, language background, gender
– this year and previous years (longitudinal comparisons)
• compared to all other ISA schools
• compared to all other Like schools (based on %s of students with ESB)
• For years 8 – 10, comparison with PISA results (Maths, Reading,
Science)
Classes within whole grades
– class mean and individual student scale scores
– performance by item classification within a domain
• compared to all other ISA schools
• compared to all other Like schools
• compared to other classes in your school
Item by item results for each student
– By item classification
– compared to other individuals
– compared to all other ISA schools
ISA Reports
• Paper-based
• Electronic
• tracking
• interactive
Science: online test delivery means
faster turn-around of results
“Improve teacher knowledge and
practice”
A growth model:
– acknowledges that each student is at some
point in their learning
– expects every student to make excellent
learning progress regardless of their
starting point
– assesses growth over time
ACER National School Improvement tool
(NSIT)
• Means of ACER’s mission of improving
learning
• Endorsed by Australian federal and state
governments
• Research-based
• Informs and assists schools’ improvement
agendas
• Available to all schools
http://www.acerinstitute.edu.au/home
Part of ACER’s Institute of Learning larger
school improvement project.
• The Institute’s services include
– Professional learning
– School review services
– Capacity building for school improvement
www.acer.edu.au/documents/NSIT.pdf
9. Schoolcommunity
partnerships
8. Effective
pedagogical
practices
7.
Differentiated
teaching and
learning
1. An explicit
improvement
agenda
2. Analysis
and
discussion of
data
National school
improvement
tool
6.
Systematic
curriculum
delivery
5. An expert
teaching
team
3. A culture
that
promotes
learning
4. Targeted
use of
school
resources
Review of a domain
Domain performance levels
Domain performance level: LOW
• Teachers do not systematically analyse
test and other data for their classes and
teachers make little use of data to
reflect on their teaching.
Domain performance level: MEDIUM
• An ad hoc approach exists to building
staff skills in the analysis, interpretation
and use of classroom data
Domain performance level: HIGH
• One or more members of staff have been assigned
responsibility for implementing the annual plan,
analysing the full range of school data, and
summarising, displaying and communicating student
outcome data for the school. The school has ensured
that appropriate software is available and that at
least these assigned staff have been trained to
undertake data analyses.
• Time is set aside (e.g. on pupil free days and in staff
meetings) for the discussion of data and the
implications of data for school policies and classroom
practices. These discussions occur at whole-school and
team levels.
Domain performance level: OUTSTANDING
• Data are used throughout the school to identify gaps in
student learning, to monitor improvement over time and to
monitor growth across the years of school.
• A high priority has been given to professional development
aimed at building teachers’ and leaders’ data literacy skills.
• Staff conversations and language reflect a sophisticated
understanding of student assessment and data concepts
(e.g. value-added; growth; improvement; statistical
significance).
• Teachers are given class test data electronically and are
provided with, and use, software to analyse, display and
communicate data on individual and class performances and
progress, including pre- and post-test comparisons.
• Teachers routinely use objective data on student
achievement as evidence of successful teaching.
Data literacy framework
US Dept of Education office of planning, evaluation and
Policy development (2011). Teachers’ ability to use data to
inform instruction: challenges and supports.
• STUDY:
• investigation of teachers’ thinking about student data by administering
hypothetical education scenarios accompanied by data displays and
questions to individual and small group interviews to teachers and
schools identified as exemplary in active data use (50 teachers, 72 small
groups)
• PURPOSES:
• 1. to investigate teachers thinking and reasoning independently about
data and how they build on each other's understandings when working
with data in small groups.
• 2. to isolate the difficulties, misconceptions and support needed.
What they found
• Teachers’ likelihood of using data is affected by
how confident they feel about their knowledge
and skills
• Working in small groups appears to promote
teachers’ engagement with data. Compared when
working individually, teachers were
– more likely to arrive at sound data interpretations
– more likely to use a wider range of skills when making
decisions about how to use and interpret data
– able to clarify and frame problems and correct data
interpretation errors
– more likely to enjoy discussing data
Data literacy skill areas:
US study framework
• Data location
– find relevant pieces of data in the data system or display
• Data comprehension
– understand what the data is saying
• Data interpretation
– figure out what the data mean
• Instructional decision making
– select an instructional approach to address the situation identified
in the data
• Question posing
– frame instructionally relevant questions that can be addressed by
the data
Bias when making decisions using data
Representative /Availability bias
• When judging the probability of something, we use some
preconceptions based on how similar two things are or the extent to
which an event matches our previous experience.
• e.g. irrelevant personal characteristics or stereotypes
Anchoring and adjustment bias
• Make a decision based on an initial calculation without following
though on the calculations.
• We ignore data that doesn’t agree with our preliminary decisions or
biases (we agree with what we expect).
Confidence in making the decision is not always associated
with quality of decision making
Group decision making can mitigate some of these biases.
PISA findings about decision making
PISA in focus #26 http://www.oecd.org/pisa/pisainfocus/
• “Countries vary in the way they use marks, but they all tend to
reward the mastery of skills and attitudes that promote learning.”
• “Teachers tend to give girls and socio-economically advantaged
students better school marks, even if they don’t have better
performance and attitudes than boys and socio-economically
disadvantaged students.”
• “It seems that marks not only measure students’ progress in school,
they also indicate the skills, behaviours, habits and attitudes that
are valued in school.“
What biases?
Narrative Writing assessment
Narrative Task
Same tasks and marking criteria for all grades
– Content 0 - 14: quality and range of ideas, development of plot,
characters and setting, the writer’s sense of audience and purpose, the
overall shape of the writing.
– Language 0 - 14:
sentence and paragraph structure, vocabulary and
punctuation, and the writer’s voice.
– Spelling 0 - 11:
considers phonetic and visual spelling patterns and the
kind of words attempted, and correctness.
Reporting
Content: 7 (max score 14)
Language: 6 (max score 11)
Spelling: 5 (max score 11)
Individual student report comment:
Level 5: Write a story with some developed detail in content and
using a variety of sentence forms. Spell correctly many words
from a student-level vocabulary.
Reading Framework
The ISA definition of Reading literacy derives from PISA:
Understanding, using, reflecting on and engaging with written
texts, in order to achieve one’s goals, to develop one’s
knowledge and potential, and to participate in society.
Goes beyond decoding and literal comprehension and
recognises the full scope of situations in which reading plays
a role in the lives of students from grades 3 to 10.
Three parts:
ASPECT
TEXT TYPE
TEXT FORMAT
Reading Framework
Retrieving Information:
Locating , selecting, and collecting one
or more pieces of information in a text.
Interpreting texts:
Making sense of the text, constructing
meaning making connections and
drawing inferences from one or more
parts of a text, e.g. cause and effect,
compare and contrast, category and
example. Involves information that is
not stated.
Reflecting:
Drawing on knowledge, ideas and values
external to the text to evaluate a text;
relating a text to one’s experience,
knowledge and ideas.
Continuous text format
composed of sentences that are, in
turn, organised into paragraphs. These
may fit into even larger structures
such as sections, chapters and books.
Narrative pieces, exposition,
description, argument and
instructional passages
Non–continuous text format
Essentially, texts composed of one or
more lists in which information is
presented in, e.g. tables, graphs, maps
and diagrams
Feedback from EARCOS
• This workshop was exactly what was needed to guide us. It
really helped the focus and the analysis of data. I hope ISA
continue to provide this service to support schools who use this
service.
• Very useful and I am motivated to do my work better.
• I felt that the workshop has enabled me to speak with some
authority on how we can unpack the ISA results. More
importantly I have an insight into how we can use these results
to better inform further decisions, ones that are based on
student achievement.
• Very useful. More of these sessions need to be provided to
help teachers understand how data can be used to improve
learning.
• Informative and well worth the time. Hands on, interactive
learning.
ISA Interactive diagnostic report
A guide to using the interactive diagnostic
report that was demonstrated during the
the conference session is downloadable at
http://www.acer.edu.au/documents/ISA_Using_the_Interactive_Diagnosti
c_Report.pdf
Interrogating data
Focus on the student
• Did the student perform as well as expected?
• Does the performance match expectations/reflect teacher
judgement about the student?
• What does the student’s response pattern show about the
strengths of the student?
• What does the student’s response pattern show about the
areas of concern for the student?
• Are any areas of concern preventing the student from
making progress? What might account for these?
Interrogating data
Focus on the group – group scores
• How does the group achievement relate to the bands?
• How does the class distribution against the bands match
expectations about the group?
• Did the group as a whole perform as well as expected?
• Does the relative order of students match expectations
about the students?
• What students have achieved higher than expected, or
lower than expected, in relation to others?
• Are there students in the group with similar achievements?
Interrogating data
Focus on the group – group scores
• Do students with similar scale scores have similar or
different response patterns?
• What assessment criteria do students perform well on?
• What assessment criteria do students perform less well on?
• What does the group’s response pattern show about its
strengths?
• What does the group’s response pattern show about its
areas of concern?
Interrogating data
• Focus on the teaching program
• Has any teaching impacted on the group’s results?
• Are there any areas of concern preventing the
whole group from making progress?
www.acerinstitute.edu.au/conferences/eppc
Presented by practitioners, for practitioners.
ACER recognises that, every day, teachers and school leaders
are responsible for improving learning among students.
This conference provides an opportunity to report on and
celebrate the improvements you have achieved within your
classes, across the whole school or within networks of schools.
Call for papers now open
School Improvement Tool Contact
Robert Marshall
Senior Project Director
Australian Council for Educational Research
19 Prospect Hill Road, Camberwell, Victoria
Australia 3124
Robert.Marshall@acer.edu.au
• +61 3 9277 5346
• 0439 665 965
Martina Bovell
Senior Research Fellow
Australian Council for Educational Research
7/1329 Hay Street, West Perth
Western Australia 6005
martina.bovell@acer.edu.au
+61 8 9235 4821
+61 439 926 277
References
• Barber, M. and Mourshed, M. (2007). How the world’s best-performing
school systems come out on top. Mckinsey and Co.
• Dweck, C.S. (2006). Mindset: The new psychology of success. New
York: Balantine Books.
• Fullen, M., Hill, P., Crevola, C. (2006). Breakthrough. California: Corwin
Press.
• International Baccalaureate Organisation. (2010). IB learner profile
booklet. www.ibo.org
• International Baccalaureate Organisation. (2010). Programme
standards and practices. www.ibo.org
• Masters, G. (2013). Towards a grown mindset in assessment. ACER
Occasional essays – 2013. www.acer.edu.au
• Tversky, A. and Kahneman, D. (1982).Judgement under uncertainty:
Heuristics and biases. In Judgement under uncertainty: Heuristics
and biases, eds. D Kahneman, P. Slovic and A.Tversky, 3-20. NY:
Cambridge University Press – cited in US Dept of Education office of
planning, evaluation and Policy development (2011). Teachers’ ability
to use data to inform instruction: challenges and supports .
Download