Exceeds Standard

advertisement
Changes in Student and Teacher
Attitudes and Behaviors in an
Integrated High School Curriculum
Presented by:
Nicolle Gottfried & Catherine Saldutti
Background
Design of an integrated, standards- & inquirybased 2-year Biology and Chemistry program
Use of research-based best practices
In cooperation with: teachers, students,
Teachers College, Columbia and The
Rockefeller University
Piloted in 2 Public Schools in New York City
Teachers of varying levels of experience
Students representing diverse needs
Background, cont.
Evaluation of the pilot test included many
data sources:
Classroom Observations
Interviews of teachers, administrators and
students
Written student and teacher reflections
Review of student work and Regents test
scores
Pre- and post-participation questionnaires
including an attitudinal battery
Our Approach
Multiple sources of data designed to provide
a mosaic of understanding
To be used both formatively and summatively
All data were collected by EduChange and
Gottfried, and reported anonymously
Data Analysis
Data analyzed in aggregate, by teacher,
school and ethnicity
Qualitative data and observations reported as
notable trends, quotes and describable
behavioral changes
Quantitative data (e.g. survey responses and
reflections) have been categorized, analyzed
and reported as percentages and means
These data are compared to reasonable
matched non-participant normative data for
New York City students, when available
Key Findings
Student and teacher attitudes toward program
participation change over 2-years in a predictable
pattern
Model of Student and Teacher Attitude Change
100
80
60
Teacher
Student
40
20
0
Fall 1st
Year
Spring 1st
Year
Fall 2nd
Year
Spring
2nd Year
Key Findings, cont.
Students in the program perform on par or better than matched
counterparts on the Regents Living Environment (biology)
exams
Teachers improve relative to “best practices,” in parallel to
student improvements
Students demonstrate very sophisticated conceptual learning,
habits of mind relating to science, and how to facilitate their own
learning
“When I felt confused, I tended to just tune everything out.
When I’m confused now, I should ask questions in class.”
Additional Questions?
Contact:
Nicolle Gottfried (nmgottfried@sbcglobal.net)
Catherine Saldutti (catherine@educhange.com)
Assessing Discrete Inquiry Skills
Using a Classroom Laboratory
Rubric: Six Student Case Studies
Presented by:
Catherine Saldutti & Nicolle Gottfried
Six Case Studies
6 students who had participated in both years
of the pilot program and had most of the
requested labs available (rubric-based,
teacher-selected)
2 high-achieving student in laboratory writeups
2 mid-achieving students
2 low-achieving students Represent School A
and School B
Study Focus
All lab write-ups assessed using a 4-point rubric
designed to address performance levels of inquiry
habits
Same rubric criteria and performance levels over 2
years, with several opportunities to revisit these
inquiry habits in different laboratory contexts
Case Studies: Evaluation of a portfolio of lab writeups, Fall 2002-present
Focused on 3 of 18 rubric criteria
Criterion #1: Understanding the
Purpose of the Experiment
Performance
Levels
Insufficient
Evidence
Approaches
Standard
Achieves
Standard
Understanding
the Purpose of
the Experiment
1B. Does not
explain the main
purpose clearly
OR does not use
own words
1B. Explains the
main purpose of
the experiment
clearly in own
words
(Introduction)
(Introduction)
(Introduction)
1B. Explains the
main purpose of
the experiment
clearly in own
words, including
how we will know
if the purpose
has been
achieved
(Introduction)
Criterion
Exceeds
Standard
(Achieves
Standard plus…)
1B. Identifies 2
or more
additional
purposes for
conducting the
experiment
(Introduction)
Criterion #2: Understanding the
Design of the Experiment
Performance
Levels
Insufficient
Evidence
Approaches
Standard
Achieves
Standard
Exceeds
Standard
(Achieves
Standard plus…)
4. Predicts 1 or 2
sources of error
but does not
provide logical
explanations
4. Predicts once
source of error
when conducting
the experiment,
explaining it
logically
4. Logically
predicts 2
sources of error
when conducting
the experiment
4. Proposes 2 or
more ways to
ensure that the
experiment is
conducted safely
and accurately
Criterion
Understanding
the Design of the
Experiment
(Materials and
Methods)
Criterion #3: Analyzing and
Interpreting Data
Performance
Levels
Insufficient
Evidence
Approaches
Standard
Achieves
Standard
Exceeds
Standard
(Achieves
Standard plus…)
4. Offers 1 new
experimental
question or
purposes that is
not directly
related outcomes
of this
experience
4. Offers 1 new
experimental
question or
purposes directly
related to
outcomes of this
experience
4. Offers 2 new
experimental
questions or
purposes directly
related to
outcomes of this
experience
4. Takes one of
the new
experimental
questions and
outlines a design
for a new
experiment
Criterion
Analyzing and
Interpreting Data
(Analysis and
Discussion)
Preliminary Findings
A longitudinal assessment system that culls out discrete inquiry
habits of mind, regardless of the laboratory context, helps
students improve over time
Purpose
More detailed descriptions
Tighter connections
Error
Notion of inherent error is difficult
Requires a “cognitive leap”
Further Experimental Questions
Questions loosely related to the lab at first
Mid & high-level students move toward connections to
results
Additional Questions?
Contact:
Catherine Saldutti (catherine@educhange.com)
Nicolle Gottfried (nmgottfried@sbcglobal.net)
Download