Student Teacher Linkage Data Quality: Discrepancy Rates and

advertisement
Student Teacher Linkage Data Quality: Discrepancy Rates and Team Teaching
Jeff Watson
Chris Thorn
UW Madison
Wisconsin Center for Education Research
Value-Added Research Center
Center for Data Quality and Systems Innovation
Paper Presented at the Association for Education Finance Policy 37th Annual Conference
March, 2012
Boston, Massachusetts
Background and Purpose
Urban districts invest significant resources in data systems and information technology infrastructure
with the goal of informing decision making, supporting evaluation, and enabling planning. Increasingly,
districts are trying to use their data systems to inform decisions related to measuring and improving the
rates at which students are learning (e.g., using value-added). One necessary condition of developing
classroom-level estimates of student growth is that a district’s data systems must provide information
about which teachers are teaching which students. While it is unrealistic to expect districts to have
perfect data about who is teaching whom, we should be very concerned about if student – teacher
linkage data are good enough to support this new use. This paper reviews our work with one urban
district that was designed to assess the quality of their ST linkage data as well as the extent to which
team teaching may impact the attribution of student learning to specific teachers.
Linking students to teachers sounds like a function that any district data system should be able to do.
However, many states and districts are challenged when they try to implement systems that require
high quality student–teacher linkage data. The Data Quality Campaign (DQC) has focused on providing
resources state-level policy makers need to guide strategic planning for their states’ information
technology infrastructure. In 2009, the DQC found that 29 states self reported the ability to link
students to teachers (DQC, 2009). However, in a 2010 brief, the DQC reported that most states and
districts lack the capacity to know about team-teaching, schools’ organizational models, or the ability to
verify and improve the quality of their student–teacher linkage data (DQC, 2010). Battelle for Kids
recently published a white paper that identified several sources of errors in district data systems that
can impact the quality of student– teacher linkages. These deficiencies included not knowing about
course schedules, school- and grade-level regrouping students during the year, student and teacher
mobility, team teaching methods, course content and course-assessment alignment, as well as charter
school teaching and learning (Battelle for Kids, 2010).
As the stakes associated with these new policies increase, the need for accurate and valid student –
teacher linkage data also increases. As a result, the process of validating student – teacher linkage data
is becoming more common place. While organizations like the DQC and Battelle for Kids are helping
raise awareness of the factors that can impact the data quality of student teacher linkages, little
attention has been given to documenting the extent to which student – teacher linkage data accurately
portrays who is teaching whom. Watson et. al (2010) reported results from a pilot study where 10% of
records were validated as incorrect by teachers and another 15% of records were flagged as having
received team teaching. In other words, only 3 out of 4 records were verified as accurate and valid.
We designed the current study to expand the 2010 pilot study to a much larger sample of schools and
teachers. We assessed the accuracy of student-teacher linkage data by asking teachers to verify the
math and reading courses they were teaching as well as the student rosters for those courses. In
addition to assessing the accuracy of the district’s data, we also asked teachers to report the extent to
which they used team teaching.
Methods
Sample
Working with district leaders, we identified 49 schools to invite to participate in this study. Teachers
were invited to participate via email as well as through printed letters of support and periodic requests
for their participation from their principals. Over a three week period, 348 teachers from 47 schools
reviewed their student-teacher linkage data for 8824 students. As a result, 16,318 student records were
verified by teachers.
Procedure
We developed a web-based application called the Student Teacher Verification System (STVS) for the
purpose of efficiently verifying teacher – course assignments and student rosters. Initial student-teacher
linkage data was extracted from the district’s student information system and loaded into the STVS. We
also worked with the district prior to verification to recruit schools, train principals on the system, and
describe the purpose of the study. STVS was field tested in 4 schools 2 weeks prior to the
implementation of the study to ensure that they system was working properly on the computers that
were typical for the district.
Teachers completed the following steps:
1)
2)
3)
4)
5)
6)
7)
Log In
Confirm courses
Indicate content area
Indicate course-level team teaching
Confirm students for Math and Reading courses
Indicate student-level team teaching
Log out
The entire process required about 15 minutes to complete. Teachers who did not wish to participate
simply did not log in to the system. After 3 weeks, the teacher participation window was closed and
Principals logged into the system to resolve any discrepancies that may have resulted from teachers'
participation. Specifically, STVS asks principals to review and assign any unclaimed students or classes.
Findings and Conclusions
Student-teacher linkage discrepancies
In general, our findings suggest that teacher-course assignments were pretty accurate and student
rosters were also in pretty good shape for this district. Out of 16318 records verified, only 428 (2.6%)
were flagged by teachers as being discrepant. Incorrect teacher – course assignments accounted for .7%
of these while incorrect student-roster data accounted for 1.9% (see Figure 1).
17443
13296 (81.5%)
Records
reviewed
No Team Instruction
100%
Self-Attribution
15890 (97.4%)
16318 (100%)
Records Included
in Analysis
Correct at both
course- and studentlevel
2594 (15.9%)
Some Team
Instruction
775 (4.7%)
Data Quality
111 (.7%)
428 (2.6%)
Incorrect at either
course- or studentlevel
Course-level
Course-level
discrepancy
1422 (8.7%)
317 (1.9%)
Student-level
58.4%
Average SelfAttribution
63.5%
Course Level
Self-Attribution
49.7%
Student Level
Self-Attribution
Student-level
discrepancy
397 (2.4%)
1125 incomplete records excluded
BOTH courseand studentlevels
38.4%
Both Levels
Self-Attribution
Team Teaching
Figure 1- Data Quality Logic Model. Discrepancy frequencies for disputed records, Team teaching
frequencies, and the average self-attribution of instruction.
However findings showed that discrepancy rates varied by school and teacher. Using a weighted
average, we observed that 1.8 records per teacher were disputed (SDweighted = 7.3). Furthermore, 80%
of the discrepancies originated from just 6% of participating teachers (N=22) (Figure 2). The highest
discrepancy rate observed for a teacher was 98% (83 out of 84 records) and another 4 teachers had
discrepancy rates at 50% or higher. Ten teachers reported discrepancy rates at 10% or higher and less
than 50%. 19 teachers reported rates between 5% and 9.9%. Thirty-seven teachers had rates higher
than 0% but less than 5% and 278 teachers reported no discrepancies.
Cumulative Percent of Discrepancies by
Teacher
1
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
1
13
25
37
49
61
73
85
97
109
121
133
145
157
169
181
193
205
217
229
241
253
265
277
289
301
313
325
337
0
Figure 2- Cumulative percent of discrepancies by teacher - Twenty-two teachers accounted for 80% of
reported discrepancies.
We observed similar patterns when we aggregate discrepancies at the school-level. Using a weighted
average, we observed only 9.1 discrepancies per school (SDweighted = 13.7). However, eleven schools
accounted for 80% of the discrepancies reported and seventeen schools reported no discrepancies
(Figure 3). The highest observed school-wide discrepancy rate was 49% (based on 169 records). Two
other schools had rates above 10% and three more had rates between 5 and 10%. Twenty-five schools
had rates between 0% and 5%, whereas sixteen schools had no discrepancies.
One of the challenges that remain, despite the generally high quality data in most schools, is how to
successfully identify the schools and classrooms in which error rates are high. Our finding of 80% of the
errors in 11 of the schools 49 schools suggests that the deployment of a modern student information
system is not enough to address data quality challenges by itself. The task for policy makers is to explore
the sources of error in this group of schools. Anecdotal evidence from other urban districts suggests that
issues of staff training, leadership turnover, and novel organizational models (such as looping or multigrade classrooms that don’t fit the standard SIS use case) have all been associated with higher error
rates. Without a systemic understanding of the source of this error to more explicitly address its causes,
system-wide verification may be the only way to ensure fairness for high stakes use of student-teacher
linkage data.
Cumulative Percentage of Discrepancies
by School
1
0.8
0.6
0.4
0.2
0
V
I AF C AP D AJ H AO S AC AA AQ AL AT AB E G
L
P
Z AH AK AS
Figure 3- Cumulative Percentage of Discrepancies by School. About 80% of discrepancies were observed
in about 20% of the schools included in this study.
Team teaching discrepancies
Our data suggests that team teaching practices are commonly used, and on average teachers only
attributed about 60% of a student’s instruction to themselves when team teaching methods were used.
Of the 15890 records that were verified as accurate, teachers flagged 2594 as receiving instruction
through team teaching practices such as shared classrooms, pull-out instruction for English language
learners, or push-in services for special education instruction (Figure 1). When aggregated by school,
team teaching was reported unevenly across schools in that about half of the schools accounted for
about 80% of the records flagged as receiving team teaching (Figure 4). When teachers reported team
teaching at the course-level (e.g., shared classrooms), the percent of instruction they attributed to
themselves was bimodally distributed at 50% and 75% levels. Self-attribution of instruction for studentlevel team teaching also appeared bimodal, but a higher percentage of teachers estimated selfattribution below <40% (Figure 5).
Cumulative Percentage of Team Teaching
by School
100.0%
80.0%
60.0%
40.0%
20.0%
0.0%
M AI AF Y
F N T AN J AR B R P A G AU AT U E
L AM AA AJ AS
Figure 4- Cumulative Percentage of Team Teaching by School. Team teaching practices were less
clustered than data quality discrepancies, but were nonetheless distributed unevenly across schools.
0.4
0.35
0.3
0.25
0.2
0.15
Course
0.1
Student
0.05
91-100
81-90
71-80
61-70
51-60
41-50
31-40
21-30
11-20
1-10
0
0
% of Instruction Attributed to Self
Figure 5- Distribution of Self-Attribution of Instruction in Team Teaching Situations.
Team teaching presents a similar, but slightly different challenge than the “distribution of error”
problem presented above. While team teaching at the course and individual student level reflects of the
reality of a substantial number of students in the sample, it is also analytically much more complicated
and is a policy challenge for school leaders who are now required in many states and districts to provide
individual effectiveness ratings to individual educators . The fact that approximately 16% of the sample
included some level of team teaching indicates that for those teachers, the system will only be able
produce a single measure of the teachers’ joint effort – not a measure of each teacher’s contribution.
While not a problem for program evaluation of school-level effectiveness studies, it does provide a
substantial challenge for those charged with producing individual measures of educator productivity.
References
Battelle for Kids, 2009 – The Importance of Accurately Linking Instruction to Students to Determine
Teacher Effectiveness (http://static.battelleforkids.org/images/BFK/Link_whitepagesApril2010web.pdf)
Data Quality Campaign, 2010 - Strengthening the Teacher-Student Link to Inform Teacher Quality Efforts
(http://www.dataqualitycampaign.org/files/TSDL_abstract.pdf)
Data Quality Campaign, 2010 - Effectively Linking Teachers and Students
(http://www.dataqualitycampaign.org/files/DQC7_14.pdf)
Watson, J., Thorn, C., Ponisciak, S. & Boehm, F. (2011, March). Measuring the Impact of Team Teaching
on Student-Teacher Linkage Data. Paper presented at the 36th annual meeting of the Association for
Education Finance and Policy, Seattle, Washington.
Download