Employer feedback on graduates and summaries of the results

advertisement
Exhibit 1.4.h
1
Exhibit 1.4.h Employer Feedback on Graduates and Summaries of the Results
Précis:







Mean ratings since 2005-2006 indicate program improvement from initial “good” to “very good” ratings in content
area knowledge, general and content specific teaching methods, with specific strengths noted in professional
behavior and ethics and addressing issues of diversity.
Mean ratings since 2007-2008 suggest that candidates are as well or better prepared than employees from other
programs.
Issues related to ratings in assessing learning are being addressed through program and assessment changes such
as the Teacher Performance Assessment, co-teaching to emphasize co-teaching, and the use of an analysis of
student work in methods courses.
Inconsistent reports between mentors and employers related to infusing technology are being addressed through
conversations with faculty members to modify course content.
Classroom management findings will be continued to be explored with partnership panels, mentor teachers, and
our emphases on increasing engagement and differentiation.
Since 2009-2010 graduates have performed well on general education competencies as rated by employers.
Despite various efforts and consultation with our professional community, response rates are not what we would
desire.
The educator preparation unit recognizes that schools and school districts are stakeholders in our efforts to
improve the outcomes for students with whom our graduates work. During the 2003-2004 academic year, we
mailed brief survey forms to principals and superintendent faculty members that we knew had employed our
graduates. The return of these surveys was too poor to aggregate or report meaningfully (4 responses). During
the 2004-2005 academic year, we posed the question to our Partnership Panel, an advisory board designed to
provide additional stakeholder participation in the design, implementation, and evaluation of our field
experiences. In these and program specific advisory group meetings, the need for a revised survey and a means
for distribution and collection was discussed. In 2005-2006, several strategies were employed to increase
participation in the survey, as suggested by these groups. In addition, discussions took place as to whether the
survey should be annual or biannual. Beginning 2007-2008 the survey was administered annually.
In our effort to increase responses, we identified principals of schools and superintendents of districts who were
reported as employers of our graduates on the Ohio Department of Jobs and Families Services (ODJFS) Graduate
Employment Report. We also used a state Teacher Quality Performance measure follow-up study data base to
identify additional principals of schools in which our graduates worked. The results of these efforts were not
what we hoped:
 Of the 43 sent to principals on the ODJFS listing, 12 were received (28%)
 Of the 25 sent to superintendents on the ODJFS listing, 3 were received (12%)
 Of the 102 sent to Principals on the Teacher Quality Performance follow-up list, 18 were received (18
percent.
 Overall, 170 surveys were sent, and 33 were received (19%).
During the 2007-2008 academic year, the University of Cincinnati opted out of continuing the Teacher Quality
Partnership survey. During this year, we initiated sending emails with a link to a survey to principals and
superintendents named on the ODFJS listings. Fifteen were returned (25%). However, feedback from principals
and superintendent partnership panel meetings indicated that they were “always getting surveys.” Beginning
2008-2009, surveys were distributed at the partnership meetings, with postage-paid envelopes, and a request to
complete and return the survey. Limited responses were returned. We continue to talk with principals,
Exhibit 1.4.h
2
superintendents, members of the Partnership Panel, and community members of the University Council for
Educator Preparation about strategies to increase responses.
The Survey and Aggregated Data The survey was developed using phrases generated from institutional
standards. In 2007-2008, at the suggestion of the Partnership Panel, an additional section comparing candidates
to those from other programs was added. In 2009-2010 an additional set of questions, developed from the
general education student outcomes, was added to gather stakeholder data. The overall ratings are provided in
Table 1.7 T1. Scale for quality of graduate in each area: 5 Excellent
4 Very Good 3 Good 2 Fair 1 Poor
Scale for comparison: 5 much stronger 4 Stronger 3 about the same 2 weaker 1 much weaker Table 1.7 T1
Employer Perceptions of Unit Program Completers
Ratings of observations of UC program
completers/employees:
Content area knowledge
General teaching methodology
Content specific teaching methodology
Planning lessons and units
Delivering lessons to students
Assessing learning
Professional behavior and ethics
Classroom management
Infusing technology
Dealing with issues of diversity
Communication Skills
Social Responsibility
Critical Thinking
Knowledge integration
Information literacy
Comparison with grads of other programs:
Content area knowledge
General teaching methodology
Content specific teaching methodology
Planning lessons and units
Delivering lessons to students
Assessing learning
Professional behavior and ethics
Classroom management
Infusing technology
Dealing with issues of diversity
Communication Skills
Social Responsibility
Critical Thinking
Knowledge integration
Information literacy
N
2005-2006
2007-2008
2008-2009
2009-2010
2010-2011
3.97
3.91
3.78
3.88
3.72
3.94
3.56
3.84
3.76
4.36
4.18
4.18
4.27
4.36
4.36
4.60
4.45
4.30
4.18
4.11
4.05
3.96
3.93
3.85
4.17
3.83
3.56
3.86
3.79
3.79
3.79
3.57
4.21
3.93
3.64
3.69
4.69
4.65
4.62
4.38
4.38
3.46
4.77
3.86
2.92
3.76
4.64
4.14
4.21
4.08
3.79
4.50
4.29
3.64
4.29
2009-2010
4.92
4.92
4.69
3.92
4.30
2010-2011
3.14
3.79
3.21
3.21
3.14
3.21
3.36
3.14
3.29
3.86
3.21
3.93
4.14
3.21
3.14
14
3.15
3.07
3.15
3.07
3.23
3.23
3.07
3.15
3.15
4.23
3.15
3.46
3.07
3.07
3.07
13
Items added 2009-2010 to gather additional
information regarding general education
outcomes
2005-2006
2007-2008
2008-2009
Items added
2007-2008
suggestion of
partnership
panel
3.36
3.45
3.45
3.55
3.45
3.36
3.55
3.36
3.40
3.50
4.29
4.14
3.70
3.87
3.35
3.57
3.34
3.30
3.31
3.27
Items added 2009-2010 to gather additional
information regarding general education
outcomes
33
15
18
Exhibit 1.4.h
3
Analysis The following charts include (a) the mean scores of University of Cincinnati graduates as judged by
employers; and (b) the mean comparison of University of Cincinnati graduates with peers from other programs.
Trend lines are provided for mean scores of University of Cincinnati graduates.
Scale for quality of graduate in each area: 5 Excellent
4 Very Good 3 Good 2 Fair 1 Poor
Scale for comparison: 5 much stronger 4 Stronger 3 about the same 2 weaker 1 much weaker
Mean ratings of content area knowledge show a positive trend between 2005 and 2011 as judged by employers. Mean
ratings of comparisons with employees from other programs indicated that graduates were consistently judged to be at
least as strong as employees from other programs; during 2008-2009 mean ratings were much stronger for University of
Cincinnati graduates. Mean ratings of general teaching methods showed a positive trend as well; mean ratings suggested
that our graduates were at least as strong as other graduates from other programs.
Exhibit 1.4.h
4
Mean ratings suggest that content specific teaching methods, planning, and delivering lessons to students are strengths for
our candidates; mean ratings demonstrate a mild positive advantage over employees from other programs.
Exhibit 1.4.h
5
Mean ratings of graduates’ ability to assess learning have been inconsistent with a negative trend line. Programs addressed
this issue through a variety of strategies. Initial licensure programs initiated a pilot the Teacher Performance Assessment in
autumn 2009, requiring an analysis of student work and documentation of the findings of this assessment to differentiate
instruction. Secondary and middle childhood education coursework was co-taught with general and special education
faculty members, with a fresh emphasis on assessment and differentiation. We anticipate that in subsequent years we will
see a positive change in employers’ ratings of graduates’ ability to assess learning in response to these program changes.
Data gathered from our candidates’ dispositions reports document that for most programs, 98-100% of all candidates
demonstrate strong professional behavior and ethics. This finding is confirmed by employers’ ratings of graduates in this
area.
Exhibit 1.4.h
6
Classroom management is an area of concern of employers consistently voiced at Partnership Panel meetings. Graduates,
however, are demonstrating management comparable to that of employees from other programs. With the use of the
Teacher Performance Assessment, increased emphasis on student engagement, and increased differentiation we anticipate
that ratings of classroom management will improve. However, we have encountered a subtheme of “controlling students”
among some participants in these meetings. This “control” is inconsistent with our emphases on collaborative learning
strategies and developing students’ voice. Further conversations may clarify expectations related to classroom
management.
Our mentor teachers indicate that candidates use a wide range of technology in their placements. The data from
employers, however, is of great concern. We have also found that not all candidates have the technology skills necessary
for the videotaping, compressing, and uploading of evidence for the Teacher Work Sample. During summer 2011 meetings
were conducted with faculty members who teach this course to discuss these finding and concerns. With the addition of
the Information Technology program to the College of Education, Criminal Justice, and Human Services in Autumn 2011
there is an increased potential for instructional growth in technology.
Exhibit 1.4.h
7
With the emphasis on culturally responsive teaching and developing a conversation of teaching and learning among
candidates and students representing a wide range of diversity, we are please with employers’ recognition that our
candidates are very strong both in their practice and in comparison with employers from other programs.
In 2009 we added items to the employer survey which addressed the general education student outcomes for the
University of Cincinnati. Mean ratings suggest that candidates are both “very good” in these skill areas and are as strong as
or stronger than employees from other programs.
Download