Developing and validating an instrument for student ratings of teaching

advertisement
Developing and Validating an Instrument for Student Evaluation of
Teaching
A Final Report from the Student Experience and Ratings of Teaching
Committee
A subcommittee of the Faculty of Science Academic Standards Committee
Members: Lyn Baldwin, Chair, (Biological Sciences) Les Matthews (Respiratory Therapy) Ernest
Tsui (Physics), Gary Hunt (Center for Teaching and Learning)
Winter 2011
Executive Summary
In May 2007, the Faculty of Science Academic Standards committee established an ad hoc
subcommittee to develop a new student evaluation of faculty teaching. After careful review of the
literature evaluating the effectiveness of student evaluations, we decided that our student evaluation
should focus on the dimensions of teaching that had been shown in previous studies to be highly
correlated with student achievement. We also decided to develop statements that were studentcentred, discipline and pedagogically neutral and were aspects of teaching that could be reasonably
evaluated by students. With this overall philosophy, we embarked on the following steps to develop
the survey: (1) review TRU mission statement, (2) identify teaching dimensions to include in the
survey, (3) prepare statements, (4) review draft survey with faculty and student focus groups, (5)
modify statements, if necessary, (6) run field test of draft survey, (7) analyze results of field test
according to established methods and (8) review statements and finalize form. Based on the results
of our literature survey, the committee decided to include the following six dimensions of teaching
in the survey: preparation and organization, clarity and understandableness, perceived outcome or
importance, stimulation of interest in content, encouragement and openness, and availability and
helpfulness. We received input from 26 students and 17 faculty during focus groups. More than
300 students in 14 different classes participated in our field test of the draft survey. Based on the
results of the field test and the focus groups, statements were re-written, assigned to different
dimensions or eliminated from the survey. In total, the proposed survey includes 40 statements.
The first eight statements ask that students provide background information regarding why they
took the course and their level of involvement with the course. The subsequent 32 statements ask
the students to rate their experience of the course. Finally, we included four open-ended questions
that asked students to reflect upon their experience within the course. The following report
summarizes the findings of the committee and includes our final recommendations to the Science
Faculty Council.
Table of Contents
Pg
I. Introduction
1
II. Methods
7
II. Summary Results
9
IV. Recommendations
10
V. References
11
Appendix A: Proposed student evaluation form
13
Appendix B: Results of Field Test
16
Appendix C: Draft student evaluation form used in field test
22
I. Introduction
Collective Agreement Language
In May 2007 the TRU Science Faculty Council Academic Standards Committee created an ad
hoc subcommittee with the mandate to accept, develop or acquire an instrument for use in
evaluating its members’ teaching performance. This action was mandated by the new TRU /
TRUFA collective agreement which required the inclusion of student evaluations as part of teaching
portfolios that would be used for summative purposes but allowed individual departments, faculties
or schools to develop their own student evaluation tools.
Article 7.3.6.2. states
“ Student evaluations are mandated to be a part of teaching portfolio
— Summative purposes (personnel decisions) Article 7.3.6.2. The summative performance
review for each Member shall be based on:
(a) The Member’s Annual Professional Activity Report per Article 7.2
(b) Evaluation of Member’s performance in teaching for the calendar year under review,
as evidenced by student questionnaires [new instruments and procedures to be
developed based upon department requirements]; and
(c) Class room visitation and assessment by colleague;
Any other evidence that might be relevant.”
Committee Membership
This committee was initially established as a subcommittee of Science Academic Standards
with Ernest Tsui, Les Matthews, and Lyn Baldwin as representatives of the larger committee. In
addition Gary Hunt was invited to join the committee. While he was Resource Librarian at TRU
Library, Daniel Brendle-Moczuk was a member of this committee and was especially helpful in
locating background literature for the committee. Initially, Penny Heaslip, as Coordinator of the
Centre for Teaching and Learning, participated in the committee. Gary Hunt continued on in
Penny’s role following his acceptance of the Coordinator of the Teaching and Learning Center
position.
Underlying philosophy and rationale
The need for a new student evaluation of teaching form can be met through one of several
ways: (1) TRU Science could continue to use the existing student evaluation form, (2) TRU Science
could purchase a commercially produced evaluation form, (3) TRU Science could modify an existing
form, or (4) TRU Science could research and develop its own form. Although the existing
evaluation form has been in place for many years, an informal poll of Science faculty quickly
convinced us that the general attitude towards the document was negative. It became apparent that
Student Experience and Ratings of Teaching Committee Final Report
1
the existing tool did not meet our needs. We were also unable to locate the original development
and validation documentation for the existing form.
The option of adopting a published instrument was appealing because it appeared to be a
simple and practical way to complete our mandate. We gave this option very serious consideration
while weighing the potential time commitment inherent in development of our own evaluation tool.
After reviewing many of the forms used by other universities we came to believe that no existing
form addressed the unique and diverse academic environments represented by TRU Science. In
addition, we also realized that many forms do not incorporate the findings of the latest research on
student evaluation forms into their design. Finally, as this form would be an important tool by
which student input would be gathered for summative evaluations of Science faculty, we believed
that it was critically important for Science faculty to be active participants in the development of this
form. Thus, we decided to develop our own evaluation form.
Our recognition of the failings of many existing evaluation forms made it critical that this
committee delve into the vast body of research surrounding the use of student evaluation forms.
However, given the extraordinary amount of research that has been completed in this area, we
realized that our first task was to clearly articulate the role that we wanted a new student evaluation
form to play within the Faculty of Science. The committee decided that an important goal of the
form would be that it support the educational intent of the Faculty of Science (Gravestock and
Gregor-Greenleaf 2008). While the Faculty of Science currently lacks a mission statement,
committee members felt that we could use TRU’s learner-centred mission statement to guide our
process:
Thompson Rivers University is a comprehensive, learner-centred,
environmentally responsible institution that serves its regional, national, and
international learners and their communities through high quality and flexible
education, training, research and scholarship.
Although learner-centred teaching has many components (Barr and Tagg 1995), we felt that a
critical aspect of learner-centred teaching was linking the influence of teaching activities to student
learning. Given that learner-centred teaching requires that teachers evaluate the role that teaching
activities play in student learning, the committee decided that it would focus the evaluation on those
dimensions of teaching that previous research has found to be most strongly linked to student
learning. By actively tailoring the evaluation to aspects (or dimensions) of teaching that have been
found to support student learning, we believed that we would be constructing a student evaluation
of teaching form that would best support the goals of TRU Science.
However, we also recognized that in order to effectively evaluate their experience within a
course, students cannot be asked to respond to questions for which they are not qualified to answer.
Students are most qualified to report on their experience within the class and are not qualified to
report on the experience of other students (Murray 2005). Thus, we decided that our student
evaluation tool would only include statements that directly reflect a student’s experience within a
Student Experience and Ratings of Teaching Committee Final Report
2
course. To emphasize this distinction, we decided that all statements in the evaluation form would
be written from the student’s perspective. For example, compare the two statements below:
Instructor centred: The instructor was well organized for class
Student centred: I think the instructor was well prepared for class.
Likewise, students do not have the expertise to comment on scope and currency of the
curriculum or an instructor’s depth of knowledge. This is best evaluated by faculty peers (Berk
2006). Finally, the committee recognized that the Faculty of Science represents diverse departments
and it is important that individual statements be pedagogically neutral. That is, as much as possible,
each statement would be equally applicable to all students within all classes. We developed the
following overall principle to guide our selection of teaching dimensions and statements for each
dimension:
We will select measurable factors that relate to student achievement and are both
discipline-independent and pedagogically-independent. All items will be factors that
students are the most qualified to report upon and that faculty can control and
improve upon through targeted faculty development activities.
We decided to use a four-point, forced-choice response scale based on our research of the
literature. A goal of a good survey is to have every response contribute information about teaching
performance. Including a choice of “neutral” in the response scale undermines this goal for a
number of reasons. A rating of “neutral” provides no useful information about teaching
performance. It can mean several different things, including “I have no opinion,” “I don’t
understand this one,” “I don’t care about this one,” “I am too lazy to think about this right now,”or
“I am undecided.” Therefore, in surveys designed to measure teaching effectiveness, “neutral” is an
escape anchor that means the respondent is refusing to take a position on something that was part
of their course experience. If all items are well written and reflect the teaching experiences that all
students were exposed to (as supported by the focus group feedback), then all students are capable
of rendering an informed opinion on all items (Berk 2006). When ratings are generally favourable,
“neutral” responses directly lower the average rating for an item and therefore disadvantage faculty
by producing an underestimate of the true rating (Berk 2006). In addition, when items are well
written and validated through a focus group procedure, the “neutral” position is rarely picked by
students. In a study by Ron Berk (personal communication, October 2010), “neutral” was picked in
fewer than 5% of responses in a sample of 1,500 students in 30 different courses. There is clearly no
need for a choice that is rarely chosen given the effects of the lost information.
Finally, in our form, we wanted to select only those measurable factors or dimensions
relating to student achievements (Table 1) that are discipline and pedagogy-independent so that we
are not biased against any particular teaching method or learning style. And, above all, we wanted to
ensure that, as much as possible, that the factors assessed are within the instructor’s control so that
professional development activities could be targeted towards them.
Student Experience and Ratings of Teaching Committee Final Report
3
Table 1. Dimensions that explain the greatest amount of variance in student achievement
as assessed through studies of multi-section studies.
Dimension Questions from each dimension included in the proposed Student Evaluation Form (Appendix A) % Variance* Preparation and organization 9. I think the instructor was well prepared for class.
10. I think the class sessions were well organized.
11. I clearly understood the relevance of the assignments to the course
objectives.
12. I think the evaluation (all graded material) clearly reflected the course
content.
13. I think the course content was well organized. 30-­‐35%
Clarity and understandableness
14. I clearly understood what I was expected to learn in this course.
15. The time I spent in class helped my understanding of difficult course
content.
16. Examples and illustrations provided in this course aided my
understanding.
17. I think the instructor communicated the course material clearly.
18. I think the instructor delivered the course material at a pace I could
follow.
19. I clearly understood how my work would be evaluated in this course. 24-­‐30%
Perceived Outcome or Impact 20. I learned skills in this course that I will be able to use in other courses. 21. I learned ways of reasoning that I could apply to other subjects.
22. I think the instructor made the course content relevant to my overall
education
23. The instructor helped me understand the relevance of the material to the
real world.
24. I felt the instructor presented the course material in a way that challenged
me to think.
15-­‐20% Stimulation of interest in content
25.
26.
27.
28.
29.
10-­‐15%
I think the instructor was enthusiastic about the course content.
I felt comfortable participating in class activities.
My experience in the class increased my interest in the course content.
I was engaged in learning the course content during class time
My interactions with the instructor encouraged me to learn. Encouragement and 30. I think the instructor was approachable.
31. The class atmosphere supported my learning.
openness
32. I was treated with respect in this class.
33. I felt encouraged to ask questions in class.
34. I think that the instructor was receptive to suggestions from students. <10%
35. I was satisfied with the time it took for the instructor to return graded material.
36. The instructor provided me with all the information I needed to seek
help.
37. I felt welcome to seek help from the instructor.
38. I think the instructor made a genuine effort to be available outside of
class.
39. I think the instructor cared about my learning.
40. The feedback I received on work that I completed was helpful to my
learning. <10%
Availability and helpfulness
*Variance explained in student achievement: Source: Feldman (1989, 2007), work substantiated by (Abrami et al. 2007a,
Abrami et al. 2007b).
Student Experience and Ratings of Teaching Committee Final Report
4
Literature Review
A. Effectiveness of student evaluations
It is now well established that there is a general positive correlation between student ratings
and student achievements as measured by test scores (Cohen 1981, Kulik 2001). This means that for
teachers who had been rated highly by their students, their students tended to perform better in
turn. This trend had been confirmed by over 40 surveys conducted in the last 20-30 years. On the
whole, this indicates that student ratings do provide relatively reliable information on teaching
effectiveness.
Much of current research establishes that teaching is a highly complex and multifaceted
activity. Among the very many things that instructors do, it is not possible to select just one and use
it as a measure for effective teaching. Cohen (1981), Feldman (1989), Marsh (2007) and others have
identified certain “instructional dimensions” (e.g., traits and practices of the instructor) that correlate
best with student achievement and learning (e.g. organization, clarity, and stimulation). So, for
student evaluation of teaching (SET) to be useful, it must reflect this inherent multidimensionality of
teaching.
Another important point that the research literature (Murray 2005, Adams et al. 2008) makes is that
the best SET instruments are those that reflect the learners’ class experiences or perspectives, and
not the instructor’s. Students can only comment on factors that are observable in class. Therefore,
questions concerning non-classroom factors, such as course design and instructor knowledge,
should be left out of the ratings forms.
Given the complexity of teaching, it is vital that any student feedback be complemented by
evaluations from peer and expert observers, as well as self-evaluations (Murray 2005, Adams et al.
2008). This is necessary to ensure that the whole picture of an instructor’s teaching is evaluated.
However, one must be careful about any “overall” or “global” statements such as, “Overall, I would
rate this instructor’s teaching performance as …” These are quite common items in student
questionnaires, but since they do not reflect the multidimensionality of teaching, the information
they provide may not be very meaningful at best, and can be misleading at worst (Murray 2005).
Global items are recommended by some experts (Abrami 2001, Arreola 2007) because they are
correlated moderately with student learning and they provide a single-value summary of teaching
(Berk 2006). Because global items provide no value for formative feedback and are less reliable than
subscale and total scale scores, we chose not to include them in this survey.
Recent research shows a well designed SET tool is possible and it can provide valid and
reliable information on teaching; but its impact on instructional improvement is limited if it is not
accompanied by the necessary professional development activities (Cohen 1981, Feldman 1989,
2007, Wachtel 1998, Kulik 2001, Murray 2005, Marsh 2007). When we were thinking about the
design of our new ratings form, it is precisely this formative side of the form that we wanted to
emphasize. In other words, we would like the SET instrument to be used as a tool to help
instructors improve their teaching skills, in addition to its current use in summative or personnel
decision processes. Furthermore, we needed to ensure the teaching practices addressed in the form
align with the educational goals and objectives of TRU Science (e.g., learner-centredness and student
Student Experience and Ratings of Teaching Committee Final Report
5
learning). Our decision to make the new ratings form based primarily on learners’ experiences
addresses this.
B. Effectiveness of online student evaluations
A review of the literature summarizing advantages and disadvantages of administering surveys
online can be found in Gravestock and Greenleaf (2008). Advantages of the online format include:
•
Provision of security through unique login codes for each student.
•
Reducing cost compared to paper, but there is a significant cost to manage and
distribute the data (Donmeyer et al. 2004).
•
Not requiring use of class time (but see below for response rate comments).
•
Rapid processing of results and no manual typing of comments.
•
Generating customized reports, including graphical representation of changes over years
for each item (Llewellyn 2003).
•
Allowing more extensive and thoughtful comments on open-ended, online questions
when students are not restricted by time (Sorenson and Reiner 2003, Donmeyer et al.
2004).
•
Facilitating the addition of items specific to programs, courses, and individuals
(Sorenson and Reiner 2003, Donmeyer et al. 2004).
•
Standardizing the introduction and instructions for taking the survey, which makes them
less susceptible to faculty influence (Donmeyer et al. 2004).
The experience at other institutions shows that the response rate for online surveys decreases
dramatically compared to surveys completed during class time (see Avery et al. 2006 for a review of
related research). Decreases in response rate have been attributed to concern for anonymity (Avery
et al. 2006), students having to use their own time (Sorenson and Reiner, 2003), and technical
problems (Sorenson and Reiner 2003). Online surveys can be completed during class time if classes
complete them as a group in a computer lab. Large classes would need to be divided into separate
groups. When students are asked to take the survey outside of class, positive incentives and
reminders can be used to encourage participation and these can increase response rates (Donmeyer
et al. 2004).
Studies have examined the validity and reliability of online evaluation systems compared to
paper-based surveys. Although the research is limited, results suggest that ratings are generally
consistent and that variations are statistically insignificant (Donmeyer et al. 2004, Avery et al. 2006).
Student Experience and Ratings of Teaching Committee Final Report
6
It is important that the software for online surveys allows for secure data storage and controlled
distribution of data to specific individuals. It must be able to generate useful and complete summary
reports of the results. It should have the flexibility to add items specific to programs, courses, and
individuals according to the desires of faculty.
II. Methods
Based on the methods clearly outlined in (Berk 2006) and (Gravestock and Gregor-Greenleaf 2008),
the committee adopted the following steps to develop a new student evaluation form
1. Review TRU mission statement in light of literature concerning student evaluations, Summer
2008.
2. Identify teaching dimensions to be included in student evaluation, Summer 2008
Based on our survey of the literature, we chose the dimensions that showed the
highest correlation with student learning in multi-section classes. While the
committee recognizes that “learning” is extraordinarily difficult to quantify, we felt
that these studies provided the best estimate of the teaching dimensions that fit with
the overall purpose of our new student evaluation form.
3. Prepare statements, Fall 2008
To prepare statements, we compiled a list of statements that had been used in
student evaluations at other universities—particularly those universities such as the
University of Victoria and the California State University Channel Islands (Adams et
al. 2008) that had recently evaluated and reconfigured their student evaluation forms.
Once we had a master list of potential statements, we classified them by dimension.
Furthermore, we also rewrote each question to be student-centred. Because
statements can suffer from a multitude of problems that limit their effectiveness, we
used the list of rules that Berk (2006) compiled from numerous sources to ensure
that our draft statements were as problem-free as possible. These rules for writing
statements range from grammatical guidelines to rules regarding the relevance or
applicability of each statement to potential student respondents.
4. Hold focus groups with faculty and students, Fall 2009
In Sept/Oct 2009, the committee organized focus groups with both Science students
and Science faculty. In order to encourage participation, we posted signs advertising
the focus groups within the Science Building and asked faculty to announce it to
their students. Faculty were invited to participate via email solicitations.
In total, 17 faculty and 26 students participated in the focus groups. Both students
and faculty in each focus group were asked to complete three separate tasks:
1. Rank the importance of all statements in each dimension
2. Respond to a series of questions regarding the applicability and neutrality of all
statements.
Student Experience and Ratings of Teaching Committee Final Report
7
3. In groups, compile a list of behaviors that would comprise “high-ranking” and
“low-ranking” behaviors for each statement in each dimension.
5.
6.
Modify statements based on focus groups, if necessary.
Run field test with approx. 300 students, Winter 2009.
At the end of the Winter Semester, we asked for faculty volunteers to have their students
complete the online, draft student evaluation. Fourteen faculty participated volunteering 16
courses, including four lab sections, one online course and 10 lecture courses. The courses
also ranged across the curriculum from 100-400 level courses. Students who participated in
the focus groups were well distributed from within Science (Table 2).
Table 2. Program and number of students from each program who participated in the
field test of the online, draft student evaluation.
Program of participating students
ANHT
BIOL
CHEM
GEOL
NRSC
PHYS
RESP
Unreported
Total
Number of students
44
186
37
4
44
9
30
12
362
7. Analyze data, Summer 2010
In order to evaluate the reliability and validity of proposed questions, Berk (2006) suggests
the following minimum quantitative analysis:
a. Determine the mean and standard deviation of each item in the survey.
b. Assess inter-item correlations between statements in each dimension.
c. Asses item-scale correlations.
Given the importance the committee placed on creating statements that were as
pedagogically neutral as possible, it was important to identify questions that were not
applicable to a significant number of students. Thus in our field test, we included a “Not
Applicable” (NA) possibility in the response scale and determined which of the statements
received the highest number of “not applicable” from students.
Student Experience and Ratings of Teaching Committee Final Report
8
8. Review statements and finalize form, Fall 2010.
Based on the results of the field test (see below), the committee selected several statements
to eliminate from the final proposed evaluation form based on their correlation (or lack of
correlation) with other items on the scale.
9. Develop “report” for faculty/admin, Fall 2010/Winter 2011
III. Summary results of the field test
Overall, the global mean score provided for all faculty on all items was 3.36 (out of a
possible 4.0). Mean scores for individual items ranged from 2.95 to 3.64. While this is higher than
the recommended 2.0 mean (with a 4 point scale, Berk 2006), the committee believes that the
relatively high scores may have resulted from the self-selection process whereby faculty who
typically have high student evaluation scores were the most likely to volunteer their classes for the
survey. The standard deviation of each item indicates whether or not each item solicited a range of
responses from students. The standard deviations of items evaluated in the field tests ranged from
0.5 to 0.87 which Berk (2006) indicates is an adequate range of variation.
Statements, or items, within each dimension are meant to evaluate the same teaching
dimension. Correlations between mean scores for each statement evaluate how well the statements
correlate with one another. Berk (2006) recommends that all correlations should be positive and
should be as high as possible. Inter-item correlations resulting from our field test ranged from 0.30
to 0.73. Overall, teachers who score high in one category typically score high in the other teaching
dimensions and item-scale correlations evaluate how well the score for each item correlates with the
global score for all items (minus the specific item being evaluated). The item-scale correlations
ranged from 0.52 to 0.79. In addition to inter-item and item-scale correlations, we also evaluated the
number of NA responses each item solicited in the field test. Based on these three items, inter-item
correlations, item-scale correlations and the number of NA responses, we identified two statements
that should be deleted from the survey and have moved two statements into a more appropriate
dimension (Table 3).
Table 3. List of questions that were deleted or moved to another teaching dimension based
on the results of the field test. See Appendix C for full text of survey used in field test.
Dimension
Full text of Question
Q#
Action
Preparation
and
I
was
satisfied
with
the
time
it
took
for
the
Moved to Dimension 6 based
R6
Organization
instructor to return graded material.
R10
Clarity and
Understanding
R16
The work I completed in this course
(e.g., assignments, homework or class
activities) increased my learning.
Perceived
Outcome or
Impact
Stimulation of
interest in
course material
My problem solving skills improved as a
result of this course
Deleted due to high NA
responses
I felt the instructor presented the course
material in a way that challenged me to
think.
Moved to Dimension 3 based
on low inter-item correlation
coefficients
R24
Student Experience and Ratings of Teaching Committee Final Report
on low inter-item correlation
coefficients
Deleted due to high NA
responses
9
IV. Recommendations
1. Adoption of new tool
Based on the completed work, the committee recommends that the Faculty of Science
adopt the proposed student evaluation (Appendix A) as their new tool to evaluate faculty
teaching. We recommend that summary statistics (item and global means) be monitored for several
years on a faculty-wide level so that baseline data can be collected to establish faculty wide means.
This information will provide critical benchmarks to both external and internal reviewers of our
faculty teaching.
2. Global vs. sub-scale means
As indicated above, the committee intentionally avoided the use of global statements to
evaluate student experience within classrooms. Sub-scale means are often averaged across student
evaluation tools to provide a global mean for individual faculty. However, global means estimated
across all statements obscure the relative importance of each dimension (i.e., Preparation and
Organization explains 30-35% of student achievement, while Availability and Helpfulness explains
<10% of student achievement). While global means may be weighted by the amount each
dimension explains of overall student achievement (Table 1), any single value produced by a student
evaluation masks the details of variation among the dimensions. The committee is concerned that
reviews of faculty member’s teaching (especially summative reviews such as those involved in
promotion and tenure), would rely too heavily upon a global mean, if produced. Overall, the
committee strongly recommends that the Faculty of Science not include global means in
any report on a faculty member’s teaching.
3. Online versus paper implementation
Based on our review of the literature, we believe there are distinct obstacles and advantages
to using online surveys. Given the potential environmental advantage of a paper-less form, we
decided to field test our survey using an online approach. During both the field test and focus
groups, students responded very favorably to online surveys. However, as our understanding of the
logistics of implementing online surveys has developed, we recognized that there is no current
software on campus that will allow an online implementation. Based on this, the committee
recommends that the Faculty of Science collectively decide how to proceed with
implementation.
Acknowledgements
We gratefully acknowledge the contribution of the students and faculty who participated in
either the focus groups or the field test. This work would not have been possible without their
participation. We would also like to acknowledge the Dean of Science for providing financial
support for hosting the focus groups. Finally, we thank the Faculty of Science for their patience
during the three years of this project.
Student Experience and Ratings of Teaching Committee Final Report
10
V. References
Abrami, P. 2001. Improving judgements about teaching effectiveness using rating forms. in M.
Theall, P. Abrami, and L. Mets, editors. The student ratings debate: are they valid? How
can we best use them? Jossey-Bass, San Francisco, CA.
Abrami, P., S. d’Apollonia, and S. Rosenfield. 2007. The Dimensionality of Student Ratings of
Instruction: What We Know and What We Do Not*. Pages 385–456 in R. Perry and J.
Smart, editors. The Scholarship of Teaching and Learning in Higher Education: An
Evidence-Based Perspective.
Abrami, P., S. Rosenfield, and H. Dedic. 2007. The dimensionality of student ratings and instruction:
an update on what we know, do not know and what we need to do. in R. Perry and J. Smart,
editors. The Scholarship of Teaching and Learning in Higher Education: An Evidence-Based
Perspective. Springer.
Adams, V., R. Bleicher, M. Buchanan, E. Nuhfer, J. Elliot, M. Furmanski, C. Renny, P. Smith, and
G. Wood. 2008. Final report of the Task Force to Create a New Student Ratings for
California State University Channel Islands.
Arreola, R. 2007. Developing a comprehensive faculty evaluation system: a handbook for college
faculty and administrators on designing and operating a comprehensive faculty evaluation
system. Anker Publishing Co., Bolton, MA.
Avery, R., W. Bryan, A. Mathios, H. Kang, and D. Bell. 2006. Electronic course evaluation: does an
online delivery system influence student evaluation? Journal of Economic Education 37:2137.
Barr, R., and J. Tagg. 1995. From teaching to learning: a new paradigm for undergraduate education.
Change Nov/Dec:13-25.
Berk, R. 2006. Thirteen strategies to measure college teaching. Stylus, Sterling, VA.
Cohen, P. A. 1981. Student ratings of instruction and student achievement: a meta-analysis of
multisection validity studies. Review of Educational Research 51:281 -309. doi:
10.3102/00346543051003281.
Donmeyer, C., P. Baum, R. Hanna, and K. Chapman. 2004. Gathering faculty teaching evaluations
by in-class and online surveys: their effects on response rates and evaluations. Assessment
and Evaluation in Higher Education 29:611-623.
Feldman, K. 1989. The association between student ratings of specific instructional dimensions and
student achievement: refining and extending the synthesis of data from multisection validity
studies. Research in Higher Education 30:583-645.
Feldman, K. 2007. Identifying exemplary teachers and teaching: evidence from student ratings.
Pages 93-143 in R. P. Perry and J. C. Smart, editors. The Scholarship of Teaching and
Learning in Higher Education: An Evidence-Based Perspective. Springer Netherlands,
Dordrecht.
Gravestock, P., and E. Gregor-Greenleaf. 2008. Student course evaluations: research, models and
trends. Higher Education Quality Council of Ontario.
Kulik, J. A. 2001. Student ratings: validity, utility, and controversy. New Directions for Institutional
Research 2001:9–25.
Llewellyn, D. C. 2003. Online reporting of results for online student ratings. New Directions for
Student Experience and Ratings of Teaching Committee Final Report
11
Teaching and Learning 2003:61–68.
Marsh, H. 2007. Students' evaluation of university teaching,: dimensionality, reliability, validity,
potential biases and usefulness. Pages 319-383 in R. P. Perry and J. C. Smart, editors. The
scholarship of teaching and learning in higher education: An evidence-based perspective.
Springer Verlag, Dordrecht.
Murray, H. 2005. Student evaluation of teaching: has it made a difference? Paper presented at the
Annual Meeting of the Society for Teaching and Learning, Charlottetown, PEI.
Sorenson, D. L., and C. Reiner. 2003. Charting the uncharted seas of online student ratings of
instruction. New Directions for Teaching and Learning 2003:1–24.
Wachtel, H. K. 1998. Student evaluation of college teaching effectiveness: A brief review.
Assessment & Evaluation in Higher Education 23:191–212.
Student Experience and Ratings of Teaching Committee Final Report
12
Appendix A: Proposed Student Evaluation Form for TRU Faculty of Science
Student Feedback on Course Learning Experience
Introduction
You are a critical source of information about the effectiveness of the instruction that you have received.
Your thoughtful responses are appreciated and will be used to identify aspects of your instruction that are
meeting your learning needs and those that need to be improved.
Please complete the survey alone, not in consultation with your classmates.
Student feedback will remain confidential and responses will be returned to the instructor only after all grades
have been submitted to the registrar.
This information will be used by individual faculty to improve their teaching. In addition, information from
this survey will be made available to department Chairs and Deans for the purpose of assessing instructors.
This questionnaire contains three sections. In the first section, we would like you to tell us a little about
yourself.
The second section contains statements about your learning experience with your instructor. Please read each
statement carefully and the rate the extent to which you agree with the statement as a reflection of your
experience in the class. Consider each statement separately and assess it based on your actual experience.
Finally, the third section asks you to comment more generally about your experience within the course.
Please answer these questions in the most constructive and objective way possible.
I. Background Information
1. My program of study is _________________________________
2. My year in this program of study is [pull-down menu of 1 to 4]
3. My reasons for taking the course are …
(Check all that are applicable)
ž Interest
ž Program requirement
ž Program elective
ž Reputation of the instructor
ž Reputation of the course
ž Course fit in my timetable
4. Of all classes and other sessions scheduled for this course (e.g., labs, tutorials, etc), I attended
approximately …
ž 90% or more
ž 70-89%
ž 50 to 69%
ž 20 to 49%
ž less than 20%
Student Experience and Ratings of Teaching Committee Final Report
13
5. My anticipated grade in this course is …
ž A
ž B
ž C
ž D
ž F
The next three statements concern your involvement with the course. You will rate the instructor’s teaching
in the remaining statements. Please respond using the following scale.
Strongly Disagree—this statement definitely does not reflect my experience in any way
Disagree—this statement in general does not reflect my experience
Agree—this statement generally reflects my experience
Strongly Agree—this statement definitely reflects my experience in all cases
Strongly Disagree
Disagree
Agree
Strongly Agree
6. I asked the instructor for additional guidance or feedback when I needed it.
7. I came to class prepared (e.g., reviewed posted notes, read from the course text or completed other
activities as directed by the instructor) even if it was not going to be graded.
8. I think that the instructor’s main role is to explain all the course content, not to make students think about
it.
II. Ratings of Teaching in this course
9.
10.
11.
12.
13.
14.
15.
16.
I think the instructor was well prepared for class.
I think the class sessions were well organized.
I clearly understood the relevance of the assignments to the course objectives.
I think the evaluation (all graded material) clearly reflected the course content.
I think the course content was well organized.
I clearly understood what I was expected to learn in this course.
The time I spent in class helped my understanding of difficult course content.
Examples and illustrations provided in this course aided my understanding.
17. I think the instructor communicated the course material clearly.
18. I think the instructor delivered the course material at a pace I could follow.
19. I clearly understood how my work would be evaluated in this course.
20.
21.
22.
23.
24.
25.
26.
27.
28.
29.
I learned skills in this course that I will be able to use in other courses.
I learned ways of reasoning that I could apply to other subjects.
I think the instructor made the course content relevant to my overall education
The instructor helped me understand the relevance of the material to the real world.
I felt the instructor presented the course material in a way that challenged me to think.
I think the instructor was enthusiastic about the course content.
I felt comfortable participating in class activities.
My experience in the class increased my interest in the course content.
I was engaged in learning the course content during class time
My interactions with the instructor encouraged me to learn.
Student Experience and Ratings of Teaching Committee Final Report
14
30.
31.
32.
33.
34.
35.
36.
37.
38.
39.
40.
I think the instructor was approachable.
The class atmosphere supported my learning.
I was treated with respect in this class.
I felt encouraged to ask questions in class.
I think that the instructor was receptive to suggestions from students.
I was satisfied with the time it took for the instructor to return graded material.
The instructor provided me with all the information I needed to seek help.
I felt welcome to seek help from the instructor.
I think the instructor made a genuine effort to be available outside of class.
I think the instructor cared about my learning.
The feedback I received on work that I completed was helpful to my learning.
III. Knowing what you know now about the course, if it were possible to turn back time and you
could experience this course again….
1. What changes would you make in your own approach in order to improve your learning?
2. What aspects of the course would you advise your instructor to retain?
3. What suggestions would you provide to your instructor for revisions that would produce a better learning
experience for you?
4. Do you have any other comments about your learning experience in this class?
Student Experience and Ratings of Teaching Committee Final Report
15
Appendix B. Summary statistics from field test administered at the end of Winter Semester
2010.
Table B1. Mean and standard deviation for each rating questions in the online evaluation of
teaching. Refer to Appendix C entire student evaluation form.
Question
I think the instructor was well prepared for class.
R1
I think the class sessions were well organized.
R2
I clearly understood the relevance of the
R3
R4
R5
R6
R7
R8
R9
R10
assignments to the course objectives.
I think the evaluation (all graded material) clearly
reflected the course content.
I think the course content was well organized.
I was satisfied with the time it took for the
instructor to return graded material.
I clearly understood what I was expected to learn
in this course.
The time I spent in class helped my understanding
of difficult course content.
Examples and illustrations provided in this course
aided my understanding.
The work I completed in this course (e.g.,
assignments, homework or class activities)
increased my learning.
mean
3.56
3.44
3.38
sd
0.64
0.70
0.72
3.37
0.71
3.35
3.39
0.72
0.75
3.29
0.75
3.25
0.76
3.49
0.64
3.27
0.73
R11
I think the instructor communicated the
course material clearly.
3.38
0.72
R12
I think the instructor delivered the course material
at a pace I could follow
3.32
0.69
R13
I clearly understood how my work would be
evaluated in this course.
3.29
0.73
R14
I learned skills in this course that I will be able to
use in other courses.
I learned ways of reasoning that I could apply to
other subjects.
My problem solving skills improved as a result of
this course.
I think the instructor made the course content
relevant to my overall education
3.21
0.71
3.11
0.67
2.95
0.78
3.20
0.70
The instructor helped me understand the
relevance of the material to the real world.
I think the instructor was enthusiastic about the
course content.
I felt comfortable participating in class activities.
3.34
0.70
3.64
0.61
3.35
0.71
R15
R16
R17
R18
R19
R20
Student Experience and Ratings of Teaching Committee Final Report
16
R21
R22
R23
R24
R25
R26
R27
R28
R29
R30
R31
R32
R33
R34
My experience in the class increased my interest in
the course content.
I was engaged in learning the course content
during class time
3.24
0.78
3.17
0.78
My interactions with the instructor encouraged me
to learn.
I felt the instructor presented the course material
in a way that challenged me to think.
I think the instructor was approachable.
The class atmosphere supported my learning.
I was treated with respect in this class.
I felt encouraged to ask questions in class.
I think that the instructor was receptive to
suggestions from students.
The instructor provided me with all the
information I needed to seek help.
I felt welcome to seek help from the instructor.
I think the instructor made a genuine effort to be
available outside of class.
I think the instructor cared about my learning.
The feedback I received on work that I completed
was helpful to my learning.
3.24
0.80
3.26
0.70
3.59
3.35
3.61
3.43
3.46
0.71
0.73
0.57
0.71
0.70
3.44
0.67
3.52
3.52
0.69
0.66
3.46
3.28
0.70
0.77
Student Experience and Ratings of Teaching Committee Final Report
17
Table B2. Inter-item correlations for each dimension included in the draft evaluation.
Dimension 1
R1
R2
R3
R4
R5
R6
Dimension 2
R7
R8
R9
R10
R11
R12
R13
Dimension 3
R14
R15
R16
R17
R18
Dimension 4
R19
R20
R21
R22
R23
R24
R1
1.00
0.76
0.50
0.50
0.68
0.48
R2
0.76
1.00
0.58
0.52
0.72
0.41
R3
0.50
0.58
1.00
0.68
0.58
0.30
R4
0.50
0.52
0.68
1.00
0.61
0.40
R5
0.68
0.72
0.58
0.61
1.00
0.49
R6
0.48
0.41
0.30
0.40
0.49
1.00
R7
1.00
0.68
0.51
0.51
0.68
0.56
0.62
R8
0.68
1.00
0.61
0.53
0.63
0.55
0.61
R9
0.51
0.61
1.00
0.55
0.62
0.52
0.48
R10
0.51
0.53
0.55
1.00
0.59
0.49
0.44
R11
0.68
0.63
0.62
0.59
1.00
0.66
0.63
R12
0.56
0.55
0.52
0.49
0.66
1.00
0.56
R14
1.00
0.65
0.51
0.57
0.45
R15
0.65
1.00
0.65
0.55
0.50
R16
0.51
0.65
1.00
0.59
0.49
R17
0.57
0.55
0.59
1.00
0.68
R18
0.45
0.50
0.49
0.68
1.00
R19
1.00
0.46
0.52
0.50
0.56
0.43
R20
0.46
1.00
0.54
0.51
0.55
0.42
R21
0.52
0.54
1.00
0.73
0.60
0.53
R22
0.50
0.51
0.73
1.00
0.60
0.57
R23
0.56
0.55
0.60
0.60
1.00
0.57
Student Experience and Ratings of Teaching Committee Final Report
R13
0.62
0.61
0.48
0.44
0.63
0.56
1.00
R24
0.43
0.42
0.53
0.57
0.57
1.00
18
Dimension 5
R25
R26
R27
R28
R29
Dimension 6
R30
R31
R32
R33
R34
R25
1.00
0.59
0.65
0.61
0.68
R26
0.59
1.00
0.67
0.61
0.58
R27
0.65
0.67
1.00
0.63
0.65
R28
0.61
0.61
0.63
1.00
0.68
R29
0.68
0.58
0.65
0.68
1.00
R30
1.00
0.72
0.62
0.62
0.63
R31
0.72
1.00
0.73
0.75
0.62
R32
0.62
0.73
1.00
0.72
0.63
R33
0.62
0.75
0.72
1.00
0.68
R34
0.63
0.62
0.63
0.68
1.00
Student Experience and Ratings of Teaching Committee Final Report
19
Table B3. Item-scale correlations for each ranking question included in the field test.
Question Correlation
R1
0.639922
R2
0.66873
R3
0.711612
R4
0.71664
R5
0.695163
R6
0.540815
R7
0.734069
R8
0.716955
R9
0.64647
R10
0.664676
R11
0.797651
R12
0.654795
R13
0.725151
R14
0.572724
R15
0.566435
R16
0.52591
R17
0.683259
R18
0.65343
R19
0.683475
R20
0.631581
R21
0.73735
R22
0.689508
R23
0.792114
R24
0.622447
R25
0.689815
R26
0.708809
R27
0.687586
R28
0.67249
R29
0.739747
R30
0.744615
R31
0.771491
R32
0.664126
R33
0.75697
R34
0.749575
Note: each item was removed from grand mean before correlations were run.
Student Experience and Ratings of Teaching Committee Final Report
20
Table B4. Number of respondents choosing NA. “B”
Questions indicate background questions. “R” questions indicate
rating of teaching questions.
Question #
B11
B12
B13
R1
R2
R3
R4
R5
R6
R7
R8
R9
R10
R11
R12
R13
R14
R15
R16
R17
R18
R19
R20
R21
R22
R23
R24
R25
R26
R27
R28
R29
R30
R31
R32
R33
R34
Number of Respondents who chose NA when
responding to question
21
5
3
0
1
8
3
0
5
0
5
1
16
0
0
0
7
20
33
4
5
0
7
6
6
13
3
0
3
5
5
5
4
5
13
2
13
Student Experience and Ratings of Teaching Committee Final Report
21
Appendix C. Draft Student Evaluation Form used in field test
Student Feedback on Course Learning Experience
Field-tested survey at end of Winter 2010
This survey is being administered as part of a research project entitled “Validating a new student ratings of
teaching survey in the Faculty of Science.” This study is being conducted by members of the Science faculty
(Dr. Lyn Baldwin, Dr. Gary Hunt, Dr. Ernest Tsui, and Mr. Les Matthews) who have been given the job of
creating a new survey form that students will complete to evaluate and rate the teaching in their courses. The
current survey form has not been updated for many years. As part of developing a new form, we need to
administer the proposed survey to students currently enrolled in Science classes. We will then analyze the
results and make sure that the new survey form provides data that is statistically reliable and valid. The
development of a new student evaluation survey will allow students to provide constructive feedback to their
instructors. By completing this survey you are consenting to participate in the study. You may stop at any
time, or refuse to answer any question without consequence. This survey is anonymous—you do not have to
identify yourself as part of completing the survey. All information will remain confidential and will be
destroyed 10 years after the completion of this research project. The entire survey should take approximately
15-20 minutes to complete. If you have questions, please contact Dr. Lyn Baldwin (lybaldwin@tru.ca, phone
250-377-6167. Thank you for your participation.
Introduction
You are a critical source of information about the effectiveness of the instruction that you have received.
Your thoughtful responses are appreciated and will be used to identify aspects of your instruction that are
meeting your learning needs and those that need to be improved.
Please complete the survey alone, not in consultation with your classmates.
Student feedback will remain confidential and responses will be returned to the instructor only after all grades
have been submitted to the registrar.
This information will be used by individual faculty to improve their teaching. In addition, information from
this survey will be made available to department Chairs and Deans for the purpose of assessing instructors.
This questionnaire contains three different sections. In the first section, we would like you to tell us a little
about yourself.
The second section contains statements about your learning experience with your instructor. Please read each
statement carefully and the rate the extent to which you agree with the statement as a reflection of your
experience in the class. Consider each statement separately and assess it based on your actual experience.
Finally, the third section asks you to comment more generally about your experience within the course.
Please answer these questions in the most constructive and objective way possible.
I. Background Information
1. My program of study is _________________________________
2. My year in this program of study is [pull-down menu of 1 to 4]
3. My reasons for taking the course are …
(Check all that are applicable)
Student Experience and Ratings of Teaching Committee Final Report
22
ž
ž
ž
ž
ž
ž
Interest
Program requirement
Program elective
Reputation of the instructor
Reputation of the course
Course fit in my timetable
4. Of all classes and other sessions scheduled for this course (e.g., labs, tutorials, etc), I attended
approximately …
ž 90% or more
ž 70-89%
ž 50 to 69%
ž 20 to 49%
ž less than 20%
5. My anticipated grade in this course is …
ž
ž
ž
ž
ž
A
B
C
D
F
The next three questions ask you to reflect upon your involvement with the course. Please answer these
questions using the following response scale.
Strongly Disagree—this statement definitely does not reflect my experience in any way
Disagree—this statement in general does not reflect my experience
Agree—this statement generally reflects my experience
Strongly Agree—this statement definitely reflects my experience in all cases
Strongly Disagree
Disagree
Agree
Strongly Agree
Not applicable
6. I asked the instructor for additional guidance or feedback when I needed it.
7. I came to class prepared (e.g., reviewed posted notes, read from the course text or completed other
activities as directed by the instructor) even if it was not going to be graded.
8. I think that the instructor’s main role is to explain all the course content, not to make students think about
it.
Student Experience and Ratings of Teaching Committee Final Report
23
II. Ratings of Teaching in this course
Response scale:
Strongly Disagree—this statement definitely does not reflect my experience in any way
Disagree—this statement in general does not reflect my experience
Agree—this statement generally reflects my experience
Strongly Agree—this statement definitely reflects my experience in all cases
Not Applicable—this statement does not apply in any way to the experience I had with the instructor.
Strongly Disagree
Disagree
Agree
Strongly Agree
Not applicable
Dimension 1: Preparation and Organization
1. I think the instructor was well prepared for class.
2. I think the class sessions were well organized.
3. I clearly understood the relevance of the assignments to the course objectives.
4. I think the evaluation (all graded material) clearly reflected the course content.
5. I think the course content was well organized.
6. I was satisfied with the time it took for the instructor to return graded material. (Question moved to
Dimension 6)
Dimension 2: Clarity and Understanding
1. I clearly understood what I was expected to learn in this course.
2. The time I spent in class helped my understanding of difficult course content.
3. Examples and illustrations provided in this course aided my understanding.
4. The work I completed in this course (e.g., assignments, homework or class activities) increased
my learning. Deleted due to high NA responses
4. I think the instructor communicated the course material clearly.
5. I think the instructor delivered the course material at a pace I could follow.
6. I clearly understood how my work would be evaluated in this course.
Dimension 3: Perceived Outcome or Impact
1.
2.
3.
4.
5.
I learned skills in this course that I will be able to use in other courses.
I learned ways of reasoning that I could apply to other subjects.
I think the instructor made the course content relevant to my overall education
The instructor helped me understand the relevance of the material to the real world.
My problem solving skills improved as a result of this course. Delete due to high NA responses
Dimension 4: Stimulation of interest in course content
1. I think the instructor was enthusiastic about the course content.
2. I felt comfortable participating in class activities.
3. My experience in the class increased my interest in the course content.
4. I was engaged in learning the course content during class time
5. My interactions with the instructor encouraged me to learn.
Student Experience and Ratings of Teaching Committee Final Report
24
6. I felt the instructor presented the course material in a way that challenged me to think. (This question
moved to Dimension 3 based on low inter-item correlation coefficients).
Dimension 5: Encouragement and Openness
1. I think the instructor was approachable.
2. The class atmosphere supported my learning.
3. I was treated with respect in this class.
4. I felt encouraged to ask questions in class.
5. I think that the instructor was receptive to suggestions from students.
Dimension 6. Availability and helpfulness
1.
2.
3.
4.
5.
The instructor provided me with all the information I needed to seek help.
I felt welcome to seek help from the instructor.
I think the instructor made a genuine effort to be available outside of class.
I think the instructor cared about my learning.
The feedback I received on work that I completed was helpful to my learning.
III. Knowing what you know now about the course, if it were possible to turn back time and you
could experience this course again….
1. What changes would you make in your own approach in order to improve your learning?
2. What aspects of the course would you advise your instructor to retain?
3. What suggestions would you provide to your instructor for revisions that would produce a better
learning experience for you?
4. Do you have any other comments on your learning experience in this class?
Student Experience and Ratings of Teaching Committee Final Report
25
Download