Response rate - Institutional Research

advertisement
On-line Course Evaluation Implementation
and Improvement of Response Rates
Marcia Belcheir, Ph.D.
Robert Anson, Ph.D.
James A. Goodman, Ph.D.
Boise State University
© 2012 Boise State University
Goals of the Session
• Offer suggestions for implementing on-line
course evaluations
• Report research results of how faculty can
improve response rates in their courses
© 2012 Boise State University
A short history on implementing on-line course
evaluations at Boise State
• Diffused ownership – collaboration of stakeholders
from all parts of campus examined viability of the
move
• Most departments wanted to participate; some
skeptical, needed to see it work on campus
• No university-wide question set; instead, from
college, department, and a variety of academic
support programs
• Purchased CollegeNet, “WDYT” product
© 2012 Boise State University
Academic-Oriented Team
Led by Institutional Research
• Teamed with 2 faculty—one in Center for Teaching and
Learning, one teaches IT
• Minimized OIT involvement (only prepare data source
and set up user authentication)
• Implementation required:
– No new FTE resources (“other duties as assigned”)
– Need “sensitive touch” for faculty buy-in and participation
• To Opt-In
• To encourage student involvement
• To trust data for annual performance evaluations
© 2012 Boise State University
Implementing On-Line Evaluations
in a Streamlined Environment
• Coordinated many hands (and no FTE body) to run online evaluations
– Developed upload files and checked for accuracy
– Handled help questions
– Prepared survey calendar each semester
• Recommendations
– Reduce # of sessions = more sessions = more cycles to
support
– Don’t assume course registration data is accurate
– Communication is key
© 2012 Boise State University
Implementing On-Line Evaluations
in a Voluntary Environment
• Opt-In for colleges, departments (down to course level)
created challenges to selectively load course enrollment data
• Also opt-in for trans-departmental academic programs-“matrix university”
• Solution
– Developed MS Access Database to check, filter and restructure data
flexibly and systematically
• Recommendation
– Overcomes potential faculty and department resistance to how
teaching is evaluated
– (Optional) University-wide question set should be universally adopted
early in the process
© 2012 Boise State University
Implementing On-Line Evaluations
in a Decentralized Environment
• Questions controlled by colleges, departments, and academic
support programs; no required university question-set
• Solution
– Vendor (CollegeNet, WDYT) selected for flexibility to create and bind
specific question sets to type of course
– Centralized question review, assistance, creation in Center for
Teaching and Learning
• Recommendations
– Decentralized requires centralized editing & question support
– Helps to incorporate academic support programs (Honors? On-line
programs? SL?) in on-line evaluations to reduce extra survey demands
© 2012 Boise State University
Part 2: Improving response rates
The key to good student response rates for on-line
evaluations often lie with the faculty
➢What does the literature say?
➢What do our faculty do?
➢How well do different tactics work to raise
response rates?
➢Are there other factors?
© 2012 Boise State University
Background and literature review
• Response rates for online evaluations tend to be
lower than of pen/paper evaluations (~70-80%/~5060%) (Avery et al, 2006; Dommeyer 2002; many
others)
• Likert-type quantitative ratings remain constant even
with lower response rates (Anderson et al, 2005;
Dommeyer, 2004; many others)
© 2012 Boise State University
Background and literature review
• Qualitative responses tend to increase
(Handwerk, 2000; Heath, et al, 2007)
• Institutions and instructors that do something
tend increase response rates (Nulty, 2008)
• The more tactics an institution and/or
instructor uses, the higher the response rate
(Nulty, 2008)
© 2012 Boise State University
Response Rates and
Adoption of On-line Evaluations
* Estimate
based on
literature
© 2012 Boise State University
➢Online Evaluations
–
Fall 2012
76,172 enrollments (88% of total were online evaluations)
•
•
1,134 instructors
3,213 classes
➢Response Rates
– Overall: 54.9%
– Colleges
• High: 73% in Engineering
• Low: 52% in Arts & Sciences
– Departments
• High: 82% in Construction Mgmt
• Low: 31% in Respiratory Care
– Classes: 0% to 100%
© 2012 Boise State University
University Measures/Tactics
• No iPad lotteries, web advertising, or grade hooks
• Sent 5 student emails
– Start
– Reminders (3)
– End
• Sent 5 instructor emails
–
–
–
–
–
Pre-eval (enter custom questions)
Start
Reminder
End
Results available
© 2012 Boise State University
Research Methods: Data
1. Studied Fall 2012 evaluation
– Used individual response rate data
– Received anonymous data set from vendor
2. Surveyed all 1,134 instructors
– 678 (60%) instructors responded about tactics used in
1 or 2 classes taught that Fall
– 1,128 (35%) classes studied
– Survey asked
• Tactics used to increase student response rates or otherwise
obtain student feedback
• Basic demographic questions
© 2012 Boise State University
Research Methods: Analysis
• Applied series of t-tests and ANOVAS on
individual variables for statistical differences
• Applied regression for modeling the best set
of factors for predicting response rates
© 2012 Boise State University
Research Questions Addressed
• What tactics do instructors use to improve
their response rates?
• Does it help to use more tactics?
• What is the impact of course and instructor
characteristics?
• What is the best model for explaining
variation in response rates?
• Are there interactions between tactics used
and course or faculty characteristics?
© 2012 Boise State University
Which Tactics Were Instructors
Most Likely to Use?
• Reminded students during class (61%)
• Explained to class how I use the results to improve
teaching (57%)
• Sent personal emails to students as reminders (33%)
• Posted a reminder or assignment on Blackboard
(32%)
• Provided incentives to complete the evaluation
(22%)
© 2012 Boise State University
What instructor tactics improved response rates?
Tactics
Used tactic Didn’t use
Response
Response
rate
rate
Provided incentives
79%
57%
Provided time in class to complete
70%
61%
Sent personal e-mails as reminders
66%
60%
Reminded students during class time
65%
57%
Explained to class how I use the results
65%
57%
Posted reminder/assignment on BlackBoard
64%
61%
I did nothing
50%
63%
© 2012 Boise State University
Incentives: Does the kind of incentive
and approach affect response rates?
Type
of
Points
Incentive
Other incentive
Total
© 2012 Boise State University
Basis for Awarding
Incentive
Individual % of class
Total
78%
(n=62)
77%
(n=15)
79%
(n=156)
82%
(n=18)
79%
(n=218)
80%
(n=33)
78%
(n=77)
79%
(n=174)
79%
(n=251)
Class-based Incentive Examples
• Point-based class-wide rewards
– If 80% complete the eval, everyone gets 1 point
added to final grade
– If 90% complete eval, all students gain back all
participation points
• Non-Point-based class-wide rewards
– If 70% complete eval, instructor will bring cupcakes
to final
– If 90% complete eval, everyone can bring a 3 X 5
notecard to the final
© 2012 Boise State University
Impact of threshold on response rate
Incentive threshold
100%
N who used that
threshold
8
Average response
rate
89%
90%+
10
87%
80%+
22
86%
Sliding scale
8
73%
70%+
12
68%
60%+
3
62%
Unknown
9
74%
© 2012 Boise State University
Did using more tactics improve response rates?
80%
71%
70%
60%
60%
50%
50%
74%
64%
54%
40%
30%
20%
10%
0%
0
© 2012 Boise State University
1
2
3
4
5 or more
Course and instructor characteristics:
Higher response rates were associated with:
Small courses (<20)
vs larger (>40)
In-person courses
vs On-Line
Graduate courses
vs Undergraduate
Tenure/tenure-track faculty
vs Adjunct
© 2012 Boise State University
Response
Rate
64%
vs 59%
63%
vs 50%
65%
vs 61%
64%
vs 57%
N
477
vs 154
955
vs 119
153
vs 974
748
vs 379
Final model (R2=.32):
Variable
Standardized estimate
Provided incentives
0.41
Number of tactics used
0.43
Took class time to complete evaluation
0.07
Small class
0.09
Taught on-line
-0.09
Undergraduate course
0.04
Taught by adjunct
-0.12
Interaction: Number of tactics &
undergraduate course level
-0.32
© 2012 Boise State University
Response rates based on number of tactics and course level
0.90
0.80
0.70
Response rate
0.60
0.50
0.40
Graduate course
0.30
Undergraduate course
0.20
0.10
0.00
0
1
2
3
4
5
Number of tactics employed to increase response
rates
© 2012 Boise State University
Conclusions
• Incentives provide the biggest boost
– If you use incentives, set the threshold for the
class (not individual) and set it at 80% or more
– If using incentives, best to set high threshold (80%
or greater) for class as a whole
• Use multiple tactics
• On-line and large section classes have a built
in disadvantage—use even more tactics
© 2012 Boise State University
Future Changes
• Institutionalize small core set of universitywide questions to enable comparisons
• Pushing grade hook proposal
• Keep expanding use (now 91.5%)
• Explore means to impact response rates for
on-line courses
© 2012 Boise State University
Questions?
• Marcia Belcheir - mbelcheir@boisestate.edu
• Robert Anson – ranson@boisestate.edu
• James A. Goodman –
jamesgoodman1@boisestate.edu
© 2012 Boise State University
Download