Using Surveys to Evaluate Sakai: Goals & Results

advertisement
Using Surveys to Evaluate Sakai:
Goals & Results
Moderator: Stephanie Teasley, University of Michigan
Panelists:
Steve Lonn, University of Michigan
Salwa Khan, Texas State University
Jeff Narvid, University of California, Berkeley
Angelica Risquez, University of Limerick
Confluence Page: http://bugs.sakaiproject.org/confluence/pages/viewpage.action?pageId=18063
Goals/Purpose of This Session
• Discuss why feedback from users is
important for implementing Sakai.
• Share general purposes of evaluations.
• Share common question types and
categories.
• Share general survey statistics.
• Answer specific questions about
evaluation using surveys and other
methods.
2
Panel Participants
• Steve Lonn
CTools
Usability, Support, and Evaluation (USE) Lab, University of Michigan
• Salwa Khan
TRACS
Instructional Technologies Support, Texas State University
• Angelica Risquez
Sulis
Centre for Teaching and Learning, University of Limerick
• Jeff Narvid
bSpace
Training & Support Team, University of California, Berkeley
3
Stage of Implementation
• Michigan - CTools
•
•
Full Production (about 45,000 users in Winter 2006)
Was running Sakai 2.0 during survey (now running Sakai 2.1.2)
• Texas State - TRACS
•
•
Pilot (F 05), Expanded Pilot (Sp 06), Voluntary Production (F 06)
Running Sakai 2.1
• Limerick - Sulis
•
•
Pilot (Spring 2006), Pilot (Fall 06-Spring 07)
Running Sakai 2.1.2
• Berkeley - bSpace
•
•
Pilot (F 05) / Pre-Production (Sp 06) / Full Production (F 06)
Running Sakai 2.1.1
4
Purpose of Survey Evaluation
(Slide 1 of 2)
• Gather users’ demographic data
• Determine users’ level of computer proficiency
• Determine level of system usage
• Users’ perceptions of the system (ease of use,
satisfaction, reliability, usefulness of tools)
5
Purpose of Survey Evaluation
(Slide 2 of 2)
• Determine users’ perception of system benefits
• Determine quality of support / user preferences
in obtaining support
• Determine users’ desired improvements
• Determine users’ pain points / barriers to
migration
6
General Survey Statistics:
Michigan
• 1,357 instructor respondents (19% response rate)
All instructional faculty (including student instructors) invited to
respond
• 2,485 student respondents (27% response rate)
25% of students (stratified random sample) invited to respond
• Surveys administered online
April 2006 (end of winter semester)
Incentive: Random drawing for four $100 gift certificates
• Surveys completed in 10-15 minutes, on average
7
General Survey Statistics:
Texas State
• 11 faculty/staff respondents
Surveys completed in 15-20 minutes, on average
• 798 student respondents
Specific courses surveyed during class period
Surveys completed in 5-7 minutes, on average
• Surveys administered online for faculty/staff, paper for
students
End of semester
8
General Survey Statistics:
Limerick
• 175 faculty/staff respondents (no questions related to Sakai)
•
All faculty/staff invited via email by the Dean of T&L, aprox
46% response rate
• 897 student responses (aprox 9% total population), 200 of
them Sulis (Sakai) users (aprox 58% users)
•
•
All undergraduate and taught postgraduate students invited
via general email
Instructors using Sulis offered a voluntary evaluation of their
students in a class setting
• Surveys administered online (www.markclass.com) and paper
based (Sulis users)
•
Completed in 10-20 minutes
9
General Survey Statistics:
UC Berkeley
• Fall 2005: 15 faculty/staff respondents
• Spring 2006: 22 faculty/staff respondents
All faculty/staff invited via Message of the Day and an
announcement on the bSpace community page
• Fall 2005: 101 student respondents
• Spring 2006: 110 student respondents
All students invited via Message of the Day and an
announcement on the bSpace community page
• Surveys administered online (zoomerang)
Administered at the end of Fall and Spring semesters
Completed in 10 minutes, on average
10
Demographic Questions
•
•
•
•
•
•
Gender
Age
Years Teaching
Year in School (students)
Department / College
etc.
11
Demographic Data
• Michigan: Asked college/school and students’
year in program
• Texas State: Asked gender, age group,
years at Texas State, staff/faculty member
• Limerick: Asked gender, age, rural/urban,
nationality, college, year of study, working relation,
position, main activity in university
• UC Berkeley: N/A
12
General Computer Use & Proficiency
• Rate your expertise with computers
Novice, Intermediate, Advanced
• Rate expertise with other computerbased tools
• Use of / preference for information
technology in your courses
None, Limited, Moderate, Extensive,
Exclusive (Online Only)
13
General Computer Use & Proficiency:
Michigan, Texas State, Limerick Results
•Computer Expertise - Instructors:
Q: Rate your expertise with computers
100%
91.2%
*
83.0%
75%
62.2%
50%
32.8%
25%
17.0%
8.8%
5.0%
0%
Novice
UM: Instructors
Intermediate
TXSU: Instructors
Advanced
UL: Instructors
10
14
General Computer Use & Proficiency:
Michigan, Texas State, Limerick Results
•Computer Expertise - Students:
Q: Rate your expertise with computers
100%
75%
65.0%
*
64.0%
68.0%
50%
32.5%
27.0%
25%
17.0%
3.6%
8.2%
4.0%
0%
Novice
UM: Students
Intermediate
TXSU: Students
Advanced
UL: Students
10
12
15
Benefits of Using a System Like Sakai
• Most valuable benefit of using
information technologies
Improved teaching/learning
Convenience
Manage course activities
Faculty-student communication
Student-student communication
16
Benefits of Using a System Like Sakai
• Michigan:
• Instructors: Improved instructor to student communication (38%)
•
and efficiency (23%)
Students: Efficiency (45%) and helps manage course activities
(17%)
• Texas State:
• Instructors: Most answered helped manage course activities and
helped communicate with students
• Limerick
• Instructors: Improved instructor to student communication (31.4%)
•
and improved students’ learning (29%)
Students: (Sulis users) Manage activities (62%); increased student to
student communication (28,5%). (Non-Sulis users): improved
learning (27%) and faculty-student commun. (19%)
17
Sakai Use
• How often do you visit your Sakai sites?
Stopped using
Few times a semester
Few times a month
Once a week
Few times a week
Daily (once or more)
• For how many courses have you used
Sakai?
None, 1-2, 3-4, 5 or more
18
Sakai Use:
Michigan, Texas State, Berkeley Results
Q: (Instructors) How often do you visit your Sakai sites?
75%
64.0%
58.3%
60%
45%
41.7%
40.0%
33.4%
30%
19.0%
15%
9.4%
7.5%
0%
Few Times a Semester
Few Times a Month
UM: Instructors
Weekly (Once or more)
TXSU: Instructors
Daily (or More)
UCB: Instructors
19
Sakai Use:
Michigan, Texas State, Limerick, & Berkeley Results
Q: (Students) How often do you visit your Sakai sites?
75%
68.3%
60%
55.7%
54.6%
48.0%
45%
35.0%
30.0%
30%
21.0%
15%
12.6%
3.4%
0%
18.0%
11.9%
10.9%
5.9%
Few Times a Semester
UM: Students
7.0%
5.8%
2.4%
Few Times a Month
TXSU: Students
Weekly (Once or more)
UL: Students
Daily (or More)
UCB: Students
17
20
Overall Perceptions of Sakai
• Overall dimensions of use
Frustrating – Satisfying
Difficult – Easy
Useless – Useful
Unreliable – Reliable
•
•
•
•
Learning to use (Easy - Hard)
Time to learn (Slow – Fast)
Exploration of features (Risky – Safe)
Tasks can be performed in straightforward
manner (Never – Always)
• Feedback on completion of steps (Unclear –
Clear)
21
Overall Perceptions of Sakai:
Reliability: Michigan & Texas State Results
Instructors
75%
58.3%
60%
45%
41.4%
41.7%
45.4%
30%
15%
0%
9.8%
1.1%
2.3%
Unreliable Somewhat Unreliable
UM: Instructors
Neutral
Somewhat Reliable
Reliable
TXSU: Instructors
22
Overall Perceptions of Sakai:
Reliability: Michigan & Texas State Results
Students
75%
60%
50.2%
45%
41.0%
41.0%
31.2%
30%
27.0%
15%
6.0%
2.0%
14.0%
12.0%
9.0%
26.0%
24.0%
13.4%
3.2%
0%
Unreliable Somewhat Unreliable
UM: Students
Neutral
TXSU: Students
Somewhat Reliable
Reliable
UL: Students
23
Overall Perceptions of Sakai:
Learning to Use: Michigan & Texas State Results
Instructors
75%
60%
50.0%
44.9%
45%
30%
41.6%
23.9%
23.5%
15%
6.0%
8.3%
1.6%
0%
Difficult
Somewhat Difficult
UM: Instructors
Neutral
Somewhat Easy
Easy
TXSU: Instructors
20
24
Overall Perceptions of Sakai:
Learning to Use: Michigan & Texas State Results
Students
75%
60%
48.6%
45%
41.0%
43.0%
35.3%
30%
15%
0%
25.7%
22.2%
8.2%
11.0%
14.5%
17.5%
16.4%
7.0%
1.3% 3.0%
Difficult
Somewhat Difficult
UM: Students
Neutral
TXSU: Students
Somewhat Easy
Easy
UL: Students
20
22
25
Specific Uses of Sakai Tools/System
•
•
•
•
•
•
•
•
•
•
Announcements
Assignments – submission & grading
Exams & quizzes (sample & real)
Online folders for student work (Drop Box)
Peer review
Posted questions for students
Resources / online readings & materials
Schedule / Online Calendar
Syllabus
Other tools / uses
26
Specific Uses of Sakai Tools/System:
Online Readings, Documents (Resources)
Instructors
100.0%
100%
89.6%
80%
66.7%
60%
40%
33.3%
20%
8.9%
0%
1.5%
Have Not Used
UM: Instructors
Not Valuable
TXSU: Instructors
Valuable/Very Valuable/Used
UCB: Instructors
17
27
Specific Uses of Sakai Tools/System:
Online Readings, Documents (Resources)
Students
100%
94.7%
80%
72.0%
65.0%
63.3%
60%
40%
25.0%
20%
20.8%
16.0%
14.9%
10.0%
2.9%
0%
4.0%
2.4%
Have Not Used
UM: Students
Not Valuable
TXSU: Students
UL: Students
Valuable/Very Valuable
UCB: Students
17
25
28
Specific Uses of Sakai Tools/System:
Calendar / Schedule Tool
Instructors
75%
62.9%
58.3%
60%
45%
41.7%
30%
23.2%
13.9%
15%
0%
Have Not Used
Not Valuable
UM: Instructors
Valuable/Very Valuable
TXSU: Instructors
17
25
29
Specific Uses of Sakai Tools/System:
Calendar / Schedule Tool
Students
75%
69.0%
60%
56.0%
52.5%
45%
38.9%
38.5%
34.0%
30%
22.6%
15%
21.8%
24.7%
19.0%
11.0%
5.0%
0%
Have Not Used
UM: Students
Not Valuable
TXSU: Students
UL: Students
Valuable/Very Valuable
UCB: Students
17
25
26
30
Level / Quality of Training & Support
• Most effective way to get help
Colleague/Classmate/Friend
Online help documentation
Keep trying on my own
Email Support
Call Support
Department IT Staff
Others (campus-specific)
• Satisfaction with quality of support
Very satisfied – very dissatisfied
Can be broken up into different categories:
• Timeliness, Quality of response, Friendliness
31
Level / Quality of Training & Support:
Most Common Responses
Instructors
75%
60.0%
60%
56.3%
45.0%
45%
29.5%
30%
25.0%
21.9%
20.0%
15.6%
15%
6.3%
0%
17.0%
6.7%
Ask colleague/friend
UM: Instructors
12.5%
14.0%
21.5%
16.0%
6.7%
Help Docs
TXSU: Instructors
Keep Trying on Own
UL: Instructors
Email/Call Support
UCB: Instructors
32
Level / Quality of Training & Support:
Most Common Responses
Students
75%
60%
45%
42.8%
36.3%
31.0%
30%
34.0% 34.0%
17.0%
15%
6.8%
3.0%
0%
Ask colleague/friend
UM: Students
Help Docs
3.0%
2.9%
Keep Trying on Own
TXSU: Students
3.0%
5.0%
Email/Call Support
UL: Students
30
33
Attitudes / Suggestions for Improvement
• Most important improvement
• Importance of potential / new tools
34
Attitudes / Suggestions for Improvement:
Michigan Results
• Importance of Potential CTools Features
•
•
Helps identify which tools in the Sakai pipeline might be most
beneficial to our users (using Sakai 2.0 when surveyed)
Compared with qualitative data and comments from Support queue
• Most important tools for users:
•
•
Instructors: Gradebook, New Discussion Tool, Group Control
Students: Gradebook, Tests & Quizzes, New Discussion Tool, Group
Control
• Least important tools for users:
•
•
Instructors: Concept Mapping, Podcasting, Blog
Students: ePortfolio, Content Authoring, Blog
35
Attitudes / Suggestions for Improvement:
Limerick Results
•
Faculty:
• Solve problems with test and quizzes, make discussion
•
•
tool more intuitive (qualitative)
Two most important motivators: More information about
it and successful teaching practices (22,3%); training
and support (17,1%); and added teaching and
learning value (10,3%)
Students (Sulis users):
• Faster performance, avoid problems with file upload
capability (22,4%); improve design (21,7%); and
wide adoption across campus (15,5%)
34
36
Attitudes / Suggestions for Improvement:
Berkeley Results
•
Improve the discussion board
•
Add file upload capability
•
Improve the interface
•
Add customization options (image on home
page, customization of synoptic view,
reordering of tools and resources, etc)
34
35
37
Barriers to Use / Pain Points
• Changes that you feel need to be made
• List of potential barriers (skills, time, etc.)
• Would you recommend Sakai to a
colleague?
• What would deter you from using Sakai?
38
• Michigan
Barriers to Use / Pain Points
• Instructors: Gradebook, Service Integration wanted
• Students: Consistency of site usage, UI changes wanted
• Texas State
• Some instructors report that they don’t have time or don’t have skills
• Need ability to set availability dates for documents
• Need ability to set order of documents and to include a visible description of
documents
• Limerick
• Preserve control over course administration while integrating with other IS
•
(qualitative)
Two most important barriers: Need to free up time to get traing and devote to
design (24%) and Dont know or understand teaching implications of using a
CMS (17%)
• Berkeley
•
•
•
•
•
Discussion board!
Bug with roster data (affecting gradebook & site info)
User interface ("reset arrow", font size, etc)
Transition issues from Blackboard to bSpace
39
Other Methods for Evaluation
• 1:1 interactions
• Support queue
• Use logs
• Online help documentation tracking
• Training session feedback
40
Questions?
41
Contact Information
• Stephanie Teasley – CTools, University of Michigan
steasley@umich.edu
• Steve Lonn – CTools, University of Michigan
slonn@umich.edu
• Salwa Khan – TRACS, Texas State University
sk16@txstate.edu
• Jeff Narvid – bSpace, University of California, Berkeley
jeffn@media.berkeley.edu
• Angelica Risquez – Sulis, University of Limerick
angelica.risquez@ul.ie
•
Confluence Page: http://bugs.sakaiproject.org/confluence/pages/viewpage.action?pageId=18063
42
Download