Learning Analytics: On the Way to Smart Education Dr. Griff Richards Athabasca University Moscow, 08 October 2012 Distance Learning In Canada - Is growing, especially in K-12 public schools 60oN -15% of 2008 high school graduates in BC took at least one on-line course -Increase in “Blended Learning”, online activities for F2F courses. 49oN -Increasing concern for course quality and student retention Athabasca, Alberta Learning Engagement The more a learner interacts with the content and with their peers about the content, the more they are likely to internalize it, and remember it... (Richard Snow, 1980) Learning Engagement Learning engagement promotes student retention and academic success. (George Kuh,2001) • Engagement is “action based on internal motivation” • Can not directly measure engagement • Must look for external traces, actions that indicate interest, and involvement (NSSE) CAUTION: Engagement • Arum & Roska (2011) • Students who studied alone had higher marks than students studying in groups. • Engagement that is not task-focused is unlikely to improve learning. Interaction Equivalency Theory (Garrison & Anderson,1995) Course Content Other learners Instructor The quality of the interaction is more important than the source 4 places to improve learning Course Content 1. Clearer designs 2. Collaborative Learning activities 3. Less lecturing, more mentoring Other learners Instructor 4. More study skills We have insufficient data about which instructional strategies actually work best. Hybrid or Blended Learning? The use of online technologies to augment face to face delivery. -> replaces some F2F “face time” (forums save classroom space) -> uses LMS to track assessments and assignments -> uses technology in class e.g. pop quizzes, collaboration tools Face to Face On-Line Analytic Measures Engagement is inferred from activity. • Student interaction with Online systems leaves an “exhaust trail” of their activity • Learning Analytics is not a statistical sample, it is all the data for all learners • Questions: What patterns to look for? How to interpret them? Example: Snapp • LMS interaction data e.g. Moodle discussion • Extract linked list • Plot interactions in star map SNAPP Visualization SNAPP shot of Conference Students “engaged” SNAPP shot of Conference Individuals with lower engagement, 3 or less messages Does Activity = Engagement ? Beer (2010) plotted total LMS interactions with academic grades. Does activity = engagement? Is Blackboard more engaging than Moodle? Limitations: Activity Outside LMS LMS • As learners become engaged, they move to more engaging channels: email, Elluminate, Skype • This activity is not tracked by LMS. • No data available. Interpretation of Analytics • Data patterns require investigation • Quantitative data requires interpretation --> make and test hypotheses --> create useful models • When we measure something we risk changing it. – e.g. If learners know we count hits they may make meaningless hits to fool the system. Analytics for Learners! • The same analytics should be able to provide easy to understand information dashboards for students. Analytics for Learners SIGNALS Arnold (2010) inserted a traffic light on each student’s course page to provide guidance on course success. A/B C D/F Dashboard for Faculty Arnold (2010) reported 14 % shift from D’s to C’s & B’s. Same number of withdrawals, but W’s occurred earlier before damaging student grade point averages. 12% increase in students requesting assistance. N=220 My webquest data Mesmots a dashboard for learners (Richards & Sehboub, 2008) Data for my class How to start: Analytics for online & blended learning? • Measuring something is the first step • “Better to measure something than to measure nothing” (Scrivens) • Need more data than just page hits. We also need to ask learners about their experience, what worked, what needs improvement. Analytics at the activity level Dynamic Evaluation Model (Richards & Devries,2011) » Preparation Conduct • Design • Facilitation • Learning Reflection Timely feedback enables quick fixes Dynamic Evaluation Model » Preparation Conduct • Design • Facilitation • Learning Reflection If Analytics, Then What? • If analytics show students are failing, is there a moral obligation to help them? • If analytics show a course has weaknesses, is there a business obligation to fix it? • If analytics reveal weak instruction, who is responsible to intervene? • If analytics are inaccurate, who fixes them? • What about privacy & ethics? Who owns the data? Who has the right to see it? • joannaparypinski.com The Analytics Box? Learning Analytics: On the Way to Smart(er) Education Dr. Griff Richards Athabasca University griff@sfu.ca Moscow, 08 October 2012