SEJ DPS Seminar Nov 2015

advertisement
Assessment for learning:
Learning from assessment?
Sally Jordan (@SallyJordan9)
DPS Seminar, 19th November 2015
My background
● Longstanding interests in assessment and maths skills
development;
● Introduced online interactive assessment into S151 Maths
for Science (2002);
● Several CETL and other externally-funded projects in
particular:
● Project which investigated the use of short-answer freetext questions;
● “Remote observation” of student engagement with eassessment;
● I am passionate about using scientific methodology to find
out what is going on in learning.
Why have I used computer-marked
assessment?
• In my work, the focus has been on ‘assessment for
learning’, so feedback and giving students a second and
third attempt is important (Gibbs & Simpson, 2004-5).
• We aim to ‘provide a tutor at the student’s elbow’ (Ross
et al., 2006).
• My work has been at the limit of what is possible with
computer-marked assessment (not just multiple-choice
questions!).
• …to learn more about what is going on…[Learning
analytics/assessment analytics]
Analysis of student errors
● At the most basic – look for questions that
students struggle with;
● Look at responses in more detail to learn more
about the errors that students make;
● This can give insight into student
misunderstandings.
● So what topics in Maths for Science do students
find difficult?
So what topics in Maths for Science do
students find difficult?
Analysis of student responses to individual
questions
Gives information about student errors, linked to
their misconceptions. The confidence in the findings
is increased when
• The questions require a ‘free-text’ (constructed)
response;
• The questions are in summative use (students
are trying);
• Similar errors are seen in different variants.
See Jordan (2014)
Why is the answer 243? (instead of 9)
The question was:
Evaluate 36/3
7
Why is the answer 243? (instead of 9)
The question was:
Evaluate 36/3
Students were evaluating
Instead of 36/3 = 32 = 9
6
3
5
 3  243
3
For another variant the answer was 5000
instead of 100
The question was:
Evaluate 104/2
Students were evaluating
4
10
10000

 5000
2
2
Instead of 104/2 = 102 = 100
9
Measuring student engagement… “750
students used my iCMA”
Measuring student engagement…
Measuring student engagement…
When do students do iCMAs? (overall
activity)
When do students do iCMAs?
(impact of deadlines)
When do students do iCMAs (typical
patterns of use)
Student engagement with feedback
Student engagement with feedback
(identical question)
Module A
Module B
General conclusions
●Analysis of student responses to interactive
computer-marked questions can give information
about student misunderstandings and student
engagement with assessment;
●Generally, students do what they believe their
teachers expect them to do;
●Engagement with computer-marked assessment
can act as a proxy for more general engagement
with a module (and so act as an early warning if
engagement is not as deep as we might wish).
The future?
● Redecker, Punie and Ferrari (2012, p. 302)
suggest that we should “transcend the testing
paradigm”; data collected from student interaction
in an online environment offers the possibility to
assess students on their actual interactions rather
than adding assessment separately.
A short-answer question (PMatch)
https://students.open.ac.uk/openmark/s104-11b.icma48/
A short-answer question (PMatch)
A short-answer question (PMatch)
Short-answer free-text questions: humancomputer marking comparison
●A linguistically-based system was used to mark
and give feedback on student responses of ‘a
sentence’ in length.
●The computer marking was compared with that of
6 human markers.
Question
Number of
responses in
analysis
Percentage of responses where
the human markers were in
agreement with question author
Range for the 6
human markers
Mean
percentage for
the 6 human
markers
Percentage of
responses where
computer marking
was in agreement
with question author
A
189
97.4 to100
98.9
99.5
B
248
83.9 to 97.2
91.9
97.6
C
150
80.7 to 94.0
86.9
94.7
D
129
91.5 to 98.4
96.7
97.6
E
92
92.4 to 97.8
95.1
98.9
F
129
86.0 to 97.7
90.8
97.7
G
132
66.7 to 90.2
83.2
89.4
Short-answer free-text questions: computercomputer marking comparison
●An undergraduate student (not of computer
science) developed answer matching using two
algorithmically based systems, Java regular
expressions and OpenMark PMatch;
●These are not simple ‘bag of words’ systems;
●Student responses were used in the development
of the answer matching, as had been the case for
the linguistically based IAT system;
●The results were compared.
Question
Responses in
set
Percentage of responses where computer marking
was in agreement with question author
Computational
linguistics
Algorithmic manipulation of
keywords
IAT
OpenMark
Regular
Expressions
A
189
99.5
99.5
98.9
B
248
97.6
98.8
98.0
C
150
94.7
94.7
90.7
D
129
97.6
96.1
97.7
E
92
98.9
96.7
96.7
F
129
97.7
88.4
89.2
G
132
89.4
87.9
88.6
Word-length for Snowflake question
●With no restriction on the number of words
allowed
1
Word-length for Snowflake question
●With warning ‘Your answer should be no more
than 20 words in length.’
1
Word-length for Snowflake question
●With warning ‘Your answer should be no more
than 20 words in length.’
1
Exploring the gender gap
● S207 (The Physical World) was a 60-credit OU level 2 (FHEQ
level 5) module (replaced by S217 from autumn 2015).
● Men were significantly and consistently more likely to complete
S207 than women.
● Of those who complete, men were more likely to pass.
● The difference in outcomes was larger in 2013-14.
● The effect is not present in our level 1 modules or in any other
Science level 2 modules except for S282 (Astronomy), where
women also do less well but the difference is not so stark.
● Women do slightly better on our level 3 physical science
modules.
30
So what’s causing the attainment
gap?
Initial hypotheses:
● It has something to do with the type of assessment we are
using.
● There are other demographic differences (e.g. previous
educational qualifications) between our male and female
students.
● There are other differences between men and women e.g. in
the amount of time they have available for study.
● It has something to do with role models and/or the nature of
the support being offered.
● It has something more fundamental to do with the way we are
teaching physics.
● On average, women and men handle certain physical
concepts and skills of physics in a different way.
31
Results of data analysis
Concentrating on the 2013-14 presentation
N (male) = 455
N (female) = 157
32
Different outcomes for male and
female students
33
When do students stop submitting
assignments?
34
Performance on different parts of
the exam
Part A is multiple-choice questions; Part B is short-answer
questions; Part C is longer questions (with choice)
35
Performance on different questions
in the exam
For Part C (Choice of 3 out of 7 long questions)
36
Choice of different questions in the
exam
For Part C (Choice of 3 out of 7 long questions)
37
Performance on iCMAs (interactive
computer-marked assignments)
Note that a wide range of questions are represented, but
there is no obvious correlation between male/female
performance and question type.
38
Demographics
 There are no obvious differences between the distribution of
other demographic factors (e.g. age, previous educational
qualifications etc.) for men and women.
 However, there is some indication that women with other
characteristics, in particular
o having less than two A levels
o not having English as a first language
may be particularly likely to withdraw.
 There is also some evidence that women appreciate different
aspects of student support/tuition than men.
 Although the proportion of students with A levels is similar for
men and women, we don’t know how many have A levels in
maths and physics.
39
Summary and some questions
● We have a significant attainment gap between men and women on our
level 2 physics module; we have ruled out many possible explanations
but further investigation is of vital importance.
● Block 2 (Describing motion) and Block 3 (Predicting motion) may be
creating particular problems.
● Particular questions (not question types) may be causing particular
problems.
● We know that most girls who do maths and physics A level go straight
to University; does this mean that the female OU population has a
relatively smaller proportion of students with A levels in maths and/or
physics?
● Our level 1 modules prepare students for the content of S207/S217, but
do they prepare students for the question types?
● Are our teaching materials and questions sufficiently clear for students
for whom English is not their first language?
● Women appear to be more likely to give up if they are finding the
module difficult. Is this because they have a different motivation for
student/less time/less confidence?
40
What next? Your thoughts and
suggestions please
● S217 has a different tuition and assessment strategy – review the
impact
● Surveying (male and female) students to find out more detail
about their previous qualifications, whether they consider English
to be their first language, and their perception of their
preparedness to study S207/S217.
● Work with a statistician to model which demographic factors may
be combining to contribute to success or the lack of it – for all
students and across several universities
● Combine the work with another project that is using free-text
questions to establish the “Force concept inventory”
41
References
Clow, D. (2013). An overview of learning analytics. Teaching in
Higher Education, 18(6), 683-695.
Ellis, C. (2013). Broadening the scope and increasing the
usefulness of learning analytics: The case for assessment
analytics. British Journal of Educational Technology, 44(4), 662664.
Nicol, D. & Macfarlane‐Dick, D. (2006). Formative assessment and
self‐regulated learning: a model and seven principles of good
feedback practice. Studies in Higher Education, 31(2), 199-218.
Redecker, C., Punie, Y., & Ferrari, A. (2012). eAssessment for 21st
Century Learning and Skills. In A. Ravenscroft, S. Lindstaedt, C.D.
Kloos & D. Hernandez-Leo (Eds.), 21st Century Learning for 21st
Century Skills (pp. 292-305). Berlin: Springer.
For more about what I’ve discussed
Jordan, S. (2011). Using interactive computer-based
assessment to support beginning distance learners of science,
Open Learning, 26(2), 147-164.
Jordan, S. (2012). Student engagement with assessment and
feedback: Some lessons from short-answer free-text eassessment questions. Computers & Education, 58(2), 818-834.
Jordan, S. (2013). Using e-assessment to learn about learning.
In Proceedings of the 2013 International Computer Assisted
Assessment (CAA) Conference, Southampton, 9th-10th July
2013. Retrieved from http://caaconference.co.uk/proceedings/
Jordan, S. (2014). Adult science learners’ mathematical
mistakes: an analysis of student responses to computer-marked
questions. European Journal of Science and Mathematics
Education, 2(2), 63-87.
For more on OU e-assessment systems
Butcher, P. G. (2008). Online assessment at the Open University using open
source software: Moodle, OpenMark and more. In Proceedings of the 12th
International Computer Assisted (CAA) Conference, Loughborough, 8th-9th July
2008. Retrieved from
http://caaconference.co.uk/pastConferences/2008/proceedings
Hunt, T. J. (2012). Computer-marked assessment in Moodle: Past, present and
future. In Proceedings of the 2012 International Computer Assisted Assessment
(CAA) Conference, Southampton, 10th-11th July 2012. Retrieved from
http://caaconference.co.uk/proceedings/
Ross, S. M., Jordan, S. E. & Butcher, P. G. (2006). Online instantaneous and
targeted feedback for remote learners. In C. Bryan & K. Clegg (Eds.), Innovative
Assessment in Higher Education (pp. 123-131). London: Routledge.
Sangwin, C. J. (2013). Computer aided assessment of mathematics. Oxford:
Oxford University Press.
Much of what I have said is discussed in more detail in:
Jordan, S. E. (2014). E-assessment for learning? Exploring the potential of
computer-marked assessment and computer-generated feedback, from shortanswer questions to assessment analytics. PhD thesis. The Open University.
Retrieved from http://oro.open.ac.uk/4111
Sally Jordan
email: sally.jordan@open.ac.uk
twitter: @SallyJordan9
blog: http://www.open.ac.uk/blogs/SallyJordan/
Download