LG675
Session 5: Reliability II
Sophia Skoufaki sskouf@essex.ac.uk
15/2/2012
What is item analysis?
How can we conduct item analysis for a) norm-referenced data-collection instruments?
Only statistical analyses provided through SPSS b) criterion-referenced data-collection measures?
How can we examine the reliability of criterionreferenced data-collection instruments?
Work with some typical scenarios
2
The kind of reliability analysis used to identify items in a data-collection instrument (e.g., questions in a questionnaire, tasks/questions in a language test) which do not measure the same thing as the other items.
It is conducted on data from the pilot study.
The aim is to improve our data-collection instrument by removing any irrelevant items.
3
NB: This item analysis is different from item analysis (also called ‘analysis by items’) which is part of data analysis in experiments. This analysis is done to ensure that the findings of an experiment are generalisable not only to people with similar characteristics to those who participated in the experiment but also to items similar to those in the experiment (Clark 1973).
If you plan to conduct an experiment, see Phil’s discussion of this term and SPSS how-to: http://privatewww.essex.ac.uk/~scholp/statsquib s.htm#item
4
Reminder: Classification of datacollection instruments according to the basis of grading
Item analysis for norm-referenced measures
•
According to the traditional approach to item analysis, items are examined in terms of:
1.
Item facility: It is a measure of how easy an item is. High facility means easy item.
An easy way to assess it is by looking at the percentage of people who answer each item correctly.
The data-collection instrument as a whole should have facility of 0.5 and most items should have around such a level of facility.
6
This is an activity from http://www.caacentre.ac.uk/dldocs/BP2final.pdf
Input the file ‘three_tests_IF.sav’ into SPSS.
This file shows the item facility for each question in three tests.
Examine the item facilities in each test and try to spot problematic item facilities.
Which test seems to be the best in that it contains items which will be able to distinguish among students of various proficiency levels?
7
•
•
2.
Item discrimination: It is a measure of how different performance on an item in comparison to performance on the other items.
It can be assessed via a correlation between the item’s score and the score of the whole measure.
It can also be assessed via Cronbach’s a if item deleted .
8
Do the activity described in the box on pages 26-27 from Phil’s ‘Simple statistical approaches to reliability and item analysis’ handout.
Then do the activity described in the box on pages 29-30.
Calculate also item facility as a percentage of correct answers.
9
Item analysis for criterion-referenced measures (Brown 2003)
Difference Index : Item facility in the posttest – item facility in the pre-test
B-Index : Item facility for students who passed the test – item facility for those who failed it
10
SPSS: Item analysis for criterion-referenced measures
This is an activity from Brown (2003). He used excel to calculate DI and B-I on two data sets.
Download this article as a pdf file from http://jalt.org/test/bro_18.htm
Input the data from page 20 in SPSS.
Calculate DI via Transform…Compute .
11
Reliability of criterion-referenced measures
There are two basic approaches:
1.
Threshold loss agreement
This approach examines the proportion of people who consistently did better than the cutoff point (‘masters’) and the proportion of those who consistently did worse (‘nonmasters’). It uses a test-retest method.
Example statistic: Cohen’s Kappa (AKA ‘kappa coefficient’)
12
The structure of Cohen’s kappa table in this scenario (figure from Brown and
Hudson 2002: 171)
13
2.
Squared error loss agreement
These statistical tests are like the previous ones but they also assess how consistent the degree of mastery/non-mastery is.
Example: phi(lamda) dependability index
(Not available in SPSS, see Brown 2005: 206-
207)
14
SPSS: Assessing reliability of a criterionreferenced measure through Cohen’s Kappa
Go to page 172 at http://books.google.co.uk/books?id=brDfGghl3qI
C&pg=PA169&source=gbs_toc_r&cad=3#v=one page&q&f=false.
Input the data in SPSS.
Conduct the Kappa test.
15
Statistics for validity assessment
ANOVA with one independent variable
16
Brown, J.D. 2003. Criterion-referenced item analysis (The difference index and B-index). Shiken: JALT Testing & Evaluation SIG Newsletter 7 (3) , 18-
24.
Brown, J.D. 2005. Testing in language programs: a comprehensive guide to
English language assessment.
New York: McGraw Hill.
Clark, H.H. 1973. The language-as-fixed-effect fallacy. Journal of Verbal
Learning and Verbal Behavior 12, 335-359.
Scholfield, P. 2011. Simple statistical approaches to reliability and item analysis. LG675 Handout. University of Essex.
17
On the statistics used for item analysis
Brown, J.D. 2003. Criterion-referenced item analysis (The difference index and B-index). Shiken: JALT Testing & Evaluation SIG Newsletter
7 (3) , 18-24.
Scholfield, P. 2011. Simple statistical approaches to reliability and item analysis. LG675 Handout. University of Essex. (pp. 24-33)
On the statistics used to assess the reliability of criterion-referenced measures
Brown, J.D. 2005. Testing in language programs: a comprehensive guide to English language assessment.
New York: McGraw Hill.
(chapter 9)
Brown, J.D. and Hudson, T. 2002. Criterion-referenced language testing.
Cambridge: Cambridge University Press. (chapter 5)
18