Statistics

advertisement
Hui Bian
Office for Faculty Excellence
Fall 2011
• Purpose of data screening
•To find wrong entries
•To find extreme responses or outliers
•To see if data meet the statistical
assumptions of analysis you are going to
use.
2
• Distribution diagnosis
•
•
•
•
Frequency tables
Histograms and bar graphs
Stem-and-leaf plots
Box plots
3
• Frequencies procedure provides statistics and
graphical display to describe your variables.
• Frequency tables can give you: frequency
counts, percentages, cumulative percentages,
mean, median, mode, sum, standard deviation,
variance. Skewness and kurtosis, etc.
• Assumptions
• Tabulations and percentages are useful for
categorical data.
• Mean and standard deviation are based on normal
theory and are appropriate for quantitative
variables.
• Robust statistics, such as median, percentiles, and
quartiles are appropriate for quantitative variables
that may or may not meet the assumption of
normality.
• Example: what are the distributions of
demographic variables? Such as age,
gender, and grade (Q1, Q2, Q3).
• First recode Q1 into Q1r because we want
real ages for the participants.
• How to recode Q1?
• Recode Q1 into Q1r
• Obtain frequency table: Analyze > Descriptive Statistics >
Frequencies
9
• Click Statistics to get this window
• Click Charts to get this window
11
• Frequency statistics
• Frequency statistics
• Skewness and kurtosis are statistics that characterize
the shape and symmetry of the distribution
• Skewness: a measure of the asymmetry of a
distribution. The normal distribution is symmetric and
has a skewness value of zero.
• Positive skewness: a long right tail.
• Negative skewness: a long left tail.
13
• Skewness
• Kurtosis: a measure of the extent to which
observations cluster around a central point. For a
normal distribution, the value of the kurtosis statistic is
zero.
• Leptokurtic data values are more peaked (positive
kurtosis) than normal distribution.
• Platykurtic data values are flatter and more
dispersed along the X axis (negative kurtosis) than
normal distribution.
15
16
• Unacceptable level of skewness and kurtosis:
departure from normality
• Some use: ± 0.5
• Some use: ± 1.00
• As a guideline, a skewness value more than twice its
standard error is taken to indicate a departure from
symmetry (from SPSS help menu).
17
• Frequency tables of Q2 and Q3
• From frequency table of Q3, we decide that we do
not want Ungraded or Other grade group in our data
analysis.
• Then, we need to recode Q3.
• How to recode Q3 into Q3r
• Keep four levels (9th - 12th ).
• Recode the fifth level (Ungraded or Other grade) into system
missing.
• Run frequency for Q3r.
• Recode Q3 into Q3r
• Histogram
1. It is plotted along an
equal interval scale. The
height of each bar is the
count of values of a
variable.
2. It shows the shape,
center, and spread of
the distribution.
3. A normal curve helps
you judge whether the
data are normally
distributed.
• Example: Frequency of Q41(frequency of drinking)
• The Explore procedure produces summary statistics
and graphical displays, either for all of your cases
or separately for groups of cases. Reasons of using
Explore procedure:
•
•
•
•
•
Data screening
Outlier identification
Description
Assumption checking
Characterizing differences among subpopulations
(groups of cases).
23
• Exploring the data can help
•
•
•
to determine whether the statistical techniques that
you are considering for data analysis are
appropriate.
the exploration may indicate that you need to
transform the data if the technique requires a normal
distribution.
or you may decide that you need nonparametric
tests.
24
• Statistics: Mean, median, standard error, variance, standard
deviation , range, skewness, kurtosis, and confidence interval for
the means, the Kolmogorov-Smirnov statistic with a Lilliefors
significance level for testing normality, and the Shapiro-Wilk
statistic. Boxplots, stem-and-leaf plots, histograms, normality
plots, and spread-versus-level plots with Levene tests and
transformations.
• Data: quantitative variables or categorical variables with a
reasonable number of distinct categories.
• Assumptions: the distribution of your data does not have to be
symmetric or normal.
• Example: let’s explore Q41 again, by Q2 (gender).
• Obtain Explore: Analyze > Descriptive Statistics >
Explore
• Click Statistics to get this window
• Click Plots to get this window
Descriptives
Q2 What is your sex
Q41 How many days drink
Female
alcohol 30 days
Statistic
Mean
95% Confidence Interval for
Lower Bound
1.73
Mean
Upper Bound
1.79
5% Trimmed Mean
.013
1.63
Median
1.00
Variance
1.230
Std. Deviation
1.109
Minimum
1
Maximum
7
Range
6
Interquartile Range
Male
Std. Error
1.76
1
Skewness
1.647
.029
Kurtosis
2.501
.058
1.85
.015
Mean
95% Confidence Interval for
Lower Bound
1.83
Mean
Upper Bound
1.88
5% Trimmed Mean
Median
1.69
1.00
Variance
1.732
Std. Deviation
1.316
Minimum
1
Maximum
7
Range
6
Interquartile Range
1
Skewness
1.726
.028
Kurtosis
2.624
.055
• Test of normality
The normality
assumption is not
satisfied.
• Stem-and-Leaf Plots: use the original data values to
display the distribution's shape.
Female
Male
• Stem-and-Leaf Plots
• Normal Q-Q Plots:
• The straight line in the plot represents expected values
when the data are normally distributed.
• A normal distribution is indicated if the data points
fall on or very near the diagonal line.
32
• Normal Q-Q Plot of Q41among Females
• Normal Q-Q Plot of Q41among Males
• Why use box plots?
• Provide some indication of the data's symmetry and
skewness.
• We can detect outliers.
• By using a boxplot for each categorical variable
side-by-side on the same graph, one quickly can
compare data sets.
• Box plots
• A boxplot splits the data set into quartiles.
• Calculate the median and the quartiles (the lower
quartile is the 25th percentile and the upper quartile
is the 75th percentile).
• For 25th percentile, 25% of the data values are
below it.
• The box represents the middle 50% of the data--the
"body" of the data.
36
No outliers
Response
variable
Median
Highest
value
25th
percentile
Smallest
value
37
• Box plots of Q41 by gender
Median
• Box Position:
• The location of the box within two outer lines can
provide insight on the normality of the sample's
distribution.
• When the box is not centered between the outer
lines, the sample may be positively or negatively
skewed.
• If the box is shifted significantly to the low end, it is
positively skewed; if the box is shifted significantly
to the high end, it is negatively skewed.
• Box Size
• The size of the box can provide an estimate of the
kurtosis.
• A very thin box relative to the outer lines indicates a
distribution with a thinner peak.
• A wider box relative to the outer lines indicates a
wider peak. The wider the box, the more U-shaped
the distribution becomes.
• If there is a violation of statistical assumptions, you
might need a data transformation to correct this
matter.
• Concern: interpretation problem
• Log transformation, square root, logarithm, inverse,
reflect and square root, reflect and logarithm, and
reflect and inverse
41
• The Crosstabs procedure forms two-way and multiway
tables and provides a variety of tests and measures
of association for two-way tables.
• Crosstabs’ statistics and measures of association are
computed for two-way tables only.
• Example 1: the association (independence) of gender
(Q2) and physical activity (Q80r).
• Q2 has two levels: female and male
• Q80r has four levels: 0 days, 1-2 days, 3-5 days,
and more than 5 days
• Obtain Crosstabs: Analyze > Descriptive Statistics >
Crosstabs
• Click Statistics to get this window
• Click Cell to get this window
• Crosstabs statistics (SPSS): Chi-square
• For tables with two rows and two columns, select Chisquare to calculate the Pearson chi-square, the
likelihood-ratio chi-square, Fisher’s exact test, and
Yates’ corrected chi-square.
• For 2 × 2 tables, Fisher’s exact test is computed
when columns in a larger table has a cell with an
expected frequency of less than 5 (more than 20%).
• Yates’ corrected chi-square is computed for all other
2 × 2 tables.
47
• Crosstabs: Chi-square
• For tables with any number of rows and columns,
select Chi-square to calculate the Pearson chi-square
and the likelihood-ratio chi-square.
• When both table variables are quantitative, Chisquare yields the linear-by-linear association test.
48
• Q80r by Q2
• Chi-square results
• Nominal. For nominal data (no intrinsic order, such as
Catholic, Protestant, and Jewish), you can select Contingency
coefficient, Phi (coefficient) and Cramer’s V, Lambda
(symmetric and asymmetric lambdas and Goodman and
Kruskal’s tau), and Uncertainty coefficient.
51
• Ordinal. For tables in which both rows and columns contain
ordered values, select Gamma (zero-order for 2-way tables
and conditional for 3-way to 10-way tables), Kendall’s tau-b,
and Kendall’s tau-c.
• For predicting column categories from row categories, select
Somers’ d.
52
• Nominal by Interval. When one variable is
categorical and the other is quantitative, select Eta.
The categorical variable must be coded numerically.
53
• Kappa. Cohen's kappa measures the agreement
between the evaluations of two raters when both are
rating the same object (interrater reliability).
 A value of 1 indicates perfect agreement.
 A value of 0 indicates that agreement is no better
than chance.
 Kappa is available only for tables in which both
variables use the same category values and both
variables have the same number of categories.
• Example 2. Whether Q24 (ever consider suicide) and
Q25 (ever make suicide plan) agree each other.
• Sometimes, one measure is from an objective test and
another is self-report measure.
• We need to get Kappa statistic.
55
• Example 2: SPSS output
56
• Kappa statistics
Kappa = 0.589 with p < 0.001. This measure of agreement, while
statistically significant, is only marginally convincing. As a rule of
thumb values of Kappa from 0.40 to 0.59 are considered moderate,
0.60 to 0.79 substantial, and 0.80 outstanding. Most statisticians
prefer for Kappa values to be at least 0.6 and most often higher than
0.7 before claiming a good level of agreement.
• Risk. For 2 x 2 tables, a measure of the strength of
the association between the presence of a factor and
the occurrence of an event.
• relative risk (RR) is the risk of an event (or of
developing a disease) relative to exposure. Relative
risk is a ratio of the probability of the event occurring
in the exposed group versus a non-exposed group.
58
• Example 3: examine the effect of thinking
about suicide on suicide behavior. We have two
groups: Q24, thinking about suicide (yes) and
not thinking about suicide (no). Then, they plan
suicide, Q25 (yes/no).
59
• Risk
• Relative Risk: is the risk of making suicide plan relative
to thinking about suicide. Relative risk is a ratio of the
probability of making plan (yes) in the thinking about
suicide group versus a non-thinking about suicide
group.
Relative
risk
• RR = (1274/2231)/(483/13926)= 16.46
• It means the risk of making plan for people thinking
about suicide is more than 16 times that of people
non-thinking about suicide.
• Odds ratio is a way of comparing whether the
probability of making suicide plan is the same for
thinking and non-thinking groups.
• Odds ratio = (1274/957)/(483/13443)=37.05
• This odds ratio illustrates that making plan is far more
likely in the thinking group.
62
• Bivariate correlation is about the association of two
variables.
 Research question is like: whether two variables
are related to each other.
 Bivariate correlation statistic: Pearson productmoment coefficient (Pearson correlation: r).
63
• Before calculating a correlation coefficient, screen
your data for outliers and evidence of a linear
relationship.
• Pearson’s correlation coefficient is a measure of linear
association.
• Two variables can be perfectly related, but if the
relationship is not linear, Pearson’s correlation
coefficient is not an appropriate statistic for measuring
their association.
64
• Data. Use symmetric quantitative variables for
Pearson’s correlation coefficient and quantitative
variables or variables with ordered categories for
Spearman’s rho and Kendall’s tau-b.
• Assumptions. Pearson’s correlation coefficient
assumes that each pair of variables is bivariate
normal.
65
• Correlation Coefficients. For quantitative, normally
distributed variables, choose the Pearson correlation
coefficient.
• If your data are not normally distributed or have
ordered categories, choose Kendall’s tau-b or
Spearman, which measure the association between
rank orders.
66
• Example: Correlation between Q3r and Q80
• Go to Analyze > Correlate > Bivariate
67
• SPSS output
68
• Correlation coefficients range in value from –1 (a
perfect negative relationship) and +1 (a perfect
positive relationship).
• A value of 0 indicates no linear relationship. When
interpreting your results, be careful not to draw any
cause-and-effect conclusions due to a significant
correlation.
69
Pearson correlation: -.064
Kendall’s tau-b : -.051
Question: the coefficients are small but significant, why?
• For large samples, it is easy to achieve significance,
and one must pay attention to the strength of the
correlation to determine if the relationship explains
very much.
70
• Partial correlation coefficients describe the linear
relationship between two variables while controlling
for the effects of one or more additional variables.
• Correlations are measures of linear association
71
• Example: Is there a relationship between Q39 and
Q80 while controlling Q81?
• Analyze > Correlate > Partial
72
• SPSS output
73
 T test
 One-Sample T test
 Independent-Samples T test
 Paired-Samples T test
 One-way ANOVA
74
• One-Samples T Test: this procedure tests whether the
mean of a single variable differs from a specified
constant.
• Example: we want to know whether the mean of
Q80 is different from 5 days (we use 5-day as cutoff)
• Recode Q80 into 0 = 0 days, 1 = 1 day, 2 = 2
days … 7= 7 days.
• Go to Analyze > Compare Means > One-Samples T
Test
75
 SPSS output
76
 Independent-Samples T test
 The Independent-Samples T Test procedure
compares means for two groups of cases.
 Example: whether there is a difference in Q80
between males and females.
 Go to Analyze > Compare Means > IndependentSamples T Test
77
 Independent-Samples T test
78
 Independent-Samples T test: SPSS output
79
• Paired-Samples T Test: This procedure compares the means of
two variables for a single group.
• The procedure computes the differences between values of the
two variables for each case and tests whether the average
differs from 0.
• Each subject has two measures, often called before and after
measures. An alternative design for which this test is used is a
matched-pairs or case-control study, in which each record in the
data file contains the response for one person and also for his
or her matched control subject.
80
• Paired-Sample T Test
• Use combined data set from the example of
Merging data sets.
• We want to know if there is a difference in a05
(school performance) across the time after
intervention.
• We focus on one group. So we need to select cases
from group 1.
• Go to Analyze > Compare Means > Paired-Sample
T Test
81
• Paired-Sample T Test:
82
• Paired-Sample T Test: SPSS output
83
Download