Statistical Analysis of the Nonequivalent Groups Design

advertisement
Statistical Analysis of the
Nonequivalent Groups Design
Analysis Requirements
N
N



O
O
X
O
O
Pre-post
Two-group
Treatment-control (dummy-code)
Analysis of Covariance
yi = 0 + 1Xi + 2Zi + ei
where:
outcome score for the ith unit
coefficient for the intercept
pretest coefficient
mean difference for treatment
covariate
dummy variable for treatment(0 = control, 1=
treatment)
ei = residual for the ith unit
yi
0
1
2
Xi
Zi
=
=
=
=
=
=
The Bivariate Distribution
90
80
posttest
70
Program
group
scores
15-points
higher
on
Posttest.
60
50
40
30
30
40
Program
group
has
60
70
80 a
pretest5-point pretest
Advantage.
50
Regression Results
yi = 18.7 + .626Xi + 11.3Zi
Predictor
Coef
Constant
18.714
pretest
0.62600
Group
11.2818



StErr
1.969
0.03864
0.5682
t
9.50
16.20
19.85
Result is biased!
CI.95(2=10) = 2±2SE(2)
= 11.2818±2(.5682)
= 11.2818±1.1364
CI = 10.1454 to 12.4182
p
0.000
0.000
0.000
The Bivariate Distribution
90
80
posttest
70
Regression
line slopes
are biased.
Why?
60
50
40
30
30
40
50
pretest
60
70
80
Regression and Error
Y
No measurement error
X
Regression and Error
Y
No measurement error
X
Y
Measurement error
on the posttest only
X
Regression and Error
Y
No measurement error
X
Y
Measurement error
on the posttest only
X
Y
Measurement error
on the pretest only
X
How Regression Fits Lines
How Regression Fits Lines
Method of least squares
How Regression Fits Lines
Method of least squares
Minimize the sum of the squares
of the residuals from the
regression line.
How Regression Fits Lines
Method of least squares
Minimize the sum of the squares
of the residuals from the
regression line.
Y
Least squares
minimizes on y not x.
X
How Error Affects Slope
Y
No measurement error,
No effect
X
How Error Affects Slope
Y
No measurement error,
no effect.
X
Y
X
Measurement error
on the posttest only,
adds variability around
regression line, but
doesn’t affect the slope
How Error Affects Slope
Y
No measurement error,
no effect.
X
X
Measurement error
on the posttest only,
adds variability around
regression line, but
doesn’t affect the slope.
X
Measurement error
on the pretest only:
Affects slope
Flattens regression lines
Y
Y
How Error Affects Slope
Y
X
Measurement error
on the pretest only:
Affects slope
Flattens regression lines
Y
X
Y
Y
X
X
How Error Affects Slope
Y
X
Y
Notice that the true result in
all three cases should
be a null (no effect) one.
X
Y
Y
X
X
How Error Affects Slope
Notice that the true result in
all three cases should
be a null (no effect) one.
Null case
Y
X
How Error Affects Slope
But with measurement error
on the pretest, we get a
pseudo-effect.
Y
Pseudo-effect
X
Where Does This Leave Us?



Traditional ANCOVA looks like it should
work on NEGD, but it’s biased.
The bias results from the effect of
pretest measurement error under the
least squares criterion.
Slopes are flattened or “attenuated”.
What’s the Answer?




If it’s a pretest problem, let’s fix the
pretest.
If we could remove the error from the
pretest, it would fix the problem.
Can we adjust pretest scores for error?
What do we know about error?
What’s the Answer?





We know that if we had no error,
reliability = 1; all error, reliability=0.
Reliability estimates the proportion of
true score.
Unreliability=1-Reliability.
This is the proportion of error!
Use this to adjust pretest.
What Would a Pretest Adjustment Look
Like?
Original pretest distribution
What Would a Pretest Adjustment Look
Like?
Original pretest distribution
Adjusted dretest distribution
How Would It Affect Regression?
Y
The regression
X
The pretest
distribution
How Would It Affect Regression?
Y
The regression
X
The pretest
distribution
How Far Do We Squeeze the Pretest?
Y
X
• Squeeze inward an
amount proportionate to
the error.
• If reliability=.8, we want
to squeeze in about 20%
(i.e., 1-.8).
• Or, we want pretest to
retain 80% of it’s original
width.
Adjusting the Pretest for Unreliability
_
_
Xadj = X + r(X - X)
Adjusting the Pretest for Unreliability
_
_
Xadj = X + r(X - X)
where:
Adjusting the Pretest for Unreliability
_
_
Xadj = X + r(X - X)
where:
Xadj =
adjusted pretest value
Adjusting the Pretest for Unreliability
_
_
Xadj = X + r(X - X)
where:
Xadj =
_
X =
adjusted pretest value
original pretest value
Adjusting the Pretest for Unreliability
_
_
Xadj = X + r(X - X)
where:
Xadj =
_
X =
adjusted pretest value
r
reliability
=
original pretest value
Reliability-Corrected
Analysis of Covariance
yi = 0 + 1Xadj + 2Zi + ei
where:
yi
0
1
2
Xadj
Zi
ei
outcome score for the ith unit
coefficient for the intercept
pretest coefficient
mean difference for treatment
covariate adjusted for unreliability
dummy variable for treatment(0 = control,
1= treatment)
= residual for the ith unit
=
=
=
=
=
=
Regression Results
yi = -3.14 + 1.06Xadj + 9.30Zi
Predictor
Coef
Constant
-3.141
adjpre
1.06316
Group
9.3048



StErr
3.300
0.06557
0.6166
t
-0.95
16.21
15.09
Result is unbiased!
CI.95(2=10) = 2±2SE(2)
= 9.3048±2(.6166)
= 9.3048±1.2332
CI = 8.0716 to 10.5380
p
0.342
0.000
0.000
Graph of Means
80
75
70
65
60
55
50
45
40
35
30
Comparison
Program
Pretest
Comp
Prog
ALL
pretest
MEAN
49.991
54.513
52.252
Posttest
posttest
MEAN
50.008
64.121
57.064
pretest
STD DEV
6.985
7.037
7.360
posttest
STD DEV
7.549
7.381
10.272
Adjusted Pretest
pretest
MEAN
adjpre
MEAN
posttest
MEAN
pretest
STD DEV
adjpre
STD DEV
posttest
STD DEV
Comp 49.991
Prog 54.513
ALL
52.252
49.991
54.513
52.252
50.008
64.121
57.064
6.985
7.037
7.360
3.904
4.344
4.706
7.549
7.381
10.272


Note that the adjusted means are the
same as the unadjusted means.
The only thing that changes is the
standard deviation (variability).
Original Regression Results
90
Pseudo-effect=11.28
80
Original
posttest
70
60
50
40
30
30
40
50
pretest
60
70
80
Corrected Regression Results
90
Pseudo-effect=11.28
80
Original
posttest
70
60
50
40
Effect=9.31
30
30
40
Corrected
50
pretest
60
70
80
Download