Statistical Analysis of the Nonequivalent Groups Design Analysis Requirements N N O O X O O Pre-post Two-group Treatment-control (dummy-code) Analysis of Covariance yi = 0 + 1Xi + 2Zi + ei where: outcome score for the ith unit coefficient for the intercept pretest coefficient mean difference for treatment covariate dummy variable for treatment(0 = control, 1= treatment) ei = residual for the ith unit yi 0 1 2 Xi Zi = = = = = = The Bivariate Distribution 90 80 posttest 70 Program group scores 15-points higher on Posttest. 60 50 40 30 30 40 Program group has 60 70 80 a pretest5-point pretest Advantage. 50 Regression Results yi = 18.7 + .626Xi + 11.3Zi Predictor Coef Constant 18.714 pretest 0.62600 Group 11.2818 StErr 1.969 0.03864 0.5682 t 9.50 16.20 19.85 Result is biased! CI.95(2=10) = 2±2SE(2) = 11.2818±2(.5682) = 11.2818±1.1364 CI = 10.1454 to 12.4182 p 0.000 0.000 0.000 The Bivariate Distribution 90 80 posttest 70 Regression line slopes are biased. Why? 60 50 40 30 30 40 50 pretest 60 70 80 Regression and Error Y No measurement error X Regression and Error Y No measurement error X Y Measurement error on the posttest only X Regression and Error Y No measurement error X Y Measurement error on the posttest only X Y Measurement error on the pretest only X How Regression Fits Lines How Regression Fits Lines Method of least squares How Regression Fits Lines Method of least squares Minimize the sum of the squares of the residuals from the regression line. How Regression Fits Lines Method of least squares Minimize the sum of the squares of the residuals from the regression line. Y Least squares minimizes on y not x. X How Error Affects Slope Y No measurement error, No effect X How Error Affects Slope Y No measurement error, no effect. X Y X Measurement error on the posttest only, adds variability around regression line, but doesn’t affect the slope How Error Affects Slope Y No measurement error, no effect. X X Measurement error on the posttest only, adds variability around regression line, but doesn’t affect the slope. X Measurement error on the pretest only: Affects slope Flattens regression lines Y Y How Error Affects Slope Y X Measurement error on the pretest only: Affects slope Flattens regression lines Y X Y Y X X How Error Affects Slope Y X Y Notice that the true result in all three cases should be a null (no effect) one. X Y Y X X How Error Affects Slope Notice that the true result in all three cases should be a null (no effect) one. Null case Y X How Error Affects Slope But with measurement error on the pretest, we get a pseudo-effect. Y Pseudo-effect X Where Does This Leave Us? Traditional ANCOVA looks like it should work on NEGD, but it’s biased. The bias results from the effect of pretest measurement error under the least squares criterion. Slopes are flattened or “attenuated”. What’s the Answer? If it’s a pretest problem, let’s fix the pretest. If we could remove the error from the pretest, it would fix the problem. Can we adjust pretest scores for error? What do we know about error? What’s the Answer? We know that if we had no error, reliability = 1; all error, reliability=0. Reliability estimates the proportion of true score. Unreliability=1-Reliability. This is the proportion of error! Use this to adjust pretest. What Would a Pretest Adjustment Look Like? Original pretest distribution What Would a Pretest Adjustment Look Like? Original pretest distribution Adjusted dretest distribution How Would It Affect Regression? Y The regression X The pretest distribution How Would It Affect Regression? Y The regression X The pretest distribution How Far Do We Squeeze the Pretest? Y X • Squeeze inward an amount proportionate to the error. • If reliability=.8, we want to squeeze in about 20% (i.e., 1-.8). • Or, we want pretest to retain 80% of it’s original width. Adjusting the Pretest for Unreliability _ _ Xadj = X + r(X - X) Adjusting the Pretest for Unreliability _ _ Xadj = X + r(X - X) where: Adjusting the Pretest for Unreliability _ _ Xadj = X + r(X - X) where: Xadj = adjusted pretest value Adjusting the Pretest for Unreliability _ _ Xadj = X + r(X - X) where: Xadj = _ X = adjusted pretest value original pretest value Adjusting the Pretest for Unreliability _ _ Xadj = X + r(X - X) where: Xadj = _ X = adjusted pretest value r reliability = original pretest value Reliability-Corrected Analysis of Covariance yi = 0 + 1Xadj + 2Zi + ei where: yi 0 1 2 Xadj Zi ei outcome score for the ith unit coefficient for the intercept pretest coefficient mean difference for treatment covariate adjusted for unreliability dummy variable for treatment(0 = control, 1= treatment) = residual for the ith unit = = = = = = Regression Results yi = -3.14 + 1.06Xadj + 9.30Zi Predictor Coef Constant -3.141 adjpre 1.06316 Group 9.3048 StErr 3.300 0.06557 0.6166 t -0.95 16.21 15.09 Result is unbiased! CI.95(2=10) = 2±2SE(2) = 9.3048±2(.6166) = 9.3048±1.2332 CI = 8.0716 to 10.5380 p 0.342 0.000 0.000 Graph of Means 80 75 70 65 60 55 50 45 40 35 30 Comparison Program Pretest Comp Prog ALL pretest MEAN 49.991 54.513 52.252 Posttest posttest MEAN 50.008 64.121 57.064 pretest STD DEV 6.985 7.037 7.360 posttest STD DEV 7.549 7.381 10.272 Adjusted Pretest pretest MEAN adjpre MEAN posttest MEAN pretest STD DEV adjpre STD DEV posttest STD DEV Comp 49.991 Prog 54.513 ALL 52.252 49.991 54.513 52.252 50.008 64.121 57.064 6.985 7.037 7.360 3.904 4.344 4.706 7.549 7.381 10.272 Note that the adjusted means are the same as the unadjusted means. The only thing that changes is the standard deviation (variability). Original Regression Results 90 Pseudo-effect=11.28 80 Original posttest 70 60 50 40 30 30 40 50 pretest 60 70 80 Corrected Regression Results 90 Pseudo-effect=11.28 80 Original posttest 70 60 50 40 Effect=9.31 30 30 40 Corrected 50 pretest 60 70 80