Repeated Measures Lecture - 1 2/6/2016 Repeated Measures: Field Ch 14, 15 The analyses considered here are those involving comparison of means in two or more research conditions when either the same people or matched people have participated in the conditions. Two major categories of design 1. Participants as their own controls design. (Also called Within Subjects designs.) Same people participant in all conditions of the research. Simplest: 2 conditions; Comparing performance of a group of people before treatment and after treatment. Called a Before-After or Pre-Post design. Longitudinal studies comparing performance of people over several time periods. So, many repeated measures designs involve comparisons over time. The key advantage of repeated measures designs is the similarity of participants in the two or more conditions. For this reason, the maximum similarity is achieved by using the same people in all conditions. But there are many designs in which you can’t use the same people. Learning designs are a good example. Once you’ve learned something, you can’t unlearn it. The next design is a way around that. 2. Matched participants designs. (Matched people) Different people experience the different conditions, but they’re matched with respect to one or more variables that are correlated with the dependent variable, making them “as if” they were the same people in each condition. Of course, you can’t match on everything, so matched participants are not as nearly identical as participants matched with themselves. 3. Cloned participants designs!!!!!??? (Totally matched people) Having your cake and eating it to. This is a way of having nearly perfectly matched participants without having to use the same people in all conditions. The ultimate matching. Repeated Measures Lecture - 2 2/6/2016 Advantages of Repeated Measures designs 1. Absence of confounds from extraneous variables. Since the people in each condition are the same or are matched, differences found between conditions are more likely to be due to treatment differences rather than to participant differences on variables we’re not interested in. 2. Increase in Power. Repeated measures designs offer increased power for the same or smaller samples when compared with between subjects designs. Consider the relationship between the independent groups t-test and the correlated groups t-test. The correlated-groups t-test is often expressed as X-bar1 – X-bar2 ---------------------------------------------S12 + S22 – 2rS1S2 r is the correlation between matched participants. -----------------------N If S1=S2=S, this can be rewritten as X-bar1 – X-bar2 t = -----------------------2S2 (1-r) -------------N The key to this is the (1-r) part. r is the correlation between the paired scores. In an independent groups design, r = 0. In a repeated measures design, r is usually > 0. This means that for a given difference in means (X-bar1 – X-bar2), the larger the value of r, the smaller the denominator of the t statistic and therefore the bigger the value of t. The bigger the value of t, the more likely it is to be a rejection t. That is, power to detect a difference is a positive function of r - the larger the correlation between paired scores, the greater the likelihood of detecting a difference, if there is a difference. So designs with a positive correlation between scores in the two conditions will be more powerful than designs with zero correlation, e.g., independent groups designs. Of course, if there is no difference between the means of the two populations, then power is not an issue. But if there IS a difference between the population mean, the correlated groups design will be more likely to detect it. Repeated Measures Lecture - 3 2/6/2016 How should data be put in the data editor for repeated measures designs? Between Subjects Designs: Different conditions are represented in different rows of the data matrix. C o n d i t i o n 1 C o n d i t i o n 2 Repeated Measures Designs: Different conditions are represented in different columns of the data matrix with each row representing a person or matched persons. C o n d i t i o n C o n d i t i o n 1 2 Repeated Measures Lecture - 4 2/6/2016 Data Matrix of Designs with 2 Repeated Measures Factors This example has two repeated measures factors, one with 2 levels and one with 3. Example: Two types of material are being taught – material involving multiplication and material involving division. Three data collection time periods are used at equally spaced intervals. Multiplication Multiplication Time 1 Time 2 Levels of outer factor Division Time 3 Time 1 Time 2 Time 3 Levels of inner factor Combination Designs: Between Subjects and Repeated Measures factors . . . 1B 1W: In this example, the Between-subjects factor has two levels and the Repeated Measures factor has 3. For example, two ways of teaching statistics (with lab vs. without) measured across 3 tests. Lab Lab Lab T1 T2 T3 No Lab No Lab No Lab T1 T2 T3 Repeated Measures Lecture - 5 2/6/2016 The simplest type of repeated measures analysis – the paired samples t-test. – Start here on 9/23/15. Comparing conscientiousness under honest instructions with conscientiousness under faking instructions. The data editor 4.407 – 3.697 d = ------------------------ = 1.04 .685 This is a HUGE effect size. Repeated Measures Lecture - 6 2/6/2016 The paired sample t as an analysis of difference scores. We’ll get the exact same result by analyzing the single column of f-h difference scores. Null hypothesis is that the mean of the difference scores is 0. GET FILE='G:\MdbR\0DataFiles\Wrensen_070114.sav'. DATASET NAME DataSet1 WINDOW=FRONT. T-TEST /TESTVAL=0 /MISSING=ANALYSIS /VARIABLES=dcons /CRITERIA=CI(.95). T-Test [DataSet1] G:\MdbR\0DataFiles\Wrensen_070114.sav One-Sample Statistics N dcons Mean 166 Std. Deviation .710 Std. Error Mean .8431 .0654 One-Sample Test Test Value = 0 95% Confidence Interval of the Difference t dcons 10.854 df Sig. (2-tailed) 165 .000 Mean Difference .7102 Lower Upper .581 .839 From above, for reference . . . Many statistical procedures involving repeated measures make use of difference scores. Repeated Measures Lecture - 7 2/6/2016 One-way Repeated Measures ANOVA Example Data from Altman, D. G. Practical Statistics for Medical Research. Data are from Table 12.2, p. 327. "Table 12.2 shows the heart rate of nine patients with congestive heart failure before and shortly after administration of enalaprilat, an angiotensin-converting enzyme inhibitor. Measurements were taken before and at 30, 60, and 120 minutes after drug administration. Presumably, the drug should lower heart rate. So this is a basic longitudinal study, although the intervals between measurements are not equal. SUBJECT TIME0 1 2 3 4 5 6 7 8 9 96 110 89 95 128 100 72 79 100 TIME30 92 106 86 78 124 98 68 75 106 TIME60 86 108 85 78 118 100 67 74 104 TIME120 92 114 83 83 118 94 71 74 102 Analyze -> General Linear Model -> Repeated Measures Mike – demo this live. Name the repeated-measures factor, and enter the number of levels. Then click on [Add]. Move the names of the columns representing the levels of the repeated measures factor into the appropriate place under [Within-Subjects Variables]. (Note the change in terminology from Repeated Measures to Within-Subjects.) Repeated Measures Lecture - 8 2/6/2016 Since the 1st measurement appears to be special, I specified a dummy variable repeated measures contrast in which the all levels were compared with level 1 of the RM factor. Dummy variable coding is called Simple coding in SPSS. Since the times are not equally spaced – 30,60,120 – I can’t easily use orthogonal polynomial contrasts to look at the shape of change over time. Check the usual set of optional statistics. Output for the Parameter Estimates has been erased. Repeated Measures Lecture - 9 2/6/2016 The syntax for the analysis. GLM time0 time30 time60 time120 /WSFACTOR = time 4 Simple(1) <<<<---- Specifies the RM factor and the contrast. /METHOD = SSTYPE(3) /PLOT = PROFILE( time ) /PRINT = DESCRIPTIVE ETASQ OPOWER PARAMETER /CRITERIA = ALPHA(.05) /WSDESIGN = time . General Linear Model Hypothesis Tested in the Analysis The null hypothesis that is tested in the analysis can be viewed in two different and essentially equivalent ways. Version 1 of the Null hypothesis: Equality of Means The first is that the four means are all equal: H0: µTime0 = µTime30 = µTime60 = µTime120 Version 2 of the Null hypothesis: Means of Differences Equal 0 If the means are all equal then the differences between pairs of means are all 0. Compute µDiff1 = µTime0 - µTime30 µDiff2 = µTime30 - µTime60 µDiff3 = µTime60 - µTime120 (Any other 3 nonredundant differences would work.) Then the 2nd version of the null hypothesis is: H0: µDiff1 = µDiff2 = µDiff3 = 0 Note that the = 0 at the end of the 2nd null is very important. The null is that every difference is 0. Repeated Measures Lecture - 10 2/6/2016 The Sequence of Tests performed by GLM The Multivariate test that all mean differences are 0. The first tests performed by GLM are what are called Multivariate Tests. They're called that because they are a test of the hypothesis that the multiple difference variables (3 in our case) are all 0. So they’re multivariate tests. Four different multivariate tests are performed. Each is based on slightly different assumptions, and in some instances, the results for the four may be different. In this case, they are all equivalent. The multivariate tests are the most robust tests of the null hypothesis. This means that they are less affected by nonnormality of the distributions than are the tests that follow. The price paid for that robustness is loss of power. The multivariate tests are less powerful than those that follow. This means that if you're worried about power and are hoping to reject the null, rejecting the null with the multivariate tests means you're home free. But if you fail to reject the null with the multivariate tests, then you have to hope that you meet the restrictions for the more powerful tests that follow. In this case, we fail to reject the null using the multivariate tests. The test of sphericity. When conducting a one-way repeated measures ANOVA, the values of variances and covariances should fall within specific ranges. If this is so, the matrix is said to meet the sphericity criterion. If the sphericity criterion is met, then the most powerful test of the null can be employed. If it not met, then either the multivariate test must be used, or one of the special tests devoted to getting around the failure to find sphericity must be employed. Fortunately, SPSS gives you all the tests, so you can pick the one which is appropriate. The null hypothesis in Mauchly's test is that the sphericity condition holds, so we generally hope to not reject the null. In this instance, the condition of sphericity holds (the test is not significant). Whew – the spericity condition is met. Repeated Measures Lecture - 11 2/6/2016 The Univariate Tests. Since the sphericity condition is met, we can use the test of significance printed in the top line of the following box. If sphericity didn't hold, then we would either report the multivariate F above or report one of the F's from the 2nd, 3rd, or 4th lines of the table below. Each of them performs an adjustment for lack of sphericity. The specifics of the adjustment vary across tests. Note that in the top three lines, the null is rejected. The last line reports the most conservative adjustment. But since sphericity holds, we can report the F in the top line and conclude that there are significant differences between mean heart rates across time periods. The difference between these results and those of the Multivariate tests shown above highlight the power differences between the two types of tests. If you want to detect differences, use the tests shown below. Tests of user-specified Contrasts The following tests are of the contrast specified above. If no contrasts had been specified, it would not be printed. The individual contrasts compare the means of the 30-, 60-, and 120minute measurements with the pre-measure. The means at the 60- and 120-minute intervals were significantly lower than the pre-measure mean as we might expect if it took time for the drug to take effect. Tests of Between-Subjects Effects. What are between subjects factors doing in a repeated measures analysis? The answer is that there is only a technical between subjects factor here - the difference between the mean of all the scores and 0, known in regression parlance as the intercept. That's what's being tested in the box below. The null hypothesis is that the intercept is 0. Don’t let its presence confuse you. Repeated Measures Lecture - 12 The plot below is that printed by GLM. 2/6/2016 Repeated Measures Lecture - 13 2/6/2016 Repeated Measures ANOVA 1 Between Groups Factor / 1 Repeated Measures Factor The Bridge effectiveness study. Situation: A school system provides middle and high school students a period for whatever intellectual activity they wish to engage in. Some choose chess; some choose reading, some choose programming, some choose to play bridge. One of the people involved with the bridge classes was interested in whether playing bridge during the hour periods over a long period of time would lead to greater performance on standardized tests of school-related achievement than the other activities. The data here provide a test of that notion. File: BridgeData1_3Periods_RM090208For595.sav Key variables: group0Vs1. Variable representing the research condition. 0=Control; 1=Bridge ss_math_tot1 ss_math_tot2 ss_math_tot3 Standardized total scores on math achievement at time period 1, time period 2 and time period 3. ss_comput1 ss_comput2 ss_comput3: Standardized “computational??” achievement ss_lang_tot1 ss_lang_tot2 ss_lang_tot3: Standardized language achievement (The names of the variables are idiosyncratic because of the way the data were given to me. I’ve just been too lazy to change them to something easier to follow.) What outcome would be ideal here?? A. Groups start out at about the same level since the two groups should be equivalent at the start. B. The Bridge group scores on standardized tests increase at a faster rate than the Control group scores. C. The interaction term would be significant. Pictorially, this is the ideal outcome of this research project . . . Bridge Group Control Group Repeated Measures Lecture - 14 The data Analyze -> General Linear Model -> Repeated Measures. 2/6/2016 Repeated Measures Lecture - 15 2/6/2016 Repeated Measures Lecture - 16 2/6/2016 The syntax – analysis of math_tot scores. GLM ss_math_tot1 ss_math_tot2 ss_math_tot BY group0Vs1 /WSFACTOR=time 3 Polynomial /METHOD=SSTYPE(3) /PLOT=PROFILE(time*group0Vs1) /PRINT=DESCRIPTIVE ETASQ OPOWER PARAMETER /CRITERIA=ALPHA(.05) /WSDESIGN=time /DESIGN=group0Vs1. The cell means displayed in a 2 way table Group Control Bridge Time1 234 233 Time2 249 250 Time3 261 261 The significant “time” effect reflects the fact that the scores for all kids changed over time – the means increased. Alas, the nonsignificant “time*group0Vs1” effect means that neither group increased at a higher rate than the other. Repeated Measures Lecture - 17 2/6/2016 There is a Between-subjects effect in this analysis – the Bridge group vs Control group factor. The table below presents a test of the significance of difference between the overall mean of the Bridge group scores vs the overall mean of the Control group scores. The difference is not significant. Repeated Measures Lecture - 18 I’ve rarely seen such a striking affirmation of the null. 2/6/2016 Repeated Measures Lecture - 19 2/6/2016 Policy Capturing using Repeated Measures Those in the I-O program know that cognitive ability is the single best predictor of performance in a variety of situations, including academia. Although we know that cognitive ability is a valid predictor, it’s possible that others do not have that knowledge. If those others are involved in selection of employees, that lack of knowledge could be a serious problem for the organization at which they are employed. This is a hypothetical research project to investigate whether or not HR persons involved in selection understand these basic results. It uses a technique called policy capturing. In policy capturing, persons are given scenarios and asked to rate each scenario on one or more dimensions. The scenarios are created so that they differ in specific ways, although the raters are not told about these differences. After the ratings have been gathered, analyses of the ratings are conducted to determine if the differences in the scenarios affected the ratings. If the raters had specific, perhaps hidden, policies controlling their ratings of the scenarios, e.g., prejudice against Blacks or females, those policies would be revealed by differences in ratings of the different subgroups distinguished by the characteristics impacted by the policies. In this example, the dependent variable, the ratings, will be of suitability for employement. Let’s suppose that we have a four-item scale of suitability, with reliability = .8. We will have persons rate two different hypothetical applicants. The applicants will be described as having different levels of cognitive ability – either low or high Cognitive Ability High Low Each person rated two scenarios. The order in which each person saw the scenarios was randomized across participants. If the participants understood contemporary validity research and implemented that knowledge in their policies regarding suitability for employment, we would expect A positive relationship of suitability to cognitive ability – scenarios describing applicants with high cognitive ability will be rated higher than scenarios describing applicants with low cognitive ability. Whatever the results, we can say that our analysis has captured the policies of the raters. If the policy was to ignore cognitive ability when rating suitability for employment, that policy would be reflected by a nonsignificant difference in mean suitability ratings between the low and high cognitive ability scenarios. Repeated Measures Lecture - 20 2/6/2016 The Analysis The analysis could be a simple paired-sample t-test, but I’ll do it using the GLM Repeated Measures procedure, since I want to expand the example following this analysis. Here are some of the data . . . Repeated Measures Lecture - 21 2/6/2016 General Linear Model Notes Output Created Comments Input Missing Value Handling 23-SEP-2015 10:18:01 Filter Weight Split File N of Rows in Working Data File Definition of Missing Cases Used Syntax Resources Within-Subjects Factors Measure: MEASURE_1 ca Dependent Variable 1 lowca 2 highca Processor Time Elapsed Time <none> <none> <none> 80 User-defined missing values are treated as missing. Statistics are based on all cases with valid data for all variables in the model. GLM lowca highca /WSFACTOR=ca 2 Polynomial /METHOD=SSTYPE(3) /PLOT=PROFILE(ca) /PRINT=DESCRIPTIVE ETASQ OPOWER /CRITERIA=ALPHA(.05) /WSDESIGN=ca. 00:00:00.14 00:00:00.14 Repeated Measures Lecture - 22 2/6/2016 Descriptive Statistics lowca highca Mean .1536 .4583 Std. Deviation .91630 .86124 N 80 80 Multivariate Testsa Effect ca Pillai's Trace Wilks' Lambda Hotelling's Trace Roy's Largest Root a. Design: Intercept Within Subjects Design: ca b. Exact statistic c. Computed using alpha = .05 Value .176 .824 .214 .214 F 16.872b 16.872b 16.872b 16.872b Hypothesis df 1.000 1.000 1.000 1.000 Error df 79.000 79.000 79.000 79.000 Partial Eta Squared .176 .176 .176 .176 Sig. .000 .000 .000 .000 Noncent. Parameter 16.872 16.872 16.872 16.872 Observed Powerc .982 .982 .982 .982 Mauchly's Test of Sphericitya Measure: MEASURE_1 Epsilonb Approx. ChiGreenhouseWithin Subjects Effect Mauchly's W Square df Sig. Geisser Huynh-Feldt Lower-bound ca 1.000 .000 0 . 1.000 1.000 1.000 Tests the null hypothesis that the error covariance matrix of the orthonormalized transformed dependent variables is proportional to an identity matrix. a. Design: Intercept Within Subjects Design: ca b. May be used to adjust the degrees of freedom for the averaged tests of significance. Corrected tests are displayed in the Tests of Within-Subjects Effects table. Tests of Within-Subjects Effects Measure: MEASURE_1 Source ca Error(ca) Sphericity Assumed GreenhouseGeisser Huynh-Feldt Lower-bound Sphericity Assumed Type III Sum of Squares 3.714 GreenhouseGeisser Huynh-Feldt Lower-bound 1 Mean Square 3.714 F 16.872 Sig. .000 Partial Eta Squared .176 Noncent. Parameter 16.872 Observed Powera .982 3.714 1.000 3.714 16.872 .000 .176 16.872 .982 3.714 3.714 1.000 1.000 3.714 3.714 16.872 16.872 .000 .000 .176 .176 16.872 16.872 .982 .982 17.390 79 .220 17.390 79.000 .220 17.390 79.000 .220 17.390 79.000 .220 df a. Computed using alpha = .05 Tests of Within-Subjects Contrasts Measure: MEASURE_1 Type III Sum of Source ca Squares ca Linear 3.714 Error(ca) Linear 17.390 df 1 Mean Square 3.714 79 .220 F 16.872 Sig. .000 Partial Eta Squared .176 Noncent. Parameter 16.872 Observed Powera .982 a. Computed using alpha = .05 Tests of Between-Subjects Effects Measure: MEASURE_1 Transformed Variable: Average Type III Sum of Source Squares Intercept 14.978 Error 107.535 a. Computed using alpha = .05 df 1 Mean Square 14.978 79 1.361 F 11.004 Sig. .001 Partial Eta Squared .122 Noncent. Parameter 11.004 Observed Powera .906 Repeated Measures Lecture - 23 Profile Plots 2/6/2016 Repeated Measures Lecture - 24 2/6/2016 Policy Capturing using Factorial Repeated Measures (Pardon the repetition from the previous example.) Those in the I-O program know that cognitive ability is the single best predictor of performance in a variety of situations, including academia. Many also know that conscientiousness is also a valid, though less efficacious predictor of performance, also including performance in academia. Although we know that these two characteristics are valid predictors, it’s possible that others do not have that knowledge. If those others are involved in selection of employees, that could be a serious problem for the organization at which they are employed. This is a hypothetical research project to investigate whether or not HR persons involved in selection understand these basic results. It uses a technique called policy capturing. In policy capturing, persons are given scenarios and asked to rate each scenario on one or more dimensions. The scenarios are created so that they differ in specific ways, although the raters are not told about these differences. After the ratings have been gathered, analyses of the ratings are conducted to determine if the differences in the scenarios affected the ratings. If the raters had specific policies regarding the differences in the scenarios, e.g., prejudice against Blacks or females, those policies would be revealed by differences in ratings of the different subgroups distinguished by those policies. In this example, the dependent variable, the ratings, will be of suitability for employement. Let’s suppose that we have a four-item scale of suitability, with reliability = .8. We will have persons rate four different hypothetical applicants. The applicants will be described as having different combinations of cognitive ability and conscientiousness – four combinations arranged factorially as shown in the following table . . . Conscientiousness High Low Cognitive Ability High Low H-L L-L H-H L-H Each person rated four scenarios. The order in which each person saw the scenarios was randomized across participants. If the participants understood contemporary validity research, we would expect 1. A positive relationship of suitability to cognitive ability – scenarios describing applicants with high cognitive ability will be rated higher than scenarios describing applicants with low cognitive ability. 2. A positive reltionship of suitability to conscientiousness – scenarios describing applicants with high conscientiousness will be rated higher than scenarios describing applicants with low conscientiousness. 3. No interaction between the effects of cognitive ability and conscientiousness on ratings – the effect of cognitive ability will be the same across conscientiousness levels and the effect of conscientiousness will be the same across levels of cognitive ability. Repeated Measures Lecture - 25 2/6/2016 The data would be analyzed using a repeated measures factorial design. The data would be conceptualized in the data editor as 4 columns ID 1 2 3 4 5 Etc Low C Low CA rating rating . . . etc Low C High CA rating rating . . . etc High C Low CA rating rating . . . etc 1 Here’s part of the actual hypothetical data . . . High C High CA rating rating . . . etc Repeated Measures Lecture - 26 2/6/2016 Repeated Measures Lecture - 27 2/6/2016 General Linear Model Notes Output Created Comments Input Missing Value Handling 23-SEP-2015 10:27:13 Filter Weight Split File N of Rows in Working Data File Definition of Missing Cases Used <none> <none> <none> 80 User-defined missing values are treated as missing. Statistics are based on all cases with valid data for all variables in the model. GLM lowC_lowCA lowC_highCA highC_lowCA highC_highCA /WSFACTOR=C 2 Polynomial CA 2 Polynomial /METHOD=SSTYPE(3) /PLOT=PROFILE(CA*C) /EMMEANS=TABLES(C*CA) /PRINT=DESCRIPTIVE ETASQ OPOWER /CRITERIA=ALPHA(.05) /WSDESIGN=C CA C*CA. 00:00:00.17 00:00:00.16 Syntax Resources Processor Time Elapsed Time Within-Subjects Factors Measure: MEASURE_1 C CA Dependent Variable 1 1 lowC_lowCA 2 lowC_highCA 2 1 highC_lowCA 2 highC_highCA Descriptive Statistics lowC_lowCA lowC_highCA highC_lowCA highC_highCA Mean .0213 .3186 .2859 .5980 Std. Deviation .98327 1.00520 1.10178 .95945 N 80 80 80 80 Multivariate Testsa Effect C Value .135 .865 .156 .156 .176 .824 .214 .214 .000 1.000 .000 .000 Pillai's Trace Wilks' Lambda Hotelling's Trace Roy's Largest Root CA Pillai's Trace Wilks' Lambda Hotelling's Trace Roy's Largest Root C * CA Pillai's Trace Wilks' Lambda Hotelling's Trace Roy's Largest Root a. Design: Intercept Within Subjects Design: C + CA + C * CA b. Exact statistic c. Computed using alpha = .05 F 12.288b 12.288b 12.288b 12.288b 16.872b 16.872b 16.872b 16.872b .009b .009b .009b .009b Hypothesis df 1.000 1.000 1.000 1.000 1.000 1.000 1.000 1.000 1.000 1.000 1.000 1.000 Error df 79.000 79.000 79.000 79.000 79.000 79.000 79.000 79.000 79.000 79.000 79.000 79.000 Sig. .001 .001 .001 .001 .000 .000 .000 .000 .923 .923 .923 .923 Partial Eta Squared .135 .135 .135 .135 .176 .176 .176 .176 .000 .000 .000 .000 Noncent. Parameter 12.288 12.288 12.288 12.288 16.872 16.872 16.872 16.872 .009 .009 .009 .009 Observed Powerc .934 .934 .934 .934 .982 .982 .982 .982 .051 .051 .051 .051 Mauchly's Test of Sphericitya Measure: MEASURE_1 Epsilonb Approx. ChiGreenhouseWithin Subjects Effect Mauchly's W Square df Sig. Geisser Huynh-Feldt Lower-bound C 1.000 .000 0 . 1.000 1.000 1.000 CA 1.000 .000 0 . 1.000 1.000 1.000 C * CA 1.000 .000 0 . 1.000 1.000 1.000 Tests the null hypothesis that the error covariance matrix of the orthonormalized transformed dependent variables is proportional to an identity matrix. a. Design: Intercept Within Subjects Design: C + CA + C * CA b. May be used to adjust the degrees of freedom for the averaged tests of significance. Corrected tests are displayed in the Tests of Within-Subjects Effects table. Repeated Measures Lecture - 28 2/6/2016 Tests of Within-Subjects Effects Measure: MEASURE_1 Source C Error(C) Sphericity Assumed GreenhouseGeisser Huynh-Feldt Lower-bound Sphericity Assumed Type III Sum of Squares 5.916 GreenhouseGeisser Huynh-Feldt Lower-bound CA Error(CA) Sphericity Assumed GreenhouseGeisser Huynh-Feldt Lower-bound Sphericity Assumed GreenhouseGeisser Huynh-Feldt Lower-bound C * CA Error(C*CA ) Sphericity Assumed GreenhouseGeisser Huynh-Feldt Lower-bound Sphericity Assumed GreenhouseGeisser Huynh-Feldt Lower-bound 1 Mean Square 5.916 5.916 1.000 5.916 12.288 .001 .135 12.288 .934 5.916 5.916 1.000 1.000 5.916 5.916 12.288 12.288 .001 .001 .135 .135 12.288 12.288 .934 .934 38.034 79 .481 38.034 79.000 .481 38.034 79.000 .481 38.034 79.000 .481 7.428 1 7.428 16.872 .000 .176 16.872 .982 7.428 1.000 7.428 16.872 .000 .176 16.872 .982 7.428 7.428 1.000 1.000 7.428 7.428 16.872 16.872 .000 .000 .176 .176 16.872 16.872 .982 .982 34.780 79 .440 34.780 79.000 .440 34.780 79.000 .440 34.780 79.000 .440 .004 1 .004 .009 .923 .000 .009 .051 .004 1.000 .004 .009 .923 .000 .009 .051 .004 .004 1.000 1.000 .004 .004 .009 .009 .923 .923 .000 .000 .009 .009 .051 .051 36.940 79 .468 36.940 79.000 .468 36.940 79.000 .468 36.940 79.000 .468 df F 12.288 Sig. .001 Partial Eta Squared .135 Noncent. Parameter 12.288 Observed Powera .934 a. Computed using alpha = .05 Tests of Within-Subjects Contrasts Measure: MEASURE_1 Source C C Linear Error(C) Linear CA Type III Sum of Squares df Mean Square 5.916 1 5.916 38.034 79 .481 CA Linear 7.428 1 7.428 Error(CA) Linear 34.780 79 .440 C * CA Linear Error(C*CA) Linear Linear Linear .004 1 .004 36.940 79 .468 F Sig. Partial Eta Squared Noncent. Parameter Observed Powera 12.288 .001 .135 12.288 .934 16.872 .000 .176 16.872 .982 .009 .923 .000 .009 .051 a. Computed using alpha = .05 Tests of Between-Subjects Effects Measure: MEASURE_1 Transformed Variable: Average Type III Sum of Source Squares Intercept 29.956 Error 215.070 a. Computed using alpha = .05 df 1 Mean Square 29.956 79 2.722 F 11.004 Sig. .001 Partial Eta Squared .122 Noncent. Parameter 11.004 Observed Powera .906 Repeated Measures Lecture - 29 2/6/2016 Estimated Marginal Means C * CA Measure: MEASURE_1 C 1 2 CA 1 2 1 2 Mean .021 .319 .286 .598 Std. Error .110 .112 .123 .107 95% Confidence Interval Lower Bound Upper Bound -.197 .240 .095 .542 .041 .531 .384 .812 Profile Plots The CA difference indicates that the hypothetical raters value persons with high CA more than those with low CA. The C difference indicates that the hypothetical raters value persons high in C more than those with low C. There is no interaction of the effectsof CA and C on ratings. Repeated Measures Lecture - 30 2/6/2016 Repeated Measures 1 Quantitative Between Subject Factor 1 Repeated Measures Factor Nhung Nguyen’s dissertation data. Nhung’s dissertation 1. Gave the Big 5 to 200 persons under two instruction conditions- once to respond honestly (Honest instruction) and again to respond in a “way that would best guarantee that you would get a customer service representative job.” (Instructed Faking). 2. Gave the Wonderlic test – a test of cognitive ability that correlates very highly with standard IQ tests. 3. Computed a faking ability score for each Big 5 dimension. This score was computed as the difference between the score under the Faking instruction and the score under the Honest instruction standardized so that faking ability scores on the Big Five dimensions were comparable. Called faking ability because participants were instructed to fake in the Instructed Faking condition. Questions: 1. The Repeated Measures Main Effect: Are there differences in the mean faking amounts of the different Big 5 Personality dimensions? That is, are the Big Five dimensions equally fakable? 2. The Between Subjects Main Effect: Are there differences in overall average faking between the various levels of cognitive ability? Did people who scored high on cognitive ability fake more than those who score low? Put another way: is there a relationship between faking ability and cognitive ability? 3. The Interaction of the Between Subjects and Repeated Measures Effects: Are differences in faking between the Big 5 measures the same across levels of cognitive ability? or Is the relationship of faking ability to cognitive ability the same for different Big 5 dimensions. The analysis is a repeated measures ANOVA with a Quantitative Between subjects factor (the Wonderlic scores). Because the Between Subjects Factor is quantitative, it will be specified differently than a nominal factor and the effects associated with it will be interpreted and presented differently than those in the previous example, which involved a nominal (group0vs1) between subjects factor. GET FILE='E:\MdbR\Nhung\SJTPaper \ThesisPaperFile030907.sav'. GLM dsurg dagree dcons des dopen WITH gscore /WSFACTOR = persdim 5 Polynomial /METHOD = SSTYPE(3) /PLOT = PROFILE( persdim ) /EMMEANS = TABLES(persdim) WITH(gscore=MEAN) /PRINT = DESCRIPTIVE ETASQ OPOWER /CRITERIA = ALPHA(.05) /WSDESIGN = persdim /DESIGN = gscore . The syntax generated by the pull down menus shown below. Repeated Measures Lecture - 31 2/6/2016 Specification of Factors: Analyze -> General Linear Model -> Repeated Measures Quantitative between subjects factors must be put in the covariates box in GLM. Repeated Measures Lecture - 32 2/6/2016 Hypotheses Tested The Repeated Measures Main Effect (Surgency is extraversion) µdsurg µdagree µdcons µdes µdopen H0: Equality of means of population of faking scores on the 5 personality dimensions. Repeated Measures Lecture - 33 2/6/2016 The Between Subjects Main Effect Average Diff scores -.10 = Av Diff Score 1 -.11 = Av Diff Score 2 0 = Av Diff Score 3 1.14 = Av Diff Score 4 1.15 = Av Diff Score 5 0.86 = Av Diff Score 6 -0.04 = Av Diff Score 7 0.06 = Av Diff Score 8 And so on . . Faking Ability H0: The correlation of gscore with the average difference scores in population is 0. gscore Note that this hypothesis is NOT stated as equality of means. If we had observed, say, 50 people with CA = 20, 50 with CA=25, and 50 with CA=30, and 50 with CA=35, then it could have been stated as a hypothesis that the means of the faking scores at the 4 levels of CA were equal. But since we have multiple CA values with just a few scores at each, perhaps only 1, it is more efficient to examine the relationship of faking to CA, rather than to examine the differences in means. The Interaction of Persdim and gscore H0 : The correlation of gscore with individual dimension faking scores will be the same for each individual dimension. or Differences between dimension faking score means will be the same for each level of gscore. Repeated Measures Lecture - 34 2/6/2016 General Linear Model The multivariate tests below indicate that there is NOT a main effect of Persdim (Personality dimension). This means that there are not overall significant differences in the amount of faking between the 5 dimensions. Within-S ubj ec ts Fa ctors Me asure : MEA SURE_1 PE RSDI M 1 De pende nt Va riable DS URG 2 DA GREE 3 DCONS 4 DE S 5 DO PEN The significant interaction suggests, however, that there are gscore-specific differences in the amount of faking between the 5 dimensions, i.e., differences that vary with gscore.. De scriptiv e Statis tics Me an .47 97 DS URG Std . Deviatio n 1.0 1526 N 20 3 DA GRE E .35 56 1.0 0172 20 3 DCONS .58 54 .97 258 20 3 DE S .74 74 1.0 6989 20 3 DO PEN .39 21 .98 581 20 3 Equivalently, the interaction means that the correlation of gscore with faking of each dimensions varies from one dimension to the next. Multiv a riate Tests c Eff ect PE RSDI M PE RSDI M * GS CORE Pil lai's T race Va lue .04 0 F Hy pothe sis df 2.0 41 b 4.0 00 Error df 19 8.000 Sig . .09 0 Pa rtial E ta Sq uared .04 0 No ncen t. Pa rame ter 8.1 63 a Ob serve d Po wer .60 3 Wi lks' La mbd a .96 0 2.0 41 b 4.0 00 19 8.000 .09 0 .04 0 8.1 63 .60 3 Ho tellin g's Trace .04 1 2.0 41 b 4.0 00 19 8.000 .09 0 .04 0 8.1 63 .60 3 Ro y's La rgest Ro ot .04 1 2.0 41 4.0 00 19 8.000 .09 0 .04 0 8.1 63 .60 3 Pil lai's T race .05 1 2.6 35 b 4.0 00 19 8.000 .03 5 .05 1 10 .538 .73 0 Wi lks' La mbd a .94 9 2.6 35 b 4.0 00 19 8.000 .03 5 .05 1 10 .538 .73 0 Ho tellin g's Trace .05 3 2.6 35 b 4.0 00 19 8.000 .03 5 .05 1 10 .538 .73 0 Ro y's La rgest Ro ot .05 3 2.6 35 4.0 00 19 8.000 .03 5 .05 1 10 .538 .73 0 b b a. Co mput ed using a lpha = .05 b. Ex act st atistic c. De sign: Interc ept+ GSCORE Wi thin S ubje cts Design : PERSDIM Ma uchly's Te st of Sphe ricity b Me asure : ME ASURE_1 Ep silon Wit hin S ubje cts Eff ect PE RSDI M Ma uchly's W .88 8 Ap prox. Ch i-Squ are 23. 663 df 9 Sig . .00 5 Gre enho useGe isser .94 6 a Hu ynh-Feldt .97 1 Lower-b ound .25 0 Te sts the null hypo thesi s that the e rror covari ance matrix of t he orthono rmal ized transf orme d dep ende nt va riable s is p roportiona l to a n ide ntity matrix. a. Ma y be used to ad just th e de grees of fre edom for the a verag ed te sts of signi fican ce. Correct ed te sts are disp laye d in the Test s of Within -Subj ects E ffect s tabl e. b. De sign: Intercept+G SCO RE Wit hin S ubje cts De sign: PERSDIM The failure to meet the sphericity criterion means that we cannot use the “Sphericity Assumed” F printed below. Repeated Measures Lecture - 35 2/6/2016 Tes ts of Within-Subj ec ts Effects Me asure : ME ASURE_1 So urce PE RSDI M Ty pe III Sum of Squa res 5.4 37 Sp heric ity Assume d PE RSDI M * GS CORE Error(PE RSDIM) 4 Me an S quare 1.3 59 F 2.2 17 Sig . .06 6 Pa rtial E ta Sq uared .01 1 No ncen t. Pa rame ter 8.8 66 df a Ob serve d Po wer .65 3 Gre enho use-Geisse r 5.4 37 3.7 85 1.4 37 2.2 17 .06 9 .01 1 8.3 89 .63 5 Hu ynh-Feldt 5.4 37 3.8 85 1.3 99 2.2 17 .06 8 .01 1 8.6 13 .64 4 Lo wer-b ound 5.4 37 1.0 00 5.4 37 2.2 17 .13 8 .01 1 2.2 17 .31 7 Sp heric ity Assume d 6.0 67 4 1.5 17 2.4 74 .04 3 .01 2 9.8 94 .70 8 Gre enho use-Geisse r 6.0 67 3.7 85 1.6 03 2.4 74 .04 6 .01 2 9.3 61 .69 0 Hu ynh-Feldt 6.0 67 3.8 85 1.5 61 2.4 74 .04 5 .01 2 9.6 11 .69 8 Lo wer-b ound 6.0 67 1.0 00 6.0 67 2.4 74 .11 7 .01 2 2.4 74 .34 7 Sp heric ity Assume d 49 2.981 80 4 .61 3 Gre enho use-Geisse r 49 2.981 76 0.700 .64 8 Hu ynh-Feldt 49 2.981 78 0.974 .63 1 Lo wer-b ound 49 2.981 20 1.000 2.4 53 a. Co mput ed using a lpha = .05 Tes ts of Within-Subj ec ts Contras ts Me asure : ME ASURE_1 PE RSDI M Lin ear So urce PE RSDI M Ty pe III Sum of Squa res 3.7 15 Qu adrat i c Cu bic PE RSDI M * GS CORE F 5.4 52 Sig . .02 1 Pa rtial E ta Sq uared .02 6 No ncen t. Pa rame ter 5.4 52 .05 2 1 .05 2 .08 8 .76 8 .00 0 .08 8 .06 0 a Ob serve d Po wer .64 2 .05 8 1 .05 8 .07 9 .77 9 .00 0 .07 9 .05 9 Ord er 4 1.6 12 1 1.6 12 3.5 97 .05 9 .01 8 3.5 97 .47 1 Lin ear 2.9 96 1 2.9 96 4.3 97 .03 7 .02 1 4.3 97 .55 1 .62 2 1 .62 2 1.0 48 .30 7 .00 5 1.0 48 .17 5 Qu adrat i c Cu bic .68 1 1 .68 1 .93 4 .33 5 .00 5 .93 4 .16 1 1.7 67 1 1.7 67 3.9 43 .04 8 .01 9 3.9 43 .50 7 Lin ear 13 6.969 20 1 .68 1 Qu adrat i c 11 9.312 20 1 .59 4 Cu bic 14 6.621 20 1 .72 9 90 .079 20 1 .44 8 Ord er 4 Error(PE RSDIM) 1 Me an S quare 3.7 15 df Ord er 4 a. Co mput ed using a lpha = .05 The Tests of Within-subjects Contrasts are automatically printed, but make no sense since the personality dimensions are not on a quantitative dimension. Tes ts of Betw een-Subj ects Effec ts Me asure : ME ASURE_1 Tra nsformed Varia ble: A vera ge So urce Inte rcep t GS COR E Error Typ e III Sum of S qua res 6.2 92 1 Me an S quare 6.2 92 F 2.6 32 Sig . .10 6 Pa rtial E ta Sq uared .01 3 No ncent . Pa rame ter 2.6 32 49. 924 1 49. 924 20. 882 .00 0 .09 4 20. 882 480 .540 201 2.3 91 df a Ob serve d Power .36 5 .99 5 a. Co mput ed using a lpha = .05 The significant main effect of gscore means that there is a correlation of overall faking with cognitive ability. That is, there are differences in amount of faking at different levels of cognitive ability. Inspection of the graph below shows that it’s positive – smarter people faked more. Repeated Measures Lecture - 36 2/6/2016 Estimated Marginal Means PE RSDI M Me asure : ME ASURE_1 95 % Co nfide nce I nterva l PE RSDI M 1 Me an .48 0 a Std . Erro r .06 9 Lo wer B ound .34 4 Up per B ound .61 5 2 .35 6 a .06 8 .22 2 .48 9 3 .58 5 a .06 7 .45 3 .71 8 4 .74 7 a .07 3 .60 4 .89 1 5 .39 2 a .06 9 .25 6 .52 8 a. Co varia tes ap pearing in the mode l are evalu ated at th e fol lowin g valu es: G SCO RE wond erlic test score = 24. 61. Profile Plots These are identical to the observed means since there is only one group of respondents. Dimensions are listed as E A C S O. So the largest amount of faking was in the Emotional Stability dimension (4th in the list). The least was in Agreeableness and Openness. But remember that the differences in means were not officially significant. Estimated Marginal Means of MEASURE_1 .8 .7 Note – Persdim main effect was not significant, so this graph shouldn’t be overinterpreted. .6 .5 .4 .3 1 2 Ext PERSDIM 3 Agr 4 Con 5 Sta Opn Graphing the Main Effect of gscore. I computed an average faking score (those values illustrated on the right on p. 22 above.) graph /scatterplot = gscore with fakescor. Graph 4 3 2 1 FAKESCOR 0 -1 -2 Rsq = 0.0941 0 10 20 30 40 50 w onderlic test score This is a display of the main effect of cognitive ability. The average difference scores are called FAKESCOR in this plot. On the right, persons with similar gscore values have been grouped together. This shows the positive relationship of faking to Gscore more strikingly. Repeated Measures Lecture - 37 2/6/2016 The interactions. Recall what the interaction is: The relationship of average faking to gscore differed across dimensions. Below are displays of the relationship of faking for each dimension to cognitive ability. These displays are analogous to the “interaction” plots when the between subjects factor is a categorical factor. I’m not sure of the explanation of the interaction. Why is the relationship of faking to cognitive ability stronger for Extraversion than it is for Openness, for example? That is, smart people faked Extraversion a lot more than less smart people did. But smart people didn’t fake Openness much more than less smart people. ?? graph /scatterplot = gscore with dsurg. Graph r = +.26 Graph r = +.27 5 5 4 4 3 3 2 2 1 1 0 0 DES DSURG -1 -1 -2 Rsq = 0.0685 0 -2 10 20 30 40 50 Rsq = 0.0744 0 10 20 30 40 50 w onderlic test score w onderlic test score graph /scatterplot = gscore with dopen. graph /scatterplot = gscore with dagree. Graph Graph r = +.09 r = +.28 4 4 3 3 2 2 1 1 0 0 -1 -1 DOPEN DAGREE -2 -3 -2 -3 Rsq = 0.0076 0 -4 10 20 30 40 50 Rsq = 0.0802 0 10 20 30 40 50 w onderlic test score w onderlic test score graph /scatterplot = gscore with dcons. Graph r = +.19 It may be that for some items, there may be disagreement on what constitutes a “good” response. That is, some people might have thought that agreeing was the appropriate way to fake while others thought that disagreeing was the appropriate way to fake. The result was that the relationships of faking to cognitive ability was suppressed for those items. 6 4 2 0 DCONS -2 -4 Rsq = 0.0362 0 10 20 30 40 50 w onderlic test score graph /scatterplot = gscore with des. Repeated Measures Lecture - 38 2/6/2016 Using a correlation matrix to examine the individual dimension correlations. correlation gscore dsurg to dopen fakescore. Correlations Wa rnings Te xt: FA KES CORE A variab le na me i s more tha n 8 characters l ong. Only the first 8 characters will be used. Correlations GS CORE wo nderl ic test sco re GS CORE wo nderlic te st sco re Pe arson Co rrelat ion Sig . (2-t ailed ) DS URG DA GREE DCONS DE S DO PEN FA KESCOR DS URG DA GREE DCONS DE S DO PEN FA KESCOR 1 .27 3 .28 3 .19 0 .26 2 .08 7 .30 7 . .00 0 .00 0 .00 7 .00 0 .21 5 .00 0 N 203 203 203 203 203 203 203 Pe arson Co rrelat ion .27 3 1 .41 6 .32 9 .45 2 .37 5 .71 9 Sig . (2-t ailed ) .00 0 . .00 0 .00 0 .00 0 .00 0 .00 0 N 203 203 203 203 203 203 203 Pe arson Co rrelat ion .28 3 .41 6 1 .39 7 .26 0 .33 1 .66 6 Sig . (2-t ailed ) .00 0 .00 0 . .00 0 .00 0 .00 0 .00 0 N 203 203 203 203 203 203 203 Pe arson Co rrelat ion .19 0 .32 9 .39 7 1 .49 8 .43 5 .73 6 Sig . (2-t ailed ) .00 7 .00 0 .00 0 . .00 0 .00 0 .00 0 N 203 203 203 203 203 203 203 Pe arson Co rrelat ion .26 2 .45 2 .26 0 .49 8 1 .45 1 .75 0 Sig . (2-t ailed ) .00 0 .00 0 .00 0 .00 0 . .00 0 .00 0 N 203 203 203 203 203 203 203 Pe arson Co rrelat ion .08 7 .37 5 .33 1 .43 5 .45 1 1 .71 8 Sig . (2-t ailed ) .21 5 .00 0 .00 0 .00 0 .00 0 . .00 0 N 203 203 203 203 203 203 203 Pe arson Co rrelat ion .30 7 .71 9 .66 6 .73 6 .75 0 .71 8 1 Sig . (2-t ailed ) .00 0 .00 0 .00 0 .00 0 .00 0 .00 0 . N 203 203 203 203 203 203 203 Repeated Measures Lecture - 38 02/06/16 Repeated Measures Lecture - 39 2/6/2016 Repeated Measures ANOVA 1 Between Groups Factor 2 Repeated Measures Factors The data are from Myers & Well, p 313 although the story describing the data is different from theirs. Memory for two types of event, one of some interest to the persons (C1), the other of less interest to them (C2) is tested at three time periods (B1, B2, and B3). The tests are performed under two conditions of distraction, much distraction (A1) and little distraction (A2). The interest here is on the effects of interest on memory, the effects of distraction on memory, and the the interaction of interest and distraction – whether the effect of distraction is the same for memory of interesting and uninteresting material. The data matrix looks as follows . . Interesting material Uninteresting material C1B1 C1B2 C1B3 C2B1 C2B2 C2B3 Mike – show the pull down menus in more detail. A 80 46 51 72 68 65 48 37 49 57 40 44 45 34 36 50 33 36 76 42 45 66 58 56 45 34 38 51 38 37 41 33 30 42 30 28 1 1 1 1 1 1 Much distraction 70 88 58 63 78 84 55 69 60 57 81 82 52 66 54 52 75 80 68 91 50 61 79 80 57 74 41 58 78 73 56 70 38 56 74 76 2 2 2 2 2 2 Little distraction Specify the factor that varies slowest across columns first Then specif the factor whose levels vary fastest next, in this case the B factor . . . Repeated Measures Lecture - 39 02/06/16 Repeated Measures Lecture - 40 Click on the “Define” button, then . . . Repeated Measures Lecture - 40 02/06/16 2/6/2016 Repeated Measures Lecture - 41 2/6/2016 Click on the name of each variable in the left-hand field and click on the right-pointing arrow to move the name into the “Withing-Subjects Variables” field. GLM c1b1 c1b2 c1b3 c2b1 c2b2 c2b3 BY a /WSFACTOR = c 2 Polynomial b 3 Polynomial /METHOD = SSTYPE(3) /PLOT = PROFILE( b*a*c ) /PRINT = DESCRIPTIVE ETASQ OPOWER HOMOGENEITY /CRITERIA = ALPHA(.05) /WSDESIGN = c b c*b /DESIGN = a . W arni ngs Box's Test of Equality of Covarianc e Matric es is not comput ed because there are fewer than two nonsingular cell covariance matric es. I believe it’s given because the number of persons in each cell is not larger than the number of measures on each person. Repeated Measures Lecture - 41 02/06/16 Repeated Measures Lecture - 42 2/6/2016 Hypotheses tested The C (Interest) - Repeated Measures Main Effect C1B1 C1B2 C1B3 C2B1 C2B2 C2B3 80 46 51 72 68 65 70 88 58 63 78 84 48 37 49 57 40 44 55 69 60 57 81 82 45 34 36 50 33 36 52 66 54 52 75 80 76 42 45 66 58 56 68 91 50 61 79 80 µC1 45 34 38 51 38 37 57 74 41 58 78 73 A 41 33 30 42 30 28 56 70 38 56 74 76 1 1 1 1 1 1 2 2 2 2 2 2 Sample is treated as one giant group for the repeated measures comparisons Much distraction Little distraction µC2 The B (Time) - Repeated Measures Main Effect C1B1 C1B2 C1B3 C2B1 C2B2 C2B3 80 46 51 72 68 65 70 88 58 63 78 84 48 37 49 57 40 44 55 69 60 57 81 82 45 34 36 50 33 36 52 66 54 52 75 80 µB1 76 42 45 66 58 56 68 91 50 61 79 80 45 34 38 51 38 37 57 74 41 58 78 73 µB2 A 41 33 30 42 30 28 56 70 38 56 74 76 1 1 1 1 1 1 2 2 2 2 2 2 Sample is treated as one giant group for the repeated measures comparisons Little distraction µB3 The A – Distraction - Between-Subjects Main Effect C1B1 C1B2 C1B3 C2B1 C2B2 C2B3 Much distraction A 80 46 51 72 68 65 48 37 49 57 40 44 45 34 36 50 33 36 76 42 45 66 58 56 45 34 38 51 38 37 41 33 30 42 30 28 µA1 Much distraction 70 88 58 63 78 84 55 69 60 57 81 82 52 66 54 52 75 80 68 91 50 61 79 80 57 74 41 58 78 73 56 70 38 56 74 76 µA2 Little distraction Repeated Measures Lecture - 42 02/06/16 Repeated Measures Lecture - 43 General Linear Model Wi thin-S ubj e cts Factors Me asure: ME ASURE_ 1 C 1 2 B 1 De pend ent Va riabl e C1 B1 2 C1 B2 3 C1 B3 1 C2 B1 2 C2 B2 3 C2 B3 Be tw ee n-Subj ects Fac tors Va lue L abel A N 1 6 2 6 De scriptiv e Statis tics C1 B1 C1 B2 C1 B3 C2 B1 C2 B2 C2 B3 A 1 Me an Std . Deviatio n 63 .67 12 .88 N 6 2 73 .50 11 .86 6 To tal 68 .58 12 .87 12 1 45 .83 7.1 4 6 2 67 .33 11 .98 6 To tal 56 .58 14 .64 12 1 39 .00 6.8 7 6 2 63 .17 12 .37 6 To tal 51 .08 15 .82 12 1 57 .17 12 .75 6 2 71 .50 14 .79 6 To tal 64 .33 15 .14 12 1 40 .50 6.2 8 6 2 63 .50 14 .07 6 To tal 52 .00 15 .88 12 1 34 .00 6.0 3 6 2 61 .67 14 .50 6 To tal 47 .83 17 .91 12 Repeated Measures Lecture - 43 02/06/16 2/6/2016 Repeated Measures Lecture - 44 2/6/2016 The multivariate tests. The less powerful but more robust multivariate tests are always printed first. These tests indicate that there is a significant effect associated with Factor C (Interest), with Factor B (Time), and with the B * A (Time by Distraction) interaction. Multiv a riate Tests c Eff ect C C* A B B* A C* B Pil lai's T race Va lue .44 2 F Hyp othe sis df 7.9 35 b 1.0 00 Error df 10. 000 No ncent . Sig . Eta Squ ared Pa rame ter .01 8 .44 2 7.9 35 Ob serve d a Po wer .71 9 Wil ks' La mbd a .55 8 7.9 35 b 1.0 00 10. 000 .01 8 .44 2 7.9 35 .71 9 Ho tellin g's Trace .79 3 7.9 35 b 1.0 00 10. 000 .01 8 .44 2 7.9 35 .71 9 Ro y's La rgest Root .79 3 7.9 35 b 1.0 00 10. 000 .01 8 .44 2 7.9 35 .71 9 Pil lai's T race .10 9 1.2 26 b 1.0 00 10. 000 .29 4 .10 9 1.2 26 .17 1 Wil ks' La mbd a .89 1 1.2 26 b 1.0 00 10. 000 .29 4 .10 9 1.2 26 .17 1 Ho tellin g's Trace .12 3 1.2 26 b 1.0 00 10. 000 .29 4 .10 9 1.2 26 .17 1 Ro y's La rgest Root .12 3 1.2 26 b 1.0 00 10. 000 .29 4 .10 9 1.2 26 .17 1 Pil lai's T race .89 3 37. 678 b 2.0 00 9.0 00 .00 0 .89 3 75. 356 1.0 00 Wil ks' La mbd a .10 7 37. 678 b 2.0 00 9.0 00 .00 0 .89 3 75. 356 1.0 00 Ho tellin g's Trace 8.3 73 37. 678 b 2.0 00 9.0 00 .00 0 .89 3 75. 356 1.0 00 Ro y's La rgest Root 8.3 73 37. 678 b 2.0 00 9.0 00 .00 0 .89 3 75. 356 1.0 00 Pil lai's T race .56 7 5.8 83 b 2.0 00 9.0 00 .02 3 .56 7 11. 766 .73 4 Wil ks' La mbd a .43 3 5.8 83 b 2.0 00 9.0 00 .02 3 .56 7 11. 766 .73 4 Ho tellin g's Trace 1.3 07 5.8 83 b 2.0 00 9.0 00 .02 3 .56 7 11. 766 .73 4 Ro y's La rgest Root 1.3 07 5.8 83 b 2.0 00 9.0 00 .02 3 .56 7 11. 766 .73 4 Pil lai's T race .32 3 2.1 50 b 2.0 00 9.0 00 .17 3 .32 3 4.2 99 .32 9 Wil ks' La mbd a .67 7 2.1 50 b 2.0 00 9.0 00 .17 3 .32 3 4.2 99 .32 9 Ho tellin g's Trace .47 8 2.1 50 b 2.0 00 9.0 00 .17 3 .32 3 4.2 99 .32 9 Ro y's La rgest Root .47 8 2.1 50 b 2.0 00 9.0 00 .17 3 .32 3 4.2 99 .32 9 .18 6 1.0 28 b 2.0 00 9.0 00 .39 6 .18 6 2.0 55 .17 7 Wil ks' La mbd a .81 4 1.0 28 b 2.0 00 9.0 00 .39 6 .18 6 2.0 55 .17 7 Ho tellin g's Trace .22 8 1.0 28 b 2.0 00 9.0 00 .39 6 .18 6 2.0 55 .17 7 Ro y's La rgest Root .22 8 1.0 28 b 2.0 00 9.0 00 .39 6 .18 6 2.0 55 .17 7 C * B * A Pil lai's T race a. Co mput ed using a lpha = .05 b. Exa ct sta tistic c. De sign: Intercept+ A Wit hin S ubje cts De sign: C+B +C*B The significant C Main effect indicates that the mean amount recalled depends on the Interest. The significant B main effect indicates that the mean amount recalled depends on the time at which recall occurred. The significant B*A interaction indicates that the change in mean amount over time is different for persons under much distraction than it is for persons under little distraction. Repeated Measures Lecture - 44 02/06/16 Repeated Measures Lecture - 45 2/6/2016 The sphericity test. Mauchly's test indicates that the sphericity condition is NOT met. So we must either go with the multivariate tests or use one of the tests which adjusts for lack of sphericity. Fortunately, they all give the same result with respect to significance and they all agree with the multivariate tests with respect to significance, so the point is moot. Ma uchly's Te st of Sphe ricity b Me asure : ME ASURE_1 Ep silon Ap prox. Wit hin S ubje cts Ef fect Ma uchly's W Ch i-Squ are C 1.0 00 .00 0 df Sig . 0 a Gre enho u se-Geisser Hu ynh-Feldt Lower-b ound . 1.0 00 1.0 00 1.0 00 B .26 3 12. 031 2 .00 2 .57 6 .66 8 .50 0 C* B .49 1 6.4 02 2 .04 1 .66 3 .80 1 .50 0 Te sts the null hypo thesi s that the e rror covari ance matrix of t he orthono rmal ized transf orme d dep ende nt variable s is p roportional to an iden tity m atrix. a. Ma y be used to ad just th e de grees of fre edom for the a verag ed te sts of sign ifican ce. Corrected tests are displ ayed in th e Tests of Within -Sub jects Effects tab le. b. De sign: Intercept+ A Wit hin S ubje cts De sign: C+B +C*B Repeated Measures Lecture - 45 02/06/16 Repeated Measures Lecture - 46 2/6/2016 Tes ts of Within-Subj ec ts Effects Me asure : ME ASURE_1 So urce C C* A Error(C) B B* A Error(B) C* B Sp hericity Assume d Type III Sum of Squa res 29 2.014 Gre enho use-Geisser 29 2.014 1.0 00 29 2.014 7.9 35 .01 8 .44 2 7.9 35 .71 9 Hu ynh-Feldt 29 2.014 1.0 00 29 2.014 7.9 35 .01 8 .44 2 7.9 35 .71 9 Lo wer-b ound 29 2.014 1.0 00 29 2.014 7.9 35 .01 8 .44 2 7.9 35 .71 9 Sp hericity Assume d 45 .125 1 45 .125 1.2 26 .29 4 .10 9 1.2 26 .17 1 Gre enho use-Geisser 45 .125 1.0 00 45 .125 1.2 26 .29 4 .10 9 1.2 26 .17 1 Hu ynh-Feldt 45 .125 1.0 00 45 .125 1.2 26 .29 4 .10 9 1.2 26 .17 1 Lo wer-b ound 45 .125 1.0 00 45 .125 1.2 26 .29 4 .10 9 1.2 26 .17 1 Sp hericity Assume d 36 8.028 10 36 .803 Gre enho use-Geisser 36 8.028 10 .000 36 .803 Hu ynh-Feldt 36 8.028 10 .000 36 .803 Lo wer-b ound 36 8.028 10 .000 36 .803 Sp hericity Assume d 36 83.11 1 2 18 41.55 6 36 .410 .00 0 .78 5 72 .821 1.0 00 Gre enho use-Geisser 36 83.11 1 1.1 51 31 99.37 8 36 .410 .00 0 .78 5 41 .915 1.0 00 Hu ynh-Feldt 36 83.11 1 1.3 35 27 58.60 5 36 .410 .00 0 .78 5 48 .613 1.0 00 Lo wer-b ound 36 83.11 1 1.0 00 36 83.11 1 36 .410 .00 0 .78 5 36 .410 1.0 00 Sp hericity Assume d 61 6.333 2 30 8.167 6.0 93 .00 9 .37 9 12 .186 .83 3 Gre enho use-Geisser 61 6.333 1.1 51 53 5.385 6.0 93 .02 7 .37 9 7.0 14 .65 2 Hu ynh-Feldt 61 6.333 1.3 35 46 1.626 6.0 93 .02 1 .37 9 8.1 35 .70 1 Lo wer-b ound 61 6.333 1.0 00 61 6.333 6.0 93 .03 3 .37 9 6.0 93 .60 6 Sp hericity Assume d 10 11.55 6 20 50 .578 Gre enho use-Geisser 10 11.55 6 11 .512 87 .870 Hu ynh-Feldt 10 11.55 6 13 .351 75 .764 Lo wer-b ound 10 11.55 6 10 .000 10 1.156 df Me an S quare 1 29 2.014 F 7.9 35 No ncen t. Sig . Eta Squ ared Pa rame ter .01 8 .44 2 7.9 35 Ob serve d a Po wer .71 9 Sp hericity Assume d 5.7 78 2 2.8 89 .69 1 .51 2 .06 5 1.3 83 .15 0 Gre enho use-Geisser 5.7 78 1.3 25 4.3 59 .69 1 .46 0 .06 5 .91 6 .12 9 Hu ynh-Feldt 5.7 78 1.6 03 3.6 05 .69 1 .48 4 .06 5 1.1 08 .13 8 Lo wer-b ound 5.7 78 1.0 00 5.7 78 .69 1 .42 5 .06 5 .69 1 .11 7 Sp hericity Assume d 7.0 00 2 3.5 00 .83 8 .44 7 .07 7 1.6 76 .17 3 Gre enho use-Geisser 7.0 00 1.3 25 5.2 82 .83 8 .40 9 .07 7 1.1 10 .14 6 Hu ynh-Feldt 7.0 00 1.6 03 4.3 67 .83 8 .42 7 .07 7 1.3 43 .15 8 Lo wer-b ound 7.0 00 1.0 00 7.0 00 .83 8 .38 2 .07 7 .83 8 .13 2 Error(C* B) Sp hericity Assume d 83 .556 20 4.1 78 Gre enho use-Geisser 83 .556 13 .253 6.3 04 Hu ynh-Feldt 83 .556 16 .029 5.2 13 Lo wer-b ound 83 .556 10 .000 8.3 56 C* B*A a. Co mput ed using a lpha = .05 Repeated Measures Lecture - 46 02/06/16 Repeated Measures Lecture - 47 2/6/2016 Test of equality of variances across groups. Levene's test is for the between-subjects effects. It indicates generally that the variances are not significantly different. Le v ene 's Te st of Equa lity of Error Va rianc es a C1 B1 F .00 7 C1 B2 C1 B3 df1 1 df2 10 Sig . .93 6 3.1 86 1 10 .10 5 4.7 78 1 10 .05 4 C2 B1 .31 1 1 10 .58 9 C2 B2 5.2 05 1 10 .04 6 C2 B3 5.0 00 1 10 .04 9 Te sts th e nul l hyp othesis tha t the error varia nce of the dep ende nt va riable is eq ual a cross grou ps. a. De sign: Intercept+ A Wi thin S ubje cts Design : C+B +C*B Test of between-subjects main effect. The comparison between the two distracter conditions indicates that there is a significant difference in mean recall between the two. Tes ts of Betw een-Subj ects E ffec ts Me asure : ME ASURE_1 Tra nsformed Varia ble: A vera ge So urce Inte rcep t Typ e III Sum of S qua res 231 767. 014 df Me an S quare F 1 231 767. 014 363 .878 A 726 0.12 5 1 726 0.12 5 Error 636 9.36 1 10 636 .936 11. 399 No ncent . Sig . Eta Squ ared Pa rame ter .00 0 .97 3 363 .878 .00 7 .53 3 11. 399 a. Co mput ed using a lpha = .05 Repeated Measures Lecture - 47 02/06/16 Ob serve d a Po wer 1.0 00 .86 0 Repeated Measures Lecture - 48 2/6/2016 Profile Plots Estimated Marginal Means of MEASURE_1 70 60 Distraction. Main Effect of A: Mean recall was better in the low distraction condition. 50 40 High distraction con Low distraction cond A Estimated Marginal Means of MEASURE_1 70 60 Time. Main Effect of B: Mean recall decreased over time periods. 50 40 1 2 3 B Repeated Measures Lecture - 48 02/06/16 Repeated Measures Lecture - 49 2/6/2016 Estimated Marginal Means of MEASURE_1 59 58 Main Effect of Factor C: Mean recall was greater for more interesting material. 57 56 55 54 1 2 Less Interesting More Interesting C Estimated Marginal Means of MEASURE_1 80 70 Distraction by Time. Interaction of A and B: The effect of distraction was greater at longer recall times. (This was the only significant interaction.) 60 A 50 High distraction con dition 40 Low distraction cond 30 ition 1 2 3 B Repeated Measures Lecture - 49 02/06/16 Repeated Measures Lecture - 50 2/6/2016 B*A*C Estimated Marginal Means of MEASURE_1 Even though the three way interaction, B*A*C was not significant, it is of value to see how it would be visualized. A significant three way interaction means that a two-way interaction was not the same across levels of a third variable. At C = 1 80 70 60 A 50 High distraction con dition 40 Low distraction cond 30 ition 1 2 3 B Estimated Marginal Means of MEASURE_1 At C = 2 80 70 In this case, if the B*A*C interaction were significant, that would mean that the B*A interaction varied across levels of C. The way to visualize that is to plot the B*A interaction for each level of C. Those plots are on the left. The fact that they're nearly identical supports the fact that there was NO B*A*C interaction. If there had been, the two plots (one for C=1 and the other for C=2) would have been different. So . . . 60 A 50 High distraction con dition 40 A two-way interaction is a difference of effect of one variable across levels of a second variable. Low distraction cond 30 ition 1 B 2 3 A three-way interaction is a difference of INTERACTION of two variables across levels of a third variable. Repeated Measures Lecture - 50 02/06/16