Reporting the Strength of Effect Estimates for Simple Statistical Analyses This document was prepared as a guide for my students in Experimental Psychology. It shows how to present the results of a few simple but common statistical analyses. It also shows how to compute commonly employed strength of effect estimates. Independent Samples T When we learned how to do t tests (see T Tests and Related Statistics: SPSS), you compared the mean amount of weight lost by participants who completed two different weight loss programs. Here is SPSS output from that analysis: Group Statistics LOSS GROUP 1 2 N 6 12 Mean 22.67 13.25 Std. Deviation 4.274 4.093 Std. Error Mean 1.745 1.181 The difference in the two means is statistically significant, but how large is it? We can express the difference in terms of within-group standard deviations, that is, we can compute the statistic commonly referred to as Cohen’s d, but more appropriately referred to as Hedges’ g. Cohen’s d is a parameter. Hedges g is the statistic we use to estimate d. First we need to compute the pooled standard deviation. Convert the standard deviations to sums of squares by squaring each and then multiplying by (n-1). For Group 1, (5)4.2742 = 91.34. For Group 2, (11)4.0932 = 184.28. Now compute the SS1 SS2 91.34 184.28 pooled standard deviation this way: spooled 4.15 . n1 n2 2 16 Finally, simply standardize the difference in means: M M 2 22.67 13.25 g 1 2.27 , a very large effect. s pooled 4.15 An easier way to get the pooled standard deviation is to conduct an ANOVA relating the test variable to the grouping variable. Here is SPSS output from such an analysis: Copyright 2012, Karl L. Wuensch - All rights reserved. Strength_of_Effect.docx 2 ANOVA LOSS Between Groups W ithin Groups Total Sum of Squares 354.694 275.583 630.278 df 1 16 17 Mean Square 354.694 17.224 F 20.593 Sig. .000 Now you simply take the square root of the within groups mean square. That is, SQRT(17.224) = 4.15 = the pooled standard deviation. An easier way to get the value of g is to use one of my programs for placing a confidence interval around our estimate of d. See my document Confidence Intervals, Pooled and Separate Variances T. Here is an APA-style summary of the results: Persons who completed weight loss program 1 lost significantly more weight (M = 22.67, SD = 4.27, n = 6) than did those who completed weight loss program 2 (M = 13.25, SD = 4.09, n = 12), t(9.71) = 4.47, p = .001, g = 2.27. Do note that I used the separate variances t here – I had both unequal sample sizes and disparate sample variances. Also note that I reported the sample sizes, which are not obvious from the df when reporting a separate variances test. You should also recall that the difference in sample sizes here was cause for concern (indicating a problem with selective attrition). One alternative strength of effect estimate that can be used here is the squared point-biserial correlation coefficient, which will tell you what proportion of the variance in the test variable is explained by the grouping variable. One way to get that statistic is to take the pooled t and substitute in this formula: t2 4.5382 2 rpb 2 .56. An easier way to get that statistic to t n1 n2 2 4.5382 6 12 2 compute the r between the test scores and the numbers you used to code group membership. SPSS gave me this: Corre lations GROUP Pearson Correlation N LOSS -.750 18 When I square -.75, I get .56. Another way to get this statistic is to do a one-way ANOVA relating groups to the test variable. See the output from the ANOVA above. SSbetween 354.694 .56. Please note that 2 is the same The eta-squared statistic is SStotal 630.278 as the squared point-biserial correlation coefficient (when you have only two groups). When you use SAS to do ANOVA, you are given the 2 statistic with the standard output (SAS calls it R2). Here is an APA-style summary of the results with eta-squared. 3 Persons who completed weight loss program 1 lost significantly more weight (M = 22.67, SD = 4.27, n = 6) than did those who completed weight loss program 2 (M = 13.25, SD = 4.09, n = 12), t(9.71) = 4.47, p = .001, 2 = .56. One-Way Independent Samples ANOVA The most commonly employed strength of effect estimates here are 2 and 2 (consult your statistics text or my online notes on ANOVA to see how to compute 2). I have shown above how to compute 2 as a ratio of the treatment SS to the total SS. If you have done a trend analysis (polynomial contrasts), you should report not only the overall treatment 2 but also 2 for each trend (linear, quadratic, etc.) Consult the document One-Way Independent Samples ANOVA with SPSS for an example summary statement. Don’t forget to provide a table with group means and standard deviations. If you have made comparisons between pairs of means, it is a good idea to present d or 2 for each such comparison, although that is not commonly done. Look back at the document One-Way Independent Samples ANOVA with SPSS and see how I used a table to summarize the results of pairwise comparisons among means. One should also try to explain the pattern of pairwise results in text, like this (for a different experiment): “The REGWQ procedure was used to conduct pairwise comparisons holding familywise error at a maximum of .05 (see Table 2). The elevation in pulse rate when imagining infidelity was significantly greater for men than for women. Among men, the elevation in pulse rate when imagining sexual infidelity was significantly greater than when imagining emotional infidelity. All other pairwise comparisons fell short of statistical significance.” Correlation/Regression Analysis You will certainly have reported r or r2, and that is sufficient as regards strength of effect. Here is an example of how to report the results of a regression analysis, using the animal rights and misanthropy analysis from the document Correlation and Regression Analysis: SPSS: Support for animal rights (M = 2.38, SD = 0.54) was significantly correlated with misanthropy (M = 2.32, SD = 0.67), r = .22, animal rights = 1.97 + .175Misanthropy, n = 154, p =.006. Contingency Table Analysis Phi, Cramer’s phi (also known as Cramer’s V) and odds ratios are appropriate for estimating the strength of effect between categorical variables. Please consult the document Two-Dimensional Contingency Table Analysis with SPSS. For the analysis done there relating physical attractiveness of the plaintiff with verdict recommended by the juror, we could report: Guilty verdicts were significantly more likely when the plaintiff was physically attractive (76.7%) than when she was physically unattractive (54.2%), 2(1, N = 145) = 6.23, p = .013, C = .21, odds ratio = 2.8. Usually I would not report both C and an odds ratio. 4 Cramer’s phi is especially useful when the effect has more than one df. For example, for the Weight x Device crosstabulation discussed in the document TwoDimensional Contingency Table Analysis with SPSS, we cannot give a single odds ratio that captures the strength of the association between a person’s weight and the device (stairs or escalator) that person chooses to use, but we can use Cramer’s phi. If we make pairwise comparisons (a good idea), we can employ odds ratios with them. Here is an example of how to write up the results of the Weight x Device analysis: As shown in Table 1, shoppers’ choice of device was significantly affected by their weight, 2(2, N = 3,217) = 11.75, p = .003, C = .06. Pairwise comparisons between groups showed that persons of normal weight were significantly more likely to use the stairs than were obese persons, 2(1, N = 2,142) = 9.06, p = .003, odds ratio = 1.94, as were overweight persons, 2(1, N = 1,385) = 11.82, p = .001, odds ratio = 2.16, but the difference between overweight persons and normal weight persons fell short of statistical significance, 2(1, N = 2,907) = 1.03, p = .31, odds ratio = 1.12. Table 1 Percentage of Persons Using the Stairs by Weight Category Category Obese Percentage 7.7 Overweight 15.3 Normal 14.0 Of course, a Figure may look better than a table, for example, Percentage Figure 1. Percentage Use of Stairs Weight Category For an example of a plot used to illustrate a three-dimensional contingency table, see the document Two-Dimensional Contingency Table Analysis with SPSS. 5 Copyright 2012, Karl L. Wuensch - All rights reserved.