1 Creating a File in SPSS Create an Excel file that looks like the one below. Save the Excel program to your computer. Now, open up SPSS. SPSS will ask you if you want to open an existing program. Click OK. Richard S. Balkin, Ph.D., 2004 2 Now, use the top bar that says, Look in: to find where your data is located. Make sure Files of type: says Excel. Richard S. Balkin, Ph.D., 2004 3 I called my Excel program anova example. Open it by double-clicking (DC). Richard S. Balkin, Ph.D., 2004 4 You will see this box: Click OK Richard S. Balkin, Ph.D., 2004 5 Once you have opened the file, it should look like this: If you have not already done so, click Edit-Options-Viewer and then click the box that says “Display commands in the log.” Now you are ready to run a one-way ANOVA. The first thing we want to do is analyze our model assumptions. Consider what you would look for to evaluate independence. Richard S. Balkin, Ph.D., 2004 6 One-way ANOVA To assess for normality, we will use the Explore command on SPSS. Richard S. Balkin, Ph.D., 2004 7 You want your dependent variable (DV) to be normally distributed for each level of your independent variable (IV). Click once on to Group and select the arrow next to Factor List: Now, click once on to Score and select the arrow next to Dependent List: The next step is to select the bottom button labeled Plots… Richard S. Balkin, Ph.D., 2004 8 Click on the box in the middle that says Normality plots with tests Click Continue. Then click OK Richard S. Balkin, Ph.D., 2004 9 An output page similar to the one below will pop up. On this page is information related to normality, including the Shapiro-Wilk statistic, skewness, and kurtosis for each group if you scroll down. Richard S. Balkin, Ph.D., 2004 10 You can divide the skewness by the standard error. If your value is less than 2.576 (.01 level of significance) or 3.29 (.001 level of significance), then your distribution is approximately normal. Preferably, we will use the .01 level of significance in this class. We still have to check our HOV assumption, which we will do when we run the ANOVA. Richard S. Balkin, Ph.D., 2004 11 Once again, click Analyze, and then General Linear Model, then Univariate Richard S. Balkin, Ph.D., 2004 12 Click once on to Group and select the arrow next to Fixed Factor(s): Now, click once on to Score and select the arrow next to Dependent Variable: Next, click Options… Richard S. Balkin, Ph.D., 2004 13 Click the boxes beside Descriptive statistics and Homogeneity tests—this will tell you whether or not you had unequal variances (HOV). Click Continue, then click OK on the next screen. Richard S. Balkin, Ph.D., 2004 14 Your output will include descriptive statistics, the Levene statistic for testing Homogeneity of Variances, and your ANOVA results. Richard S. Balkin, Ph.D., 2004 15 Now you are ready to run post hoc analyses. Richard S. Balkin, Ph.D., 2004 16 Post Hoc Analysis To run a Tukey post hoc analysis to identify exactly where significant differences may be, go back to your data and click Analyze, and then General Linear Model, then Univariate Richard S. Balkin, Ph.D., 2004 17 Now click Post Hoc… Richard S. Balkin, Ph.D., 2004 18 Click once on Group and click on the arrow so that it moves underneath Post Hoc Tests for: Click the box next to Tukey. Then click Continue. Then click OK on the next screen. Richard S. Balkin, Ph.D., 2004 19 Notice your results. An * denotes significant differences between groups at the .05 level of significance. However, if you have not set your alpha level to be different in SPSS, the * will not be accurate. So, as a rule of thumb, ignore the *. Regardless of the *, always compare the p-value to the alpha level you have set for your study. Richard S. Balkin, Ph.D., 2004 20 Contrast Comparisons Now, we will take a look at running contrast comparisons. For the purposes of this exercise, we will run two contrasts: Compare groups 1 and 3 to group 2. Compare group 2 to group 3. Click Analyze, and then Compare Means, then One-Way ANOVA… Richard S. Balkin, Ph.D., 2004 21 Your screen will look like the one below: Click once on Score and click the arrow next to Dependent List: Then click once on Group and click the arrow next to Factor: Then click Contrasts… Richard S. Balkin, Ph.D., 2004 22 In comparing groups 1 and 3 to group 2, move the cursor to the box that says Coefficients: and type .5 Then click Add Richard S. Balkin, Ph.D., 2004 23 Now in the box that says Coefficients: type -1 and click Add Richard S. Balkin, Ph.D., 2004 24 Now in the box that says Coefficients: type .5 and click Add Richard S. Balkin, Ph.D., 2004 25 Now in the box that says Coefficients: type 0 and click Add Richard S. Balkin, Ph.D., 2004 26 Now click Next and repeat this process using the proper coefficients for the second contrast comparison: 0, 1, -1 and 0 Richard S. Balkin, Ph.D., 2004 27 Your screen should look like the one below: Now click Continue And click OK on the next screen Richard S. Balkin, Ph.D., 2004 28 You will see a results page like the one below: Richard S. Balkin, Ph.D., 2004 29 Richard S. Balkin, Ph.D., 2004 30 Transformations For the purposes of explaining transformations, some changes in the previous data set have been made for group four. Richard S. Balkin, Ph.D., 2004 31 Notice that when the Shapiro-Wilk statistic is evaluated, group four is nonnormal. Shapiro-Wilk for group four is less than .01 and therefore significantly deviates from a normal distribution. Richard S. Balkin, Ph.D., 2004 32 To perform a transformation, decide which procedure would be appropriate to use. For this example, we will try a logarithmic transformation. Next click Transform then Compute. Richard S. Balkin, Ph.D., 2004 33 Create a name for your transformed variable. For this example, the variable name will be trans_score. Then click the box under numeric expression. Richard S. Balkin, Ph.D., 2004 34 Using the downward arrow under functions, click on LN (natural log). Richard S. Balkin, Ph.D., 2004 35 Now click the upward arrow next to Functions. Your screen should look like the one below. Richard S. Balkin, Ph.D., 2004 36 Now double-click on Score and click OK Richard S. Balkin, Ph.D., 2004 37 A new variable called tran_score appears in your data set. Richard S. Balkin, Ph.D., 2004 38 Now, run your normality statistics again. But, in place of score for your dependent variable, use tran_score. Richard S. Balkin, Ph.D., 2004 39 Notice that all of the groups meet the assumption for normality. Richard S. Balkin, Ph.D., 2004 40 Factorial ANOVA We will use Lesson 25 Data file 1 on the CD in the Green & Salkind textbook. As in the one-way ANOVA, you will need to analyze your model assumptions. Review pp. 6-15. In running a factorial ANOVA, you will put in 2 independent variables instead of 1. For the IV, Gender and Method were used; gpaimpr is our DV. Richard S. Balkin, Ph.D., 2004 41 As before, click Analyze, and then General Linear Model, then Univariate Richard S. Balkin, Ph.D., 2004 42 Click once on to Gender and select the arrow next to Factor List: Click once on to method and select the arrow next to Factor List: These are your independent variables. Note that gender has 2 levels (male and female). method has 3 levels. This would be a 2 by 3 (2 X 3) factorial ANOVA Click once on to gpaimpr and select the arrow next to Dependent Variable: Richard S. Balkin, Ph.D., 2004 43 After you evaluate your data and analyze your model assumptions, you have to determine if there is an interaction effect. First, select Options… Richard S. Balkin, Ph.D., 2004 44 Your screen should look like the one below: Click on OVERALL and then click the arrow. It will move your selection to Display Means for: Repeat this process for gender, method, gender*method. Richard S. Balkin, Ph.D., 2004 45 Your screen should look like the one below. Also, click the boxes next to Descriptive statistics, Homogeneity tests, Estimates of effect size. Click Continue On the next screen click OK Richard S. Balkin, Ph.D., 2004 46 Notice on your output that you have a non-significant interaction, as evidenced by the results of gender*method: F (2, 54) = 2.921, p = .062. In this particular case, you would report the main effects of each independent variable and conduct a Tukey post hoc analysis on any IV with 3 or more levels. So, we would need to do a Tukey for method. Follow the steps on pp. 16-19. Richard S. Balkin, Ph.D., 2004 47 For an example of using SPSS when there is a significant interaction, we will use lesson 25 data file 2 on the CD in Green & Salkind. Follow the same steps outlined on pp. 31-35. Richard S. Balkin, Ph.D., 2004 48 This time, however, notice that you have a significant interaction, as evidenced by the results of gender*method: F (2, 54) = 10.543, p < .0001. Richard S. Balkin, Ph.D., 2004 49 At this point in time, you want to plot the interaction. You can do this by charting the means from the output below: Richard S. Balkin, Ph.D., 2004 50 You can also ask SPSS to plot the interaction. Click Analyze, and then General Linear Model, then Univariate. Then click Plots… Richard S. Balkin, Ph.D., 2004 51 Click once on method and then click the arrow next to Horizontal Axis. Click once on Sex and then click the arrow next to Separate lines: Then click Add Richard S. Balkin, Ph.D., 2004 52 Your screen should look like the one below. The numbers on the vertical axis represent scores for gpaimpr. You can see by the chart that males and females perform differently across note-taking methods. Richard S. Balkin, Ph.D., 2004 53 Since we have a significant interaction, we need to test for simple effects by sorting our data and running an analysis for each level of an IV. This can be done in one of two ways. In the first method, we will use SPSS coding. SPSS can perform the proceeding steps using the following coding formula in SPSS Syntax: UNIANOVA DV BY IV1 IV2 /EMMEANS = TABLES (IV1 * IV2) COMP (IV2) Click Analyze, and then General Linear Model, then Univariate. Then click Paste… Richard S. Balkin, Ph.D., 2004 54 Now type your SPSS code using the formula above, highlight it, and click the run button. The univariate tests below with the post hoc tests above show your results for the simple effects. Richard S. Balkin, Ph.D., 2004 55 The next method is more involved, but it replicates the results of the first method, and it also shows you how the results were derived. In other words, it replicates the SPSS code. First, click Data and Select Cases… Richard S. Balkin, Ph.D., 2004 56 Click the second selection If the condition is satisfied. Then click If… Richard S. Balkin, Ph.D., 2004 57 Click once on gender and then click the arrow. Richard S. Balkin, Ph.D., 2004 58 On the key pad (as shown below) click = 1. Click Continue. Then on the next screen click OK. Richard S. Balkin, Ph.D., 2004 59 Now, SPSS will only select cases in which gender = 1 (males only). Now, run your ANOVA following steps on pp.11-19. You will need to consult your lecture notes to interpret the output. Follow steps on pp. 53-57 to run the analysis for females (sex = 2). Richard S. Balkin, Ph.D., 2004 60 Repeated Measures We will be using Lesson 28 data file 1 from the CD in Green & Salkind. Richard S. Balkin, Ph.D., 2004 61 Click Analyze, and then General Linear Model, then Repeated Measures… Richard S. Balkin, Ph.D., 2004 62 In this particular example, we will test to see if there are significant differences among test scores on the Desire to Express Worry Scale after 0, 5, 10, and 15 years of marriage. You will see this screen below: Richard S. Balkin, Ph.D., 2004 63 Where it says factor1 by the box next to Within Subject Factor Name: erase and type Time. Then type 4 in the box next to Number of Levels: Now click Add Now click Define Richard S. Balkin, Ph.D., 2004 64 Your box should look like the one below: Richard S. Balkin, Ph.D., 2004 65 Click time1 and then the arrow for Within Subject Variables. Repeat this step for time2, time3, and time4. Click Options Richard S. Balkin, Ph.D., 2004 66 Click once on OVERALL and then click the arrow to move the selection to Display Means for. Repeat this step for time Click the box next to Descriptive Statistics and then click Continue Richard S. Balkin, Ph.D., 2004 67 On the next screen click OK. Your output should look like the following: Review your course notes for interpretation. Richard S. Balkin, Ph.D., 2004 68 To do a post hoc analysis, click Analyze, Compare Means, and then Paired Samples T-Test. Richard S. Balkin, Ph.D., 2004 69 Click one pair times to compare. In this example, I am comparing time1 to time2. Richard S. Balkin, Ph.D., 2004 70 Click the arrow beside Paired Variables: Richard S. Balkin, Ph.D., 2004 71 Now repeat this process for all possible comparisons. Then click OK Richard S. Balkin, Ph.D., 2004 72 Use your notes to interpret the results. Make sure you do a Bonferroni adjustment. Richard S. Balkin, Ph.D., 2004 73 Multiple Regression We will be using Lesson 33 Exercise file 1 on the CD of Green & Salkind. We want to examine the extent of the relationship between stateexam (criterion variable) from mathtest and engtest (predictor variables). Richard S. Balkin, Ph.D., 2004 74 Click Analyze, Regression, Linear Richard S. Balkin, Ph.D., 2004 75 Your screen should show the following: Richard S. Balkin, Ph.D., 2004 76 Click once on math aptitude and then click the arrow by Independent(s): Repeat this for English aptitude. Richard S. Balkin, Ph.D., 2004 77 Then click on Average percentage and click the arrow next to Dependent: Richard S. Balkin, Ph.D., 2004 78 Click Statistics and click the boxes next to Descriptives, Part and partial correlations, and Collinearity diagnostics. Estimates and Model fit should already be selected. Richard S. Balkin, Ph.D., 2004 79 Click Continue. On the next screen click OK. Refer to your notes for interpretation of output. Richard S. Balkin, Ph.D., 2004 80 Evaluating model assumptions in multiple regression From the Linear Regression screen click Plots Richard S. Balkin, Ph.D., 2004 81 Insert the standardized predicted (ZPRED) and residual (ZRES) values in the X & Y axes. We will use this graph to determine the linearity and homoscedascity assumptions. Richard S. Balkin, Ph.D., 2004 82 Now click on Save Richard S. Balkin, Ph.D., 2004 83 Now click Standardized Residuals. This will create a new column in our data set. We will use this variable to examine the normality of our error variances. Now click Continue and the OK. Richard S. Balkin, Ph.D., 2004 84 Notice you have a new column for your standardized residuals (ZRE_1). Now click Analyze—Descriptive Statistics--Explore. Richard S. Balkin, Ph.D., 2004 85 We will now run normality tests for the criterion variable, which must be normally distributed, and for our error variances, which should also be normally distributed. Click on the criterion variable (which in this case is the average percentage on the exam) and the standardized residuals and place them in the Dependent List. Then click Plots. Richard S. Balkin, Ph.D., 2004 86 Because multiple regression is a large sample procedure, we will focus on the appearance of the box plots to determine if the criterion variable and error variances are normally distributed. Be sure to click Normality plots with tests and then Continue followed by OK. This concludes the procedures for analyzing model assumptions. The sample output will be helpful in examining the procdures conducted. Richard S. Balkin, Ph.D., 2004