Chapter 3 - Experiments with a Single Factor: The Analysis of Variance (ANOVA) Example 3-1: Plasma Etching Experiment: Power and Etch Rate (pg. 62) Data File: Power-Etch (Ex 3-1).JMP Here are these data entered into JMP. You need two columns, one specifying the Power (W) and the other being the measured etch rate. Make sure Power, although numeric, is interpreted by JMP as a nominal/factor variable. To specify a numeric variable as a nominal/factor variable left-click on the © next Power (W) and change the variable type to nominal as shown below. Analysis in JMP ~ Fit Y by X approach Select Fit Y by X from the Analyze menu and place Power (W) in the X, Factor box and Etch Rate in the Y, Response box, then click OK. From the Oneway Analysis pulldown menu select the following options: Means/Anova Normal Quantile Plots (won't show much here with n = 5 reps) UnEqual Variances (tests whether the population variances are equal) Compare Means > All Pairs, Tukey HSD 1 The resulting output from the Oneway Analysis is shown below: ANOVA Table with Sum of Squares (see pg. 71) H o : 1 2 3 4 or Ho :1 2 3 4 0 H 1 : i j for some i j H 1 ; i 0 for some i p-value < .0001 Reject Ho , at least two mean etch rates differ for the power levels being examined. To determine which power levels significantly differ we need to use am appropriate multiple comparison procedure (see 3.5 pgs. 85-98). CI’s for Treatment Means and Differences in Treatment Means 100(1-)% CI for i (not corrected for experimentwise error rate) y i t / 2, N a SE ( y i ) where SE ( yi ) MS E 8.1695 here. n 100(1-)% CI for ( i j ) (not corrected for experimentwise error rate) ( yi y j ) t / 2, N a SE ( yi y j ) where SE ( yi y j ) 2MS E 11.5534 here. n CI’s for the Mean Etch Rates Under the Different Power Settings (see pg. 74) Model Parameter Estimates: ˆ ˆ1 ˆ2 ˆ3 ˆ4 2 Bonferroni Corrected CI’s If you are going construct r CI’s each at the 95% confidence level the probability that all r of them cover the population parameters of interest is (.95) r 1 r . The probability that least one of the CI’s misses is therefore 1 (.95) r < r . This probability is called the experimentwise error rate. To guarantee that the experimentwise error rate is no larger the specified is simply to use r instead of in the CI formulae on the previous page. This requires that we know in advance how many CI’s we are interested in constructing and that the number of planned (post-hoc) CI’s r is no too large. If r is large we will end up with CI’s that are too wide to be useful. When r is large it is better to use multiple comparison procedures more specific to types of post-hoc comparisons we are interested in doing (e.g. Scheffe’s, Tukey’s, Dunnett’s, etc...) Example: Suppose we wanted to compare all power settings to one another in the etching experiment using 95% confidence intervals for the differences in the 4 mean etch rates. This would require constructing = 6 confidence intervals. 2 So we should use .05 .00833 in place of in the CI’s for the differences r 6 the mean etch rates. Using t-Quantile Calculator we find t.00833/ 2,16 3.34 Comparing Population Variances The text on pgs. 80-82 covers two statistical tests used to assess the equality of the variances across treatment level, Bartlett’s and Levene’s tests. The p-values for both of these tests along with two others (O’Brien & Brown-Forsythe) are reported by JMP when selecting the UnEqual Variances option from the Oneway Analysis pull-down menu. Neither test here suggests inequality of the population variances as both have p-values >> .05. A warning is given because n = 5 replicates limits our ability to detect variance differences. All tests require that the response is normally distributed for each level. H O : 12 22 k2 H 1 : at least two differ 3 Multiple Comparison Procedures Multiple comparison procedures are designed to allow the analyst to make multiple mean comparisons while controlling the experimentwise error rate. Some of the most commonly used multiple comparison procedures are: Tukey’s, Hsu’s, Scheffe’s, and Dunnett’s. A brief description of these procedures is given below: Tukey’s - allows analysts to compare all possible pairs of means Hsu’s – allows the analyst to compare the “best” mean vs. rest. Scheffe’s – allows the analyst to compare all possible linear combinations of means, e.g. the means 1 and 2 contrasted vs. means 3 and 4. Dunnett’s – allows the analyst to compare all means back to a control group or setting. These are all available in some form in JMP, however we will focus our attention on Tukey’s procedure, because of these procedures it is the most widely used. To obtain the results of Tukey’s procedure select Oneway Analysis > Compare Means > All Pairs, Tukey’s HSD (Honest Significant Difference). The results for the etch rate experiment are shown below. Here we see that all power levels significantly differ from each other in terms of the mean etch rate. None of the power settings are connected by the same letter suggesting all are different. It is important to quantify significant differences or effects through the use of CI’s for the pairwise differences in the mean etch rates under the different power settings. The largest difference in etch rate is between the highest and lowest power settings. We estimate the etch rate when the power is set at 220 watts is between 122.75 and 188.85 units larger than the etch rate when the power is 160 watts. Other differences are interpreted in a similar fashion. No of the CI’s contain 0, which supports are conclusion from above, that all power settings result in significantly different mean etch rates. The experimentwise error rate for all comparisons and CI’s is . 4 The Fit Model Approach to Performing a Single Factor ANOVA in JMP As an alternative to the Fit Y by X approach used above we can also use the Analyze > Fit Model option in JMP. There are some advantages to using this approach: It allows for analyzing experiments with more than one factor Allows for checking assumptions using residuals (normality and constant error variance) Allows for checking case diagnostics like outlier-T, Cook’s distance, etc. Gives the estimated parameters for the fitted model. Put the response of interest in the Y box and all factors in the Construct Model Effects box. We will eventually put multiple factors and their interactions in this box. The important aspects of the resulting output are show below ANOVA table gives the usual stuff The parameter estimates are different than those given earlier. Here we have constrained ˆ4 0 so ˆ 617.75 represents the mean for the highest power setting (220 W) rather than the overall mean. This parameterization is required for mathematical reasons which we will not get into at this point. The effect tests section gives the p-value associated with each of the effects in the model. For a single factor experiment it is the same the results from the ANOVA table. The residual by predicted plot shows the residuals ( eij ) plotted vs. the fitted values ( ŷ ij ) ~ i.e. the treatment sample means ( yi ). We want to see no discernible pattern in this plot. If there is fanning out, it can indicated that the variation in the response is not the same across the different treatment levels. 5 To assess the normality of the errors we need to save the residuals or the standardized residuals to the spreadsheet and use Analyze > Distribution to examine their distribution. The distribution of the residuals is slightly kurtotic (S-shape), but not enough to be concerned about. To obtain the results of Tukey’s multiple comparison procedure select that option from the Power (W) pull-down menu to the right of the main section of output. The resulting output includes the connected means report (letter A,B,... one) and you can get CI’s for the pairwise differences in an ordered difference report as in the Fit Y by X case. The matrix below conveys all the same information. 6 Single Factor Experiments in Design-Expert 6.0 First select File > New Design and then choose General Factorial from the Factorial tab. This will open a window where you asked to specify the number of categorical factors you wish to use, which for the etching experiment is one. Next you specify the number of levels for the factor, the name of the factor, the units for factor, and the settings as shown below. Click Continue when you are finished. Next you are asked to enter the number of replicates to use, which in this case is n = 5. Blocking can be specified if necessary. Information about the response(s) is entered next. 7 Data Entry Spreadsheet from Design-Expert Design-Expert will randomize the run orders by default. This is great if we actually going to run the experiment ourselves, but as we already have the data from this experiment it is easier to list the runs in what is referred to as the Standard Order vs. Run Order. To do this click on View > Std Order to change the spreadsheet to standard ordering. We are now ready to conduct our analysis of the data. Begin by clicking the Analysis icon in the menu tree on the left. Design-Expert pretty much leads you by the nose at this point. Click on the name of the response (Etch Rate) to analyze it. In experiments where there are more than one response need to analyze each separately. 8 Choice of Response Transformation (Transform) You are first asked to choose a transformation for the response. Unless you have a prior knowledge about the response that suggests the need for a transformation, we begin our analysis without one. Selecting Model Effects (Effects) Here we choose which effects wewish to examine in the model. The default chosen by Design-Expert is generally recommended, but when we get into multi-factor experiments we may decide to eliminate some effects from the model before proceeding with our analysis. The significance of the chosen effects is also reported here, which can help with effect selection. 9 Analysis of Variance (ANOVA) The ANOVA tab gives a full statistical report for the model we have fit. 10 Case Diagnostics Examining case diagnostics is generally more important we more complicated designs or when regression is being used to predict the response under different experimental conditions. The only case statistic of interest to use at this point would be the Outlier-t which is used to flag outliers. Outliers will have * next to them. Diagnostic Plots Normal Probability (normal probability/quantile plot of the residuals) * Residuals vs. Predicted Values * Residuals vs. Run Order (only if we actually did the experiment in the order) * Residuals vs. Factor (only need to look at this in multi-factor experiments) Outlier T (visualization of the outlier-t statistic results) * Cook’s Distance (visualization of Cook’s influence measure) Leverage (visualization of the potential for influence) Predicted vs. Actual (visualization of how well the model fits the data) Box-Cox (used to find a transformation to improve normality of the response) * In a single factor experiment we really only need to look at the plots denoted by a (*) in the above list. Also we only need to look at the Box-Cox transformation if we see problems, non-normality or heteroscedasticity, in the residual plots. Some these plots are shown below. 11 There is no graphical evidence of the need to transform from either the normal probability plot of the residuals or from residuals plotted vs. the predicted values. If there had been evidence of assumption violations from these plots we would generally examine the results of the Box-Cox procedure to see a suitable transformation can be found. No transformation is recommended here, which is what we expected given the residual diagnostic plots above. Model Graphs For a single factor experiment the standard comparative display is used to visualize results. The raw data is denoted by the red dots and the sample means for each factor setting along with CI’s are also shown in black. We will discuss the Optimization section later when we get into multi-factor experiments and response surfaces. 12 13