Experimental Design • If a process is in statistical control but has poor capability it will often be necessary to reduce variability. • Experimental design often offers a more effective way to accomplish this than control charting. • Control charting is a passive statistical method. We monitor the process and wait for information that may lead to a useful change. • Experimental design is an active statistical method. We perform a series of tests on a process making changes in the inputs and observing the corresponding effect of the outputs. • Application of experimental design in the product/process design stage is also important and can result in improved yield and reduced costs. Guidelines for Experimental Design • Identify the problem – Have a clear question you would like to answer – Suspect that there is something to be gained from experimentation – Must be feasible to take the process off-line for experimentation • Choose the factors (inputs) that will be varied in the experiment and their levels (values). – Put together a team of individuals who understand the process at different levels(engineers, managers, technicians, operators) – Identify the controllable and uncontrollable factors that may impact process performance. – Rank the controllable factors (usually just monitor environmental factors) – Determine the levels for each controllable factor chosen • • • • • • Formulate hypothesis Select the appropriate experimental design Conduct the experiments Analyze the results Draw conclusions Take action • Good experimental design should: – Eliminate known sources of bias – Guard against unknown sources of bias – Ensure that the experiment provides useful information about the process without using excessive experimental resources. Experiments with a single factor • We will begin by studying the most simple experimental design (an experiment with 1 factor) in detail. • The data analysis for single factor designs is similar to that of more complex experiments. • We will use analysis of variance (ANOVA) as the primary statistical tool for analyzing output from experiments An Example • A manufacturer of paper used for making grocery bags is interested in improving the tensile strength of the product. The manufacturing process specs currently call for 10% hardwood concentration. The paper has an average tensile strength of 15 psi. The process engineer and the operators suspect that tensile strength is a function of the pulp hardwood concentration in the paper. Economic considerations dictate that the range of possible hardwood concentrations range between 5% and 20%. The process engineer decides to investigate 4 levels of hardwood concentration: 5%, 10%, 15%, and 20%. She takes six specimens (replicates) at each level yielding 24 total specimens. Each specimen is then tested in random order. The results are as follows: Concentration (%) 5 10 15 20 1 2 3 Observations 4 5 6 7 12 14 19 8 17 18 25 15 13 19 22 11 18 17 23 9 19 16 18 10 15 18 20 Null and Alternative Hypothesis • Null Hypothesis: The mean response under each factor (treatment) is equal • Alternative Hypothesis: At least one of the treatment means is different. • ANOVA makes inferences about means from examination of the variability in the experiment. • The total variability can be measured by the total squared deviation of each response from the overall mean. This is termed Total Sum of Squares (TSS) • ANOVA then partitions TSS into two parts: Between treatment Sum of Squares (BSS) and Within treatment Sum of Squares (ESS). • SST = BSS + ESS • BSS represents the difference between factor level means and the grand average and so gives an indication of the differences between factor levels on the response. • Differences between observations within each factor level and the factor level mean are due to random error. Therefore the within treatment sum of squares is often termed ESS (Error sum of squares) • If the difference between BSS and ESS is large the null hypothesis will be rejected (The factor effects are significant) • Before comparing the sums of squares we must scale them by their degrees of freedom: • Let a = total number of factors n = total number of observations within each factor Then, there are an = N observations SST has N - 1 degrees of freedom Since there are a levels of the factor SSB has (a-1) degrees of freedom within each factor there are n observations (replicates) providing (n-1) degrees of freedom for estimation. Since there are a factors ESS has a(n-1) degrees of freedom. The ratio of a sum of squares to its number of degrees of freedom is termed a mean square. • Once the mean squares are calculated for the BSS and ESS, the ratio BSS/ESS provides a statistic that follows the F distribution. The critical value is then compared with the calculated value to determine the outcome of the hypothesis test. • The ANOVA table provides a convenient summary of this information: Source of Variation Between Factors Within Factors Total Sum of Squares SSB SSE SST Degrees of freedom a-1 a(n-1) An Mean Square F MSfactor MEerror Msfactor/MSerror Randomized Block Designs • Blocking involves grouping experimental units which have similar effects on the response variable. That is, we seek to eliminate the effect of extraneous factors within a block so that the between treatment effect (our main concern) can be more precisely measured. • Example: Suppose we wish to investigate the differences in raw materials from three different vendors. Processing will take place on two machines. If we randomly assign raw materials to machines we will not be able to claim that differences in output are due to vendors because we have not eliminated the effects of the machines. If we instead form two blocks consisting of the two machines differences in output from each block represent differences in vendors whereas differences in output between machines will represent differences due to blocking, I.e., whether blocking was successful. • With a blocking design, the variance is partitioned into 3 parts. Sum of squares due to treatment (SSTR), Sum of squares due to the blocking factor (SSBF), and SSE. Therefore, successful blocking minimizes variation between observations within a block while maximizing the variation between blocks. Since SSBF is eliminated from the experimental error the result is an increase in the precision of the experiment. • The ANOVA table for a randomized block design would be as follows: Source of Variation Treatments Sum of Squares Mean Square F SSTR Degrees of freedom a-1 MStreatment SSBF SSE SST b-1 (a-1)(b-1) ab-1 MSblock MSerror MStreatment/Mser ror MSblock/MSerror Blocks Error Total Example • We wish to test the effect of four different chemicals on the strength of a particular fabric. It is known that the effect of these chemicals varies considerably across fabric specimens. We take 5 fabric samples and randomly apply each chemical to each fabric. We have now isolated the effect of the chemical in the fairly homogeneous environment of a single sample of fabric. The results are as follows: Chemical 1 2 3 4 1 1.3 2.2 1.8 3.9 2 1.6 2.4 1.7 4.4 Fabric Sample 3 0.5 0.4 0.6 2 4 1.2 2 1.5 4.1 5 1.1 1.8 1.3 3.4