Chapter 2 - People Server at UNCW

advertisement
Design & Analysis of Experiments 8E 2012 Montgomery
STT 511-STT411:
DESIGN OF EXPERIMENTS AND
ANALYSIS OF VARIANCE
Dr. Cuixian Chen
Chapter 2
Chapter 2: Some Basic Statistical Concepts
1
Review of STT215: Chapter 3
3.1 Design Of Experiments
(Outline of a randomized designs)
3
Completely randomized experimental designs: Individuals are randomly
assigned to groups, then the groups are randomly assigned to treatments.
Example 3.13, page 179
What are the effects of repeated exposure to an advertising message (digital
camera)? The answer may depend on the length of the ad and on how often it is
4
repeated. Outline the design of this experiment with the following information.



Subjects: 150 Undergraduate students.
Two Factors: length of the commercial (30 seconds and 90 seconds – 2 levels) and
repeat times (1, 3, or 5 times – 3 levels)
Response variables: their recall of the ad, their attitude toward the camera, and
their intention to purchase it. (see page 187 for the diagram.)
HWQ: 3.18,
3.30(b),3.32
3.1 Design Of Experiments (Block designs)
In a block, or stratified, design, subjects are divided into groups, or blocks,
prior to experiments to test hypotheses about differences between the groups.
5
The blocking, or stratification, here is by gender (blocking factor).
EX3.19
Ex: 3.17 (p182), 3.18
HWQ: 3.47(a,b), 3.126.
3.1 Design Of Experiments (Matched pairs designs)
6 Matched
pairs: Choose pairs of subjects that are closely matched—e.g., same sex,
height, weight, age, and race. Within each pair, randomly assign who will receive
which treatment.
It is also possible to just use a single person, and give the two treatments to this
person over time in random order. In this case, the “matched pair” is just the same
person at different points in time.
The most closely
matched pair
studies use
identical twins.
HWQ 3.120
STT511-411: Chapter 2 – Some Basic
Statistics Concepts
Design of Engineering Experiments
Chapter 2 – Some Basic Statistical Concepts
8

Describing sample data






Random samples
Sample mean, variance, standard deviation
Populations versus samples
Population mean, variance, standard deviation
Estimating parameters
Simple comparative experiments



The hypothesis testing framework
The two-sample t-test
Checking assumptions, validity
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Review of one sample inference in Stt215
9

Estimation of Parameters
1 n
y   yi estimates the population mean 
n i 1
1 n
S 
( yi  y ) 2 estimates the variance  2

n  1 i 1
2

Sampling Distribution
y
Z 
~ N (0,1), if σ if is known.
/ n
y
t
~ t (df  n  1), if σ if is unknown.
s/ n
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Normal and T distribution
10
When n is very large, s is a very good estimate of  and the
corresponding t distributions are very close to the normal distribution.
The t distributions become wider for smaller sample sizes, reflecting the
lack of precision in estimating  from s.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
How can we use computer to help us understand
the distributions?
11

Introducing R.
1.
Where to find? Google “R”->the first link.
2.
Download and install CRAN package.
3.
Then we have
How do we use it?

1.
Assign value to a variable x: x=3; or x<-3;
2.
A sequence of numbers: 1:5; or 6:3;
3.
A vector: x=c(4,5,6); or x=4:6; , then x[2]=5.
4.
loop: for (i in 1:5) {print(i)};
5.
Average: mean(x);
6.
sum: sum(x);
Entry level of R in 10 mins.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Normal distribution in R
12

normal dist in R: dnorm(x, µ, σ), for density;

pnorm(x, µ, σ), for left tail probability: Pr(X<=x);

qnorm(per, µ, σ), for the quantile: given Pr(X<=x)=per and find x;

rnorm(N, µ, σ), for the random number generation.
Eg: use R to find the probabilities for N(µ=3, σ =4)
1.
P(x<4);
2.
P(x>2);
3.
P(1<X<4) ;
4.
95th percentile
Note: if X ~ N( µ,σ2) , then Z=(X-µ)/σ ~ N(0,1), which is standard normal
distribution.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
T distribution in R
13

T-distribution in R:

dt(x,df), for density;

pt(x,df), for left tail probability: Pr(X<=x);

qt(per,df), for the quantile: given Pr(X<=x)=per and find x;

rt(N,df), for the random number generation.
Eg: use R to find the following probabilities for t(df=6)
P(x<4);
1.
P(x>2);
2.
P(1<X<4)
3.
95th percentile.
4.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
One sample confidence interval and
hypothesis testing
Review: Confidence levels
Confidence intervals contain the population mean  in C% of samples.
Different areas under the curve give different confidence levels C.
z*:
z* is related to the chosen
confidence level C.

C
C is the area under the standard
normal curve between −z* and z*.

The one sample Z-confidence
interval is thus:
x  (z*) 
n
−z*
z*
Example: For an 80% confidence
level C, 80% of the normal curve’s
area is contained in the interval.
Review: 5 Steps for Hypothesis testing
16
1.
State H0 and Ha
2.
State the level of significance (Usually α is 5% ).
3.
Calculate the test statistic, ASSUMING THE NULL HYPOTHESIS IS TRUE
4.
5.
Find the P-value, that is the probability (assuming H0 is true) that the
test statistic would take a value as extreme as or more extreme than
the actually observed (in the direction of Ha).
Draw Conclusion:
If P-value ≤ α, then we reject H0 (Enough evidence).
If P-value >α, then we do not reject H0 (No Enough evidence).
Note: The two possible conclusions are rejecting or not rejecting H0.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
P-value in one-sided and two-sided tests
One-sided (onetailed) test
Two-sided (twotailed) test
To calculate the P-value for a two-sided test, use the symmetry of the normal
curve. Find the P-value for a one-sided test, and double it.
17
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Review: Find P-value
The P-value is the area under the sampling distribution for values
at least as extreme, in the direction of Ha, as that of our random
sample.
Use R we just learn to find the p-value:
e.g. H0 : µ = 2.6 hours verse Ha : µ < 2.6 hours gives test statistic
Z=-1.6. Q: Find the p-value.
Sampling
distribution
σ/√n
x
µ
defined by H0
Example 1: One-sample Z-test
A test of the null hypothesis H0 : µ = µ0 gives test statistic Z=-1.6
a) What is the P-value if the alternative is Ha : µ > µ0 ?
b) What is the P-value if the alternative is Ha : µ < µ0 ?
c) What is the P-value if the alternative is Ha : µ ≠ µ0 ?
Example 1 (cont.): One-sample Z-test
A test of the null hypothesis H0 : µ = µ0 gives test statistic Z=2.1
a) What is the P-value if the alternative is Ha : µ > µ0 ?
b) What is the P-value if the alternative is Ha : µ < µ0 ?
c) What is the P-value if the alternative is Ha : µ ≠ µ0 ?
Example 2: One sample Z-test or
One sample Z-Confidence Interval
21



The National Center for Health Statistics reports that the mean
systolic blood pressure for males 35 to 44 years of age is 128
with a population SD=15. The medical director of a company
looks at the medical records of 72 company executives in this
age group and finds that the mean systolic blood pressure in this
sample is 126.07.
1) Is this evidence that executives blood pressures are different
from the national average?
2) Find the 95% confidence interval for the average SBP of all
company executives.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Example 2: One sample Z-test or
One sample Z-Confidence Interval
22
Answer: Hypothesis: H0 : µ = 128 v.s. Ha : µ≠128.
Test statistics
x  126.07
z
  15
α = 5%
n  72
x   126.07  128

 1.09
 n
15 72
P-value= 2*pnorm(-1.09)= 0.2757131. Conclusions: ….

The 95% confidence interval for the average SBP of all company
executives is: Z*=qnorm(0.975)
15
15

 


(
126
.
07

1
.
96
,
126
.
07

1
.
96
)  (122.61, 129.53)
, X  1.96
 X  1.96

72
72
n
n


The conclusions from two-sided HT (α=5%) and CI (95%) are
consistent
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Example 3: One sample t-test or
One sample t-Confidence Interval
23



The National Center for Health Statistics reports that the mean
systolic blood pressure for males 35 to 44 years of age is 128.
The medical director of a company looks at the medical records
of 72 company executives in this age group and finds that the
mean systolic blood pressure in this sample is 126.07 with sample
SD 15.
1) Is this evidence that executives blood pressures are different
from the national average?
2) Find the 95% confidence interval for the average SBP of all
company executives.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Example 3: One sample t-test or
One sample t-Confidence Interval
24
Answer: Hypothesis: H0 : µ = 128 v.s. Ha : µ ≠ 128.
α = 5%
Test statistics X  126.07 s = 15 n = 72
t
X   126.07  128

 1.09; df = 72 - 1 = 71
S/ n
15 / 72
P-value= 2*pt(-1.09, df=71)=0.2793988. Conclusions: ….
The 95% confidence interval for the average SBP of all company
executives is
t*=qt(0.975, 71)
( x  t ( n1)
*
S
S
s
s
*
, x  t ( n1) )  ( X  1.99 , X  1.99 )
n
n
n
n
 (126.07  1.99
15
15
,126.07  1.99
)  (122.55, 129.59)
72
72
The conclusions from two-sided HT (α=5%) and CI (95%) are consistent
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
One sample test
(Shall we use Z-test or T-test ??)
25
Example 4: A new medicine treating cancer was introduced to the
market decades ago and the company claimed that on average
it will prolong a patient’s life for 5.2 years. Suppose the SD of all
cancer patients is 2.52. In a 10 years study with 64 patients, the
average prolonged lifetime is 4.6 years.


1) With normality assumption, do the 10-year study’s data show
a different average prolonged lifetime?
2) Find the 95% confidence interval for the average prolonged
lifetime for all patients.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
One sample test
(Shall we use Z-test or T-test ??)
26
Example 5: A new medicine treating cancer was introduced to the
market decades ago and the company claimed that on average
it will prolong a patient’s life for 5.2 years. In a 10 years study
with 20 patients, the average prolonged lifetime is 4.7 years with
sample SD 2.50.


1) With normality assumption, do the 10-year study’s data show
a different average prolonged lifetime?
2) Find the 95% confidence interval for the average prolonged
lifetime for all patients.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Review: Link between confidence level and margin of
error for one-sample z-CI:
The margin of error depends on z.
MOE  z  
Higher confidence C implies a larger
margin of error m (thus less precision
in our estimates).
C
A lower confidence level C produces a
smaller margin of error m (thus better
precision in our estimates).
m
−z*
m
z*
n
Example 6: Finding sample size


In a clinical study with certain # of patients, a new medicine
can on average prolong 4 years of life. Suppose the SD of all
cancer patients is 0.75.
Q: How large a sample pf cancer patients would be needed
to estimate the mean within ±0.1 years with 90% confidence?
Z*=1.645; MOE=(1.645)*(0.75)/sqrt(n)=0.1; so n=(1.645*0.75/0.1)^2=152.2139.
We will take n=153.
Confidence intervals to test hypotheses
For a level a two-sided significance test:
Rejects H0:  = 0 exactly when the hypothesized value 0 falls
outside a level (1-a)100%confidence interval for  .
In a two-sided test,
C = 1 – α.
C confidence level
α significance level
α /2
α /2
One sample test in R
30

One-sample t-test;

One-sample Z-test

One-sample Z-CI’s and One-sample t-CI’s .
For H0: mu=32, v.s. Ha: mu<32
## suppose that these are the houses prices of 5
randomly selected house from Wilmington
x<-c(20,25,28,33,37);
mean(x)
var(x)
sd(x)
##
## We want to test the average house price is less than 32
## one sample t test by hand in R ###########
## For H0: mu=32, v.s. Ha: mu<32
n<-length(x)
t.val<-(mean(x)-32)/(sd(x)/sqrt(n))
df.dat<-n-1;
p.value<-pt(t.val,df.dat);
print(p.value);
## one sample test by t.test in R
## For H0: mu=32, v.s. Ha: mu<32
t.test(x,alternative="less",mu=32)
##############################
## For H0: mu=32, v.s. Ha: mu>32
t.test(x,alternative="greater",mu=32)
## For H0: mu=32, v.s. Ha: mu≠32
t.test(x,mu=32) #for Ha: mu not equal to 32
## 95% CI in R ##########
LB<-mean(x)-qt(.975,4)*sd(x)/sqrt(n);
UB<- mean(x)+qt(.975,4)*sd(x)/sqrt(n);
print(c(LB, UB));
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
One sample test in SAS
31
title 'One-Sample t Test';
proc ttest data=value h0=32 sides=2; /*default*/
data value;
/* 3 options for sides 2=not equal */
input value @@;
/* L =less than U=greater than*/
var value;
datalines;
run;
20
25
proc ttest data=value h0=32 sides=U;
28
var value;
33
run;
37
;
proc ttest data=value h0=32 sides=L;
run;
var value;
proc ttest data=value h0=32;
run;
var value;
run;
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Matched-pair sample (one-sample) confidence
interval and hypothesis testing
Matched pairs t procedures
for dependent sample
Subjects are matched in “pairs” and

outcomes are compared within each unit

Example: Pre-test and post-test studies look at data collected on
the same sample elements before and after some experiment is
performed.

Example: Twin studies often try to sort out the influence of genetic
factors by comparing a variable between sets of twins.
We
perform hypothesis testing on the difference in each unit
Matched pairs (one-sample)
The variable studied becomes Xdifference = (X1 − X2). The null
hypothesis of NO difference between the two paired groups.
H0: µdifference= 0 ; Ha: µdifference>0 (or <0, or ≠0)
When stating the alternative, be careful how you are calculating
the difference (after – before or before – after).
Conceptually, this is not different from tests on one population.
Matched Pairs (one-sample)



If we take After – Before, and we want to show that the “After
group” has increased over the “Before group”
Ha:  > 0
“After group” has decreased
xdiff   diff
t diff 
Ha:  < 0
sdiff n
The two groups are different
Ha:  ≠0
Example 4: Matched Pairs t-test
Many people believe that the moon influences the actions of some individuals. A
study of dementia patients in nursing homes recorded various types of
disruptive behaviors every day for 12 weeks. Days were classified as moon
days and other days. For each patient the average number of disruptive
behaviors was computed for moon days and for other days. The data for 5
subjects whose behavior were classified as aggressive are presented as
below:
Moon days
Other days
3.33
0.27
3.67
0.59
2.67
0.32
3.33
0.19
3.33
1.26
We want to test whether there is any difference in aggressive behavior on
moon days and other days.
Example 4: Matched Pairs t-test
Many people believe that the moon influences the actions of some individuals. A
study of dementia patients in nursing homes recorded various types of
disruptive behaviors every day for 12 weeks. Days were classified as moon
days and other days. For each patient the average number of disruptive
behaviors was computed for moon days and for other days. The data for 5
subjects whose behavior were classified as aggressive are presented as
below:
Moon days
Other days
Difference
3.33
0.27
3.06
3.67
0.59
3.08
2.67
0.32
2.35
3.33
0.19
3.14
3.33
1.26
2.07
We want to test whether there is any difference in aggressive behavior on
moon days and other days.
Answer to Example 4
38
Let difference = aggressive behavior on moon days and other days.





H 0 : d  0 verses
H a : d  0
,
a  0.05
t-statistic=12.377, df=5-1=4,
p-value=2.449*10^(-4).
Reject H0 at 5% level.
Enough evidence to conclude that there is any difference in
aggressive behavior on moon days and other days
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Does lack of caffeine increase depression?
Individuals diagnosed as caffeine-dependent are deprived of caffeine-rich foods
and assigned to receive daily pills. Sometimes, the pills contain caffeine and other
times they contain a placebo. Depression was assessed.
 Q:
Does lack of caffeine increase depression?
There are 2 data points
for each subject, but
we’ll only look at the
difference. The sample
distribution appears
appropriate for a t-test.
Depression Depression
Subject with Caffeine with Placebo
1
5
16
2
5
23
3
4
5
4
3
7
5
8
14
6
5
24
7
0
6
8
0
3
9
2
15
10
11
12
11
1
0
Does lack of caffeine increase depression?
For each individual in the sample, we have calculated a difference in depression score
(placebo minus caffeine).
There were 11 “difference” points, thus df = n − 1 = 10.
We calculate that
x= 7.36; s = 6.92
H0: difference = 0 ; H0: difference > 0

x 0
7.36
t

 3.53
s n 6.92 / 11
For df = 10, p-value=0.0027.
(1)Since p-value < 0.05, reject H0.
Depression Depression Placebo Subject with Caffeine with Placebo Cafeine
1
5
16
11
2
5
23
18
3
4
5
1
4
3
7
4
5
8
14
6
6
5
24
19
7
0
6
6
8
0
3
3
9
2
15
13
10
11
12
1
11
1
0
-1
(2) We have enough evidence to conclude that:
Caffeine deprivation causes a significant increase in depression.
Two independent samples confidence interval
and hypothesis testing
Two (independent) sample scenario
42
Portland Cement Formulation (page 26)
An engineer is studying the formulation of a
Portland cement mortar. He has added a
polymer latex emulsion during mixing to
determine if this impacts the curing time and
tension bond strength of the mortar.
The experimenter prepared 10 samples of the
original formulation and 10 samples of the
modified formulation.
Q: How many factor(s)? How many levels?
Factor: mortar formulation;
Levels: two different formulations as two treatments or as two levels.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Graphical View of the Data
Dot Diagram, Fig. 2.1, pp 26
43
Q: Visually, do you see any difference between these two samples?
Q: If yes, do you see large, modest or very small difference?
Q: How to compare the difference between these two samples?
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
The Hypothesis Testing Framework for two
sample t-test
44



Statistical hypothesis testing is a useful framework for many
experimental situations
Origins of the methodology date from the early 1900s
We will use a procedure known as the two-sample Z-test and
two-sample t-test.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Another example: If you have a large sample, a
histogram may be useful
45
Graphical description of variability: with 200 observations
Noise: called experimental error.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Box Plots, Fig. 2.3, pp. 28
46
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Inferences about the differences in means,
Randomized designs
47



The Hypothesis Testing Framework:
Sampling from a normal distribution
Statistical hypotheses: H :   
0
1
2
H1 : 1   2
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
47
Errors in Hypothesis testing
48



If the Null hypothesis is rejected, when it is true, a type I error
occurred:
α = Pr(Type I error) = Pr(Reject H0 | H0 is true).
α is also called significance level.

If the Null hypothesis is not rejected, when it is false, a type II
error occurred:
β = Pr(Type II error) = Pr(Fail to reject H0 | H0 is false)

Power = 1- β = Pr(reject H0 | H0 is false)

Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Portland Cement
Summary Statistics (pg. 38)
49
If we want to test: H 0 : 1   2
H1 : 1   2
Modified Mortar
Unmodified Mortar
“New recipe”
“Original recipe”
y1  16.76
y2  17.04
S12  0.100
S22  0.061
S1  0.316
S2  0.248
n1  10
n2  10
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Portland Cement Example
50
If we want to test: H 0 : 1   2
H1 : 1   2

We will consider three cases for this example:

Case 1: Assume σ1 and σ2 are known: let σ1 = σ2 = 0.30.

Case 2: Assume σ1 and σ2 are unknown, and σ1 = σ2.

Case 3: Assume σ1 and σ2 are unknown, and σ1 ≠ σ2.

Then Case 1 will give two-sample Z-test; Case 2 will give twosample (pooled) t-test, and Case 3 will give two-sample t-test.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Two-Sample Z-Test: if σ1 and σ2 are known
51
Case 1: Assume σ1 and σ2 are known: let σ1 = σ2 = 0.30.
Use the sample means to draw inferences about the population means
y1  y2  16.76  17.04  0.28
Difference in sample means
Standard deviation of the difference in sample means
 
2
y
2
n
, and 
2
y1  y2
=
 12
n1

 22
n2
, y1 and y2 independent REALLY? (STT315)
This suggests a statistic:
Z0 
y1  y2
 12
n1

 22
n2
If the variances were known we could use the normal distribution as the basis of a test
Z0 has a N(0,1) distribution if the two population means are equal
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
If we knew the two variances how would we use Z0
to test H0?
52
Case 1: Assume σ1 and σ2 are known: let σ1 = σ2 = 0.30.

Suppose that σ1 = σ2 = 0.30. Then we can calculate
Z0 
y1  y2

2
1
n1





2
2
n2

0.28
0.28

 2.09
2
2
0.1342
0.3 0.3

10
10
How “unusual” is the value Z0 = -2.09 if the two population means
are equal?
It turns out that 95% of the area under the standard normal curve
(probability) falls between the values Z0.025 = 1.96 and - Z0.025 = 1.96. (that is: the critical value Z*=1.96.)
So the value Z0 = -2.09 is pretty unusual in that it would happen less
that 5% of the time if the population means were equal
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Standard Normal Table (see appendix) for critical values
53
Critical Value:
Z*=qnorm(0.975)
=1.959964
Z0.025 = 1.96
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Inferences about the differences in means,
Randomized designs Case 1: Assume σ1 and σ2 are known:
let σ1 = σ2 = 0.30.
54
So if the variances were known we would conclude that we should
reject the null hypothesis at the 5% level of significance
H 0 : 1   2
H1 : 1   2
and conclude that the alternative hypothesis is true.
This is called a fixed significance level test, because we compare the value
of the test statistic to a critical value (1.96) that we selected in advance
before running the experiment.
The standard normal distribution is the reference distribution for the test.
Another way to do this that is very popular is to use the P-value approach.
The P-value can be thought of as the observed significance level.
For the Z-test, it is easy to find the P-value.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Normal Table
Case 1: Assume σ1 and σ2 are known:
let σ1 = σ2 = 0.30.
55
Find the probability above Z0 = -2.09
from the table.
This is 1 – 0.98169 = 0.01832
Z0.025 = 1.96
The P-value is twice this probability, or
0.03662.
So we would reject the null hypothesis
at any level of significance that is less
than or equal to 0.03662.
Typically 0.05 is used as the cutoff.
In R, we use
2*pnorm(-2.09) =0.0366178
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Two-sample t –Test if σ1 and σ2 are unknown
56





The two-sample Z-test just described would work perfectly if we
knew the two population variances.
Since we usually don’t know the true population variances, what
would happen if we just plugged in the sample variances?
The answer is that if the sample sizes were large enough (say both
n> 30 or 40) the Z-test would work just fine. It is a good largesample test for the difference in means.
But many times that isn’t possible (as Gosset wrote in 1908, “…but
what if the sample size is small…?).
It turns out that if the sample size is small we can no longer use the
N(0,1) distribution as the reference distribution for the test.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
How the Two-Sample t-Test Works:
57
Case 2: Assume σ1 and σ2 are unknown, and σ1 = σ2.
Use S12 and S22 to estimate  12 and  22
The test statistic is y1  y2
The previous ratio becomes
2
y2
S12 y1 S
2
t0 

n1 n12
1
S p 2 2
2
n
However, we have the case where n




2
1 1
2
Pool the individual sample variances:
2
2
(
n

1)
S

(
n

1)
S
1
2
2
S p2  1
n1  n2  2
df=n1 + n2 - 2
is an estimate of the common variance
Or call:
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
How the Two-Sample (pooled) t-Test Works:
Case 2: Assume σ1 and σ2 are unknown, and σ1 = σ2.
The test statistic is
t0 




y1  y2
1 1
Sp

n1 n2
df=n1 + n2 - 2
The denominator is called the
standard error of the difference
in means. SE(
).
Values of t0 that are very different from zero are consistent with
the alternative hypothesis
t0 is a “distance” measure-how far apart the averages are
expressed in standard deviation units
Notice the interpretation of t0 as a signal-to-noise ratio.
Chapter 2
Design & Analysis of Experiments 8E 2012 Montgomery
58
The Two-Sample (Pooled) t-Test
59
Case 2: Assume σ1 and σ2 are unknown, and σ1 = σ2.
(n1  1) S12  (n2  1) S 22 9(0.100)  9(0.061)
S 

 0.081
n1  n2  2
10  10  2
2
p
S p  0.284
t0 
y1  y2
16.76  17.04

 2.20
1 1
1 1
Sp

0.284

n1 n2
10 10
The two sample means are a little over two standard deviations apart
Is this a "large" difference?
In R: 2*pt(-2.20, 18) = 0.04110859.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Two-Sample (Pooled) t-Test: a more general case
60
Case 2: Assume σ1 and σ2 are unknown, and σ1 = σ2.
If we would like to test a more general case, for example:
H0: µ1-µ2=10 v.s. H0: µ1-µ2≠10
Then the test statistic will be (see page 43):
t0 =
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
William Sealy Gosset (1876, 1937)
Gosset's interest in barley cultivation led
him to speculate that design of
experiments should aim, not only at
improving the average yield, but also at
breeding varieties whose yield was
insensitive (robust) to variation in soil and
climate.
Developed the t-test (1908)
Gosset was a friend of both Karl Pearson
and R.A. Fisher, an achievement, for each
had a monumental ego and a loathing for
the other.
Gosset was a modest man who cut short
an admirer with the comment that “Fisher
would have discovered it all anyway.”
61
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
The Two-Sample (Pooled) t-Test



We need an objective
basis for deciding how
large the test statistic t0
really is.
In 1908, W. S. Gosset
derived the reference
distribution for t0 … called
the t distribution.
Tables of the t distribution –
see textbook appendix
page 614.
t0 = -2.20
Critical Value:
t*=qt(0.975, 18)=2.100922.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
62
Critical Value:
t*=qt(0.975, 18)
=2.100922.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
63
The Two-Sample (Pooled) t-Test




A value of t0 between:
–2.101 and 2.101 is
consistent with equality of
means.
It is possible for the means
to be equal and t0 to either
exceed 2.101 or below
–2.101, but it would be a
“rare event” … leads to the
conclusion that the means
are different.
Could also use the P-value
approach.
t0 = -2.20
Critical Value:
t*=qt(0.975, 18)=2.100922.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
64
The Two-Sample (Pooled) t-Test
t0 = -2.20
Critical Value:
t*=qt(0.975, 18)=2.100922.




The P-value is the area (probability) in the tails of the t-distribution beyond -2.20 + the
probability beyond +2.20 (it’s a two-sided test).
The P-value is a measure of how unusual the value of the test statistic is given that the null
hypothesis is true.
The P-value the risk of wrongly rejecting the null hypothesis of equal means (it measures
rareness of the event).
The exact P-value in our problem is P = 0.042 (found from a computer).
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
65
Approximating the P-value




Our t-table only gives probabilities greater than positive values
of t. So take the absolute value of t0 = -2.20 or |t0|= 2.20.
Now with 18 degrees of freedom, find the values of t in the
table that bracket this value.
These are 2.101 < |t0|= 2.20 < 2.552. The right-tail
probability for t = 2.101 is 0.025 and for t = 2.552 is 0.01.
Double these probabilities because this is a two-sided test.
Therefore the P-valuemust lie between these two probabilities, or

0.05 <P-value < 0.02

These are upper and lower bounds on the P-value.

We know that the actual P-value is 0.042.
Chapter 2
Design & Analysis of Experiments 8E
2012 Montgomery
66
Computer Two-Sample t-Test Results
67
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Checking Assumptions –
Normal Probability Plot (called QQ-plot)
68
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Importance of the t-Test
69


Provides an objective framework for simple comparative
experiments
Could be used to test all relevant hypotheses in a two-level
factorial design, because all of these hypotheses involve the
mean response at one “side” of the cube versus the mean
response at the opposite “side” of the cube. (See page 6-7)
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Confidence Intervals (See pg. 43)
70


Hypothesis testing gives an objective statement concerning
the difference in means, but it doesn’t specify “how
different” they are.
General form of a confidence interval
L    U where P( L    U )  1  a

The 100(1- α)% confidence interval on the difference in
two means:
y1  y2  ta / 2,n1  n2 2 S p (1/ n1 )  (1/ n2 )  1  2 
y1  y2  ta / 2,n1  n2 2 S p (1/ n1 )  (1/ n2 )
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Critical Value:
t*=qt(0.975, 18)
=2.100922.
Example, page 43-44:
71
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
The two sample t-test (with unequal variance)
Case 3: Assume σ1 and σ2 are unknown, and σ1 ≠ σ2.
The degrees of freedom v associated
with this variance estimate is
approximated using the
Welch–Satterthwaite equation
72
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
What if the Two Variances are Different?
73
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Example 2.1, page 48:
74
Case 3: Assume σ1 and σ2 are unknown, and σ1 ≠ σ2.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Example 2.1, page 48:
75

P-value = pt(-2.7354, 16.1955) =0.007274408.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Two sample t-test in R: Table 2.1
76

Two-sample (pooled) t-test, and two-sample t-test.
For H0: mu1=mu2, v.s. Ha: mu1≠mu2
OR: H0: mu_diff = 0, v.s. Ha: mu_diff ≠ 0
## Two sample t-test in R with t.test###############################
data.tab2.1<-read.table("http://people.uncw.edu/chenc/STT411/dataset%20backup/Tension-Bond.TXT ",
header=TRUE);
x<-data.tab2.1[,1];
y<-data.tab2.1[,2];
t.test(x,y,alternative ="two.sided", mu=0, var.equal = TRUE);
t.test(x,y,alternative ="two.sided", mu=0, var.equal = FALSE);
var.equal: If TRUE then the pooled variance is used to estimate the variance,
otherwise the Welch (or Satterthwaite) approximation to the degrees of freedom is used.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Check the normality assumption
77
########## check the normality assumption ######
qqnorm(x); qqline(x);
#### what does the qq plot do? ######
par(mfrow=c(2,2))
aa<-rnorm(100)
qqnorm(aa); qqline(aa);
bb<-rnorm(100,10,2)
qqnorm(bb); qqline(bb);
cc<-rexp(100)
qqnorm(cc); qqline(cc); dev.off(); ## to close the plotting window
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Other Chapter Topics
78


Hypothesis testing when the variances are known.
One sample inference (t and Z tests), by comparing to a specific
value.

Hypothesis tests on variances (F tests).

Paired experiments – this is an example of blocking. (chap 4)
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
How do we know whether the variance of two
samples are different?
79
So F0 follows F (n1-1,n2-1) distribution.
We call n1-1 the numerator df, and n2-1 the denominator df.
For two-sided F-test: p-value=2*(one-tail area). But how to decide which side to double?
First, find the center of Median by qf(0.5, n1-1, n2-1).
If test statistic F0 is on the RIGHT side of the center, then double the RIGHT side.
P-value= 2* (1-pf(F0, n1-1, n2-1))
& Analysis
Experiments2*
8E pf(F0,
2012 Montgomery
Otherwise, doubleDesign
the LEFT
side.ofP-value=
n1-1, n2-1). Chapter 2
Example 2.3, page 58:
80
In R: One-sided test with f0=14.5/10.8; 1-pf(f0,11,9) = 0.3344771.
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Portland Cement Formulation: Check the equal
variance assumption
81
data.tab2.1<-read.table("http://people.uncw.edu/chenc/STT411/dataset%20backup/Tension-Bond.TXT ",
header=TRUE);
x<-data.tab2.1[,1];
y<-data.tab2.1[,2];
########## check the equal variance assumption ####
var.x<-var(x);
var.y<-var(y);
n<-length(x)
2*(1-pf(var.x/var.y,n-1,n-1))
# or
var.test(x,y);
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Summary of Ch2
82

Sampling distribution
y
~ N (0,1)
/ n
y
~ t (df  n  1)
s/ n

Steps for hypothesis testing

One sample Z-test

CI for population mean when population SD is given

One sample t-test

CI for population mean when sample SD is given
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Summary of Ch2
83







Two sample Z-test
Two sample t-test with equal variance assumption
CI for the difference of two population mean with equal
variance assumption
Two sample t-test without equal variance assumption.
CI for the difference of two population mean with equal
variance assumption
Check normality assumption
Check equal variance assumption with F test
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Summary about R
84









pnorm;
qnorm;
pt;
qt;
qqnorm;
qqline;
t.test;
var.test;
pf;
Design & Analysis of Experiments 8E 2012 Montgomery
Chapter 2
Download