compute

advertisement
Chapter 13
Inference About Comparing
Two Populations
Copyright © 2009 Cengage Learning
Comparing Two Populations
Previously we looked at techniques to estimate and test
parameters for one population:
Population Mean µ
Population Proportion p
We will still consider these parameters when we are looking
at two populations, however our interest will now be:
 The difference between two means.
 The ratio of two variances.
 The difference between two proportions.
Copyright © 2009 Cengage Learning
Difference between Two Means
In order to test and estimate the difference between two
population means, we draw random samples from each of
two populations. Initially, we will consider independent
samples, that is, samples that are completely unrelated to one
another.
Population 1
Sample, size: n1
Parameters:
(Likewise, we consider
Copyright © 2009 Cengage Learning
Statistics:
for Population 2)
Difference between Two Means
Because we are comparing two population means, we use the
statistic,
which is an unbiased and consistent estimator of µ1- µ2.
Copyright © 2009 Cengage Learning
Sampling Distribution of
1.
is normally distributed if the original populations
are normal –or– approximately normal if the populations are
nonnormal and the sample sizes are large (n1, n2 > 30)
2. The expected value of
3. The variance of
is µ1- µ2
is
and the standard error is:
Copyright © 2009 Cengage Learning
Making Inferences About μ1-μ2
Since
is normally distributed if the original
populations are normal –or– approximately normal if the
populations are nonnormal and the sample sizes are large,
then:
is a standard normal (or approximately normal) random
variable.
We could use this to build the test statistic and the
confidence interval estimator for µ1 - µ2.
Copyright © 2009 Cengage Learning
Making Inferences About μ1-μ2
…except that, in practice, the z statistic is rarely used since
the population variances are unknown.
??
Instead we use a t-statistic. We consider two cases for the
unknown population variances: when we believe they are
equal and conversely when they are not equal.
More about this later…
Copyright © 2009 Cengage Learning
Test Statistic for μ1-μ2 (equal variances)
Calculate
– the pooled variance estimator as…
…and use it here:
degrees of freedom
Copyright © 2009 Cengage Learning
CI Estimator for μ1-μ2 (equal variances)
The confidence interval estimator for μ1-μ2 when the
population variances are equal is given by:
pooled variance estimator
Copyright © 2009 Cengage Learning
degrees of freedom
Test Statistic for μ1-μ2 (unequal variances)
The test statistic for μ1-μ2 when the population variances are
unequal is given by:
degrees of freedom
Likewise, the confidence interval estimator is:
Copyright © 2009 Cengage Learning
Which test to use?
Which test statistic do we use? Equal variance or unequal
variance?
Whenever there is insufficient evidence that the variances
are unequal, it is preferable to perform the
equal variances t-test.
This is so, because for any two given samples:
The number of degrees of
freedom for the equal
variances case
≥
Larger numbers of degrees of
freedom have the same effect as
having larger sample sizes
Copyright © 2009 Cengage Learning
The number of degrees
of freedom for the unequal
variances case
≥
Testing the Population Variances
Testing the Population Variances
H0: σ12 / σ22 = 1
H1: σ12 / σ22 ≠ 1
Test statistic: s12 / s22, which is F-distributed with degrees of
freedom ν1 = n1– 1 and ν2 = n2 −2.
The required condition is the same as that for the t-test of
µ1 - µ2 , which is both populations are normally distributed.
Copyright © 2009 Cengage Learning
Testing the Population Variances
This is a two-tail test so that the rejection region is
F  F / 2,1 , 2 or F  F1 / 2, ,
1 2
Copyright © 2009 Cengage Learning
Example 13.1
Millions of investors buy mutual funds choosing from
thousands of possibilities.
Some funds can be purchased directly from banks or other
financial institutions while others must be purchased through
brokers, who charge a fee for this service.
This raises the question, can investors do better by buying
mutual funds directly than by purchasing mutual funds
through brokers.
Copyright © 2009 Cengage Learning
Example 13.1
To help answer this question a group of researchers
randomly sampled the annual returns from mutual funds that
can be acquired directly and mutual funds that are bought
through brokers and recorded the net annual returns, which
are the returns on investment after deducting all relevant
fees.
Xm13-01
Can we conclude at the 5% significance level that directlypurchased mutual funds outperform mutual funds bought
through brokers?
Copyright © 2009 Cengage Learning
Example 13.1
IDENTIFY
To answer the question we need to compare the population
of returns from direct and the returns from broker- bought
mutual funds.
The data are obviously interval (we've recorded real
numbers).
This problem objective - data type combination tells us that
the parameter to be tested is the difference between two
means µ1- µ2.
Copyright © 2009 Cengage Learning
Example 13.1
IDENTIFY
The hypothesis to be tested is that the mean net annual return
from directly-purchased mutual funds (µ1) is larger than the
mean of broker-purchased funds (µ2). Hence the alternative
hypothesis is
H1: µ1- µ2 > 0
and
H0: µ1- µ2 = 0
To decide which of the t-tests of µ1 - µ2 to apply we conduct
the F-test of σ12/ σ22 .
Copyright © 2009 Cengage Learning
Example 13.1
IDENTIFY
From the data we calculated the following statistics.
s12 = 37.49 and s22 = 43.34
Test statistic: F = 37.49/43.34 = 0.86
Rejection region:
F  F / 2,1 ,2  F.025,49,49  F.025,50,50  1.60
or
F  F1 / 2,1 , 2  F.975,49,49  1 / F.025,49,49  1 / F.025,50,50  1 / 1.60  .63
Copyright © 2009 Cengage Learning
Example 13.1
IDENTIFY
Click Data, Data Analysis, and F-Test Two Sample for
Variances
Copyright © 2009 Cengage Learning
Example 13.1
IDENTIFY
A
B
C
1 F-Test Two-Sample for Variances
2
3
Direct
Broker
4 Mean
6.63
3.72
5 Variance
37.49
43.34
6 Observations
50
50
7 df
49
49
8 F
0.86
9 P(F<=f) one-tail
0.3068
10 F Critical one-tail
0.6222
The value of the test statistic is F = .86. Excel outputs the
one-tail p-value. Because we're conducting a two-tail test,
we double that value. Thus, the p-value of the test we're
conducting is 2  .3068 = .6136.
Copyright © 2009 Cengage Learning
Example 13.1
IDENTIFY
There is not enough evidence to infer that the population
variances differ. It follows that we must apply the equalvariances t-test of µ1- µ2
Copyright © 2009 Cengage Learning
Example 13.1
COMPUTE
For manual calculations click
Example 13.1 Manual Calculations
For Excel skip to next slide.
Copyright © 2009 Cengage Learning
12.22
Example 13.1
COMPUTE
Click Data, Data Analysis, t-Test: Two-Sample Assuming
Equal Variances
Copyright © 2009 Cengage Learning
Example 13.1
COMPUTE
A
B
C
1 t-Test: Two-Sample Assuming Equal Variances
2
3
Direct
Broker
4 Mean
6.63
3.72
5 Variance
37.49
43.34
6 Observations
50
50
7 Pooled Variance
40.41
8 Hypothesized Mean Difference
0
9 df
98
10 t Stat
2.29
11 P(T<=t) one-tail
0.0122
12 t Critical one-tail
1.6606
13 P(T<=t) two-tail
0.0243
14 t Critical two-tail
1.9845
Copyright © 2009 Cengage Learning
Example 13.1
INTERPRET
The value of the test statistic is 2.29. The one-tail p-value is
.0122.
We observe that the p-value of the test is small (and the test
statistic falls into the rejection region).
As a result we conclude that there is sufficient evidence to
infer that on average directly-purchased mutual funds
outperform broker-purchased mutual funds
Copyright © 2009 Cengage Learning
Confidence Interval Estimator
Suppose we wanted to compute a 95% confidence interval
estimate of the difference between mean caloric intake for
consumers and non-consumers of high-fiber cereals. The
unequal-variances estimator is
(x1  x 2 )  t  / 2
2
s p 
1
1 


 n1 n 2 
We use the t-Estimate_2 Means (Eq-Var) worksheet in the
Estimators workbook or manually (Click here).
Copyright © 2009 Cengage Learning
Confidence Interval Estimator
COMPUTE
A
B
C
D
1 t-Estimate of the Difference Between Two Means (Equal-Variances)
2
3
Sample 1 Sample 2 Confidence Interval Estimate
4 Mean
6.63
3.72
2.91
5 Variance
37.49
43.34
Lower confidence limit
6 Sample size
50
50
Upper confidence limit
7 Pooled Variance
40.42
8 Confidence level
0.95
Copyright © 2009 Cengage Learning
E
F
±
2.52
0.39
5.43
Confidence Interval Estimator
INTERPRET
We estimate that the return on directly purchased mutual
funds is on average between .39 and 5.43 percentage points
larger than broker-purchased mutual funds.
Copyright © 2009 Cengage Learning
Example 13.2
What happens to the family-run business when the boss’s
son or daughter takes over?
Does the business do better after the change if the new boss
is the offspring of the owner or does the business do better
when an outsider is made chief executive officer (CEO)?
In pursuit of an answer researchers randomly selected 140
firms between 1994 and 2002, 30% of which passed
ownership to an offspring and 70% appointed an outsider as
CEO.
Copyright © 2009 Cengage Learning
Example 13.2
For each company the researchers calculated the operating
income as a proportion of assets in the year before and the
year after the new CEO took over.
The change (operating income after – operating income
before) in this variable was recorded.
Xm13-02
Do these data allow us to infer that the effect of making an
offspring CEO is different from the effect of hiring an
outsider as CEO?
Copyright © 2009 Cengage Learning
Example 13.2
IDENTIFY
The problem objective is to compare two populations.
Population 1: Operating income of companies whose CEO is
an offspring of the previous CEO
Population 2: Operating income of companies whose CEO is
an outsider
The data type is interval (operating incomes).
Thus, the parameter to be tested is µ1- µ2, where µ1 = mean
operating income for population 1 and µ2 = mean operating
income for population 2.
Copyright © 2009 Cengage Learning
Example 13.2
IDENTIFY
We want to determine whether there is enough statistical
evidence to infer that µ1 is different from µ2. That is, that
µ1- µ2 is not equal to 0. Thus,
H1: µ1- µ2 ≠ 0
and
H0: µ1- µ2 = 0
We need to determine whether to use the equal-variances or
unequal-variances t –test of µ1- µ2.
Copyright © 2009 Cengage Learning
Example 13.2
IDENTIFY
Click Data, Data Analysis, and F-Test Two Sample for
Variances
Copyright © 2009 Cengage Learning
Example 13.2
IDENTIFY
A
B
C
1 F-Test Two-Sample for Variances
2
3
Offspring Outsider
4 Mean
-0.10
1.24
5 Variance
3.79
8.03
6 Observations
42
98
7 df
41
97
8 F
0.47
9 P(F<=f) one-tail
0.0040
10 F Critical one-tail
0.6314
The value of the test statistic is F = .47. The p-value of the
test we're conducting is 2  .0040 = .0080.
Copyright © 2009 Cengage Learning
Example 13.2
IDENTIFY
Thus, the correct technique is the unequal-variances t-test of
µ1- µ2.
Copyright © 2009 Cengage Learning
Example 13.2
COMPUTE
For manual calculations click
Example 13.2 manual calculations
For Excel skip to next slide.
Copyright © 2009 Cengage Learning
12.37
Example 13.2
COMPUTE
Click Data, Data Analysis, t-Test: Two-Sample Assuming
Unequal Variances
Copyright © 2009 Cengage Learning
Example 13.2…
COMPUTE
A
B
C
1 t-Test: Two-Sample Assuming Unequal Variances
2
3
Offspring Outsider
4 Mean
-0.10
1.24
5 Variance
3.79
8.03
6 Observations
42
98
7 Hypothesized Mean Difference
0
8 df
111
9 t Stat
-3.22
10 P(T<=t) one-tail
0.0008
11 t Critical one-tail
1.6587
12 P(T<=t) two-tail
0.0017
13 t Critical two-tail
1.9816
Copyright © 2009 Cengage Learning
Example 13.2…
INTERPRET
The t-statistic is – 3.22 and its p-value is .0017. Accordingly,
we conclude that there is sufficient evidence to infer that the
mean times differ.
Copyright © 2009 Cengage Learning
Confidence Interval Estimator
We can also draw inferences about the difference between
the two population means by calculating the confidence
interval estimator. We use the unequal-variances confidence
interval estimator of and a 95% confidence level.
We use the t-Estimate_2 Means (Uneq-Var) worksheet in the
Estimators workbook or manually (Click here).
Copyright © 2009 Cengage Learning
Confidence Interval Estimator
COMPUTE
We activate the t-Estimate_2 Means (Uneq-Var)
worksheet in the Estimators workbook and substitute the
sample statistics and confidence level.
A
B
C
D
1 t-Estimate of the Difference Between Two Means (Unequal-Variances)
2
3
Sample 1 Sample 2 Confidence Interval Estimate
4 Mean
-0.10
1.24
-1.34
5 Variance
3.79
8.03
Lower confidence limit
6 Sample size
42
98
Upper confidence limit
7 Degrees of freedom
110.75
8 Confidence level
0.95
Copyright © 2009 Cengage Learning
E
F
±
0.82
-2.16
-0.51
Confidence Interval Estimator
INTERPRET
We estimate that the mean change in operating incomes for
outsiders exceeds the mean change in the operating income
for offspring lies between .51 and 2.16 percentage points.
Copyright © 2009 Cengage Learning
Checking the Required Condition
Both the equal-variances and unequal-variances techniques
require that the populations be normally distributed. As
before, we can check to see whether the requirement is
satisfied by drawing the histograms of the data.
Copyright © 2009 Cengage Learning
Checking the Required Condition: Example 13.1
Frequency
.
Histogram
20
10
0
-5
0
5
10
15
20
More
15
20
More
Direct
Frequency
Histogram
20
10
0
-5
0
5
10
Broker
Copyright © 2009 Cengage Learning
Checking the Required Condition: Example 13.2
Frequency
.
Histogram
20
10
0
-4
-2
0
2
4
Offspring
Frequency
Histogram
40
20
0
-4
-2
0
2
4
Outsider
Copyright © 2009 Cengage Learning
6
8
10
Violation of the Required Condition
When the normality requirement is unsatisfied, we can use a
nonparametric technique-the Wilcoxon rank sum test for
independent samples (Chapter 19)--to replace the equalvariances test of µ1-µ2 .
We have no alternative to the unequal-variances test of µ1-µ2
when the populations are very nonnormal.
Copyright © 2009 Cengage Learning
Terminology
If all the observations in one sample appear in one column
and all the observations of the second sample appear in
another column, the data is unstacked.
If all the data from
both samples is in the
same column, the data
is said to be stacked.
Copyright © 2009 Cengage Learning
Developing an Understanding of Statistical Concepts 1
The formulas in this section are relatively complicated.
However, conceptually both test statistics are based on the
techniques we introduced in Chapter 11 and repeated in
Chapter 12.
That is, the value of the test statistic is the difference
between the statistic and the hypothesized value of the
parameter measured in terms of the standard error.
Statistic  Parameter
Test statistic 
S tan dard error
Copyright © 2009 Cengage Learning
Developing an Understanding of Statistical Concepts 2
As was the case with the interval estimator of p, the standard
error must be estimated from the data for all inferential
procedures introduced here.
The method we use to compute the standard error of x1  x 2
depends on whether the population variances are equal. When
they are equal we calculate and use the pooled variance estimator
sp2.
We are applying an important principle here, and we will so
again in Section 13.5 and in later chapters. Where possible, it is
advantageous to pool sample data to estimate the standard error.
Copyright © 2009 Cengage Learning
Developing an Understanding of Statistical Concepts 2
In the previous application, we are able to pool because we
assume that the two samples were drawn from populations with a
common variance.
Combining both samples increases the accuracy of the estimate.
Thus, sp2 is a better estimator of the common variance than either
s12 or s22 separately.
When the two population variances are unequal, we cannot pool
the data and produce a common estimator.
We must compute and use them to estimate σ12 and σ22,
respectively.
Copyright © 2009 Cengage Learning
Identifying Factors I…
Factors that identify the equal-variances t-test and estimator
of
:
Copyright © 2009 Cengage Learning
Identifying Factors II…
Factors that identify the unequal-variances t-test and
estimator of
:
Copyright © 2009 Cengage Learning
“Ignorance ain’t what folks don’t know, it’s what folks know
that just ain’t so”.
Will Rogers
Few things illustrate this more than how statistical results are
interpreted.
Copyright © 2009 Cengage Learning
Example 13.3
Despite some controversy, scientists generally agree that
high-fiber cereals reduce the likelihood of various forms of
cancer.
However, one scientist claims that people who eat
high-fiber cereal for breakfast will consume, on average,
fewer calories for lunch than people who don't eat high-fiber
cereal for breakfast.
Copyright © 2009 Cengage Learning
Example 13.3
If this is true, high-fiber cereal manufacturers will be able to
claim another advantage of eating their product--potential
weight reduction for dieters.
As a preliminary test of the claim, 150 people were
randomly selected and asked what they regularly eat for
breakfast and lunch.
Copyright © 2009 Cengage Learning
Example 13.3
Each person was identified as either a consumer or a
nonconsumer of high-fiber cereal, and the number of calories
consumed at lunch was measured and recorded. Xm13-03
Can the scientist conclude at the 5% significance level that
his belief is correct?
Copyright © 2009 Cengage Learning
Example 13.3
H 0 : (1   2 )  0
H1 : (1   2 )  0
.
A
B
C
1 t-Test: Two-Sample Assuming Unequal Variances
2
3
Consumers Nonconsumers
4 Mean
604.02
633.23
5 Variance
4103
10670
6 Observations
43
107
7 Hypothesized Mean Difference
0
8 df
123
9 t Stat
-2.09
10 P(T<=t) one-tail
0.0193
11 t Critical one-tail
1.6573
12 P(T<=t) two-tail
0.0386
13 t Critical two-tail
1.9794
Copyright © 2009 Cengage Learning
Example 13.3
The value of the test statistic is −2.09.
The one-tail p-value is .0193.
We conclude that there is sufficient evidence to infer that
consumers of high-fiber cereal do eat fewer calories at lunch
than do nonconsumers.
Copyright © 2009 Cengage Learning
Observational and Experimental Data
From this result, we're inclined to believe that eating a highfiber cereal at breakfast may be a way to reduce weight.
However, other interpretations are plausible.
For example, people who eat fewer calories are probably
more health conscious, and such people are more likely to
eat high-fiber cereal as part of a healthy breakfast.
In this interpretation, high-fiber cereals do not necessarily
lead to fewer calories at lunch.
Copyright © 2009 Cengage Learning
Observational and Experimental Data
Instead another factor, general health consciousness, leads to
both fewer calories at lunch and high-fiber cereal for
breakfast.
Notice that the conclusion of the statistical procedure is
unchanged.
On average, people who eat high-fiber cereal consume fewer
calories at lunch. However, because of the way the data were
gathered, we have more difficulty interpreting this result.
Copyright © 2009 Cengage Learning
Observational and Experimental Data
From the result in Example 13.3, we're inclined to believe that
eating a high-fiber cereal at breakfast may be a way to reduce
weight.
However, other interpretations are possible. For example, people
Who eat fewer calories at lunch are probably more health
conscious, and such people are more likely to eat high-fiber
cereal as part of a healthy breakfast.
In this interpretation, high-fiber cereals do not necessarily lead to
Fewer calories at lunch. Instead another factor, general health
consciousness, leads to both fewer calories at lunch and high-fiber
cereal for breakfast.
Copyright © 2009 Cengage Learning
Observational and Experimental Data
Suppose that we redo Example 13.3 using the experimental
approach.
We randomly select 150 people to participate in the
experiment.
We randomly assign 75 to eat high-fiber cereal for breakfast
and the other 75 to eat something else.
We then record the number of calories each person
consumes at lunch.
Copyright © 2009 Cengage Learning
Observational and Experimental Data
Both groups should be similar in all other dimensions,
including health consciousness. (Larger sample sizes
increase the likelihood that the two groups will be similar.)
If the statistical result is about the same as in Example 13.3,
we may have some valid reason to believe that high-fiber
cereal at breakfast leads to a decrease in caloric intake at
lunch.
Copyright © 2009 Cengage Learning
Matched Pairs Experiment…
Previously when comparing two populations, we examined
independent samples.
If, however, an observation in one sample is matched with
an observation in a second sample, this is called a matched
pairs experiment.
To help understand this concept, let’s consider Example 13.4
Copyright © 2009 Cengage Learning
Example 13.4
In the last few years a number of web-based companies that
offer job placement services have been created.
The manager of one such company wanted to investigate the
job offers recent MBAs were obtaining.
In particular, she wanted to know whether finance majors
were being offered higher salaries than marketing majors.
Copyright © 2009 Cengage Learning
Example 13.4
In a preliminary study she randomly sampled 50 recently
graduated MBAs half of whom majored in finance and half
in marketing.
From each she obtained the highest salary (including
benefits) offer (Xm13-04).
Can we infer that finance majors obtain higher salary offers
than do marketing majors among MBAs?
Copyright © 2009 Cengage Learning
Example 13.4
IDENTIFY
The parameter is the difference between two means (where
µ1 = mean highest salary offer to finance majors and µ2 =
mean highest salary offer to marketing majors).
Because we want to determine whether finance majors are
offered higher salaries, the alternative hypothesis will
specify that is greater than.
Calculation of the F-test of two variances indicates the use
the equal-variances test statistic.
Copyright © 2009 Cengage Learning
Example 13.4
The hypotheses are
H 0 : (1   2 )  0
H1 : (1   2 )  0
The Excel output is:
Copyright © 2009 Cengage Learning
IDENTIFY
Example 13.4
COMPUTE
A
B
C
1 t-Test: Two-Sample Assuming Equal Variances
2
3
Finance
Marketing
4 Mean
65,624
60,423
5 Variance
360,433,294 262,228,559
6 Observations
25
25
7 Pooled Variance
311,330,926
8 Hypothesized Mean Difference
0
9 df
48
10 t Stat
1.04
11 P(T<=t) one-tail
0.1513
12 t Critical one-tail
1.6772
13 P(T<=t) two-tail
0.3026
14 t Critical two-tail
2.0106
Copyright © 2009 Cengage Learning
Example 13.4
INTERPRET
The value of the test statistic (t =1.04) and its p-value (.1513)
indicate that there is very little evidence to support the
hypothesis that finance majors attract higher salary offers
than marketing majors.
Copyright © 2009 Cengage Learning
Example 13.4
INTERPRET
We have some evidence to support the alternative
hypothesis, but not enough.
Note that the difference in sample means is
x1  x 2 = (65,624 -60,423) = 5,201
Copyright © 2009 Cengage Learning
Example 13.5
Suppose now that we redo the experiment in the following
way.
We examine the transcripts of finance and marketing MBA
majors.
We randomly sample a finance and a marketing major whose
grade point average (GPA) falls between 3.92 and 4 (based
on a maximum of 4).
We then randomly sample a finance and a marketing major
whose GPA is between 3.84 and 3.92.
Copyright © 2009 Cengage Learning
Example 13.5
We continue this process until the 25th pair of finance and
marketing majors are selected whose GPA fell between 2.0
and 2.08.
(The minimum GPA required for graduation is 2.0.)
As we did in Example 13.4, we recorded the highest salary
offer .
Xm13-05
Can we conclude from these data that finance majors draw
larger salary offers than do marketing majors?
Copyright © 2009 Cengage Learning
Example 13.5
IDENTIFY
The experiment described in Example 13.4 is one in which
the samples are independent. That is, there is no relationship
between the observations in one sample and the observations
in the second sample. However, in this example the
experiment was designed in such a way that each
observation in one sample is matched with an observation in
the other sample. The matching is conducted by selecting
finance and marketing majors with similar GPAs. Thus, it is
logical to compare the salary offers for finance and
marketing majors in each group. This type of experiment is
called matched pairs.
Copyright © 2009 Cengage Learning
Example 13.5
IDENTIFY
For each GPA group, we calculate the matched pair
difference between the salary offers for finance and
marketing majors.
Copyright © 2009 Cengage Learning
Example 13.5
IDENTIFY
The numbers in black are the original starting salary data
(Xm13-05) ; the numbers in blue were calculated.
although a student is either in Finance OR
in Marketing (i.e. independent), that the
data is grouped in this fashion makes it a
matched pairs experiment (i.e. the two
students in group #1 are ‘matched’ by
their GPA range
the difference of the means is equal to the mean of the differences, hence
we will consider the “mean of the paired differences” as our parameter of interest:
Copyright © 2009 Cengage Learning
Example 13.5
IDENTIFY
Do Finance majors have higher salary offers than Marketing
majors?
Since:
We want to research this hypothesis: H1:
(and our null hypothesis becomes H0:
Copyright © 2009 Cengage Learning
)
Test Statistic for
IDENTIFY
The test statistic for the mean of the population of
differences (
) is:
which is Student t distributed with nD–1 degrees of freedom,
provided that the differences are normally distributed.
Copyright © 2009 Cengage Learning
Example 13.5
COMPUTE
For manual calculations click
Example 13.5 Manual Calculations
For Excel skip to next slide.
Copyright © 2009 Cengage Learning
12.80
Example 13.5
COMPUTE
Click Data, Data Analysis, t-Test: Paired Two- Sample for
Means
Copyright © 2009 Cengage Learning
Example 13.5
COMPUTE
A
B
C
1 t-Test: Paired Two Sample for Means
2
3
Finance
Marketing
4 Mean
65,438
60,374
5 Variance
444,981,810 469,441,785
6 Observations
25
25
7 Pearson Correlation
0.9520
8 Hypothesized Mean Difference
0
9 df
24
10 t Stat
3.81
11 P(T<=t) one-tail
0.0004
12 t Critical one-tail
1.7109
13 P(T<=t) two-tail
0.0009
14 t Critical two-tail
2.0639
Copyright © 2009 Cengage Learning
Example 13.5
INTERPRET
The p-value is .0004. There is overwhelming evidence that
Finance majors do obtain higher starting salary offers than
their peers in Marketing.
Copyright © 2009 Cengage Learning
Example 13.6 Confidence Interval Estimator for µD
We can derive the confidence interval estimator for
algebraically as:
A
B
1 t-Estimate: Mean
2
3
4 Mean
5 Standard Deviation
6 LCL
7 UCL
Copyright © 2009 Cengage Learning
C
Difference
5065
6647
2321
7808
Checking the Required Condition
The population of differences are required to be normally
distributed. As before, we can check to see whether the
requirement is satisfied by drawing the histogram of the
differences.
.
Frequency
Histogram
10
5
0
0
5000
10000
Difference
Copyright © 2009 Cengage Learning
15000
20000
Violation of the Required Condition
If the differences are very nonnormal, we cannot use the ttest of µD.
We can, however, employ a nonparametric technique--the
Wilcoxon signed rank sum test for matched pairs, which we
present in Chapter 19.
Copyright © 2009 Cengage Learning
Independent Samples or Matched Pairs: Which
Experimental Design is Better?
Examples 13.4 and 13.5 demonstrated that the experimental
design is an important factor in statistical inference.
However, these two examples raise several questions about
experimental designs.
1 Why does the matched pairs experiment result in rejecting
the null hypothesis, whereas the independent samples
experiment could not?
Copyright © 2009 Cengage Learning
Independent Samples or Matched Pairs: Which
Experimental Design is Better?
2 Should we always use the matched pairs experiment? In
particular, are there disadvantages to its use?
3 How do we recognize when a matched pairs experiment has
been performed?
Copyright © 2009 Cengage Learning
Independent Samples or Matched Pairs: Which
Experimental Design is Better?
1 The matched pairs experiment worked in Example 13.5 by
reducing the variation in the data.
To understand this point, examine the statistics from both
examples. In Example 13.4, we found
x1  x 2  5,201
and in Example 13.5
x D  5,065
Copyright © 2009 Cengage Learning
Independent Samples or Matched Pairs: Which
Experimental Design is Better?
Thus, the numerators of the two test statistics were quite
similar.
However, the test statistic in Example 13.5 was much larger
than the test statistic in Example 13.4 because of the standard
errors.
Copyright © 2009 Cengage Learning
Independent Samples or Matched Pairs: Which
Experimental Design is Better?
In Example 13.4, we calculated
s 2p  311,330 ,926
2
s p 
1
1 
  4,991

 n1 n 2 
Example 13.5 produced
s D  6,647
sD
 1,329
nD
Copyright © 2009 Cengage Learning
Independent Samples or Matched Pairs: Which
Experimental Design is Better?
2 Will the matched pairs experiment always produce a larger
test statistic than the independent samples experiment? The
answer is, “Not necessarily.”
Suppose that in our example we found that companies did not
consider grade point averages when making decisions about
how much to offer the MBA graduates.
In such circumstances, the matched pairs experiment would
result in no significant decrease in variation when compared
to independent samples.
Copyright © 2009 Cengage Learning
Independent Samples or Matched Pairs: Which
Experimental Design is Better?
3 As you've seen, in this course we deal with questions
arising from experiments that have already been conducted.
Thus, one of your tasks is to determine the appropriate test
statistic.
In the case of comparing two populations of interval data,
you must decide whether the samples are independent (in
which case the parameter is) or matched pairs (in which case
the parameter is) to select the correct test statistic.
Copyright © 2009 Cengage Learning
Independent Samples or Matched Pairs: Which
Experimental Design is Better?
To help you do so, we suggest you ask and answer the
following question:
Does some natural relationship exist between each pair of
observations that provides a logical reason to compare the
first observation of sample 1 with the first observation of
sample 2, the second observation of sample 1 with the second
observation of sample 2, and so on?
If so, the experiment was conducted by matched pairs. If not,
it was conducted using independent samples.
Copyright © 2009 Cengage Learning
Developing an Understanding of Statistical Concepts 1
Two of the most important principles in statistics were
applied in this section.
The first is the concept of analyzing sources of variation. In
Examples 13.4 and 13.5, we showed that by reducing the
variation between salary offers in each sample we were able
to detect a real difference between the two majors..
Copyright © 2009 Cengage Learning
Developing an Understanding of Statistical Concepts 1
This was an application of the more general procedure of
analyzing data and attributing some fraction of the variation
to several sources.
In Example 13.5, the two sources of variation were the GPA
and the MBA major. However, we were not interested in the
variation between graduates with differing GPAs.
Instead we only wanted to eliminate that source of variation,
making it easier to determine whether finance majors draw
larger salary offers.
Copyright © 2009 Cengage Learning
Developing an Understanding of Statistical Concepts 1
In Chapter 14, we will introduce a technique called the
analysis of variance which does what its name suggests; it
analyzes sources of variation in an attempt to detect real
differences.
In most applications of this procedure, we will be interested
in each source of variation and not simply in reducing one
source.
We refer to the process as explaining the variation. The
concept of explained variation also will be applied in
Chapters 16 - 18.
Copyright © 2009 Cengage Learning
Developing an Understanding of Statistical Concepts 2
The second principle demonstrated in this section is that
statistics practitioners can design data-gathering procedures
in such a way that they can analyze sources of variation.
Before conducting the experiment in Example 13.5, the
statistics practitioner suspected that there were large
differences between graduates with different GPAs.
Copyright © 2009 Cengage Learning
Developing an Understanding of Statistical Concepts 2
Consequently, the experiment was organized so that the
effects of those differences were mostly eliminated.
It is also possible to design experiments that allow for easy
detection of real differences and minimize the costs of data
gathering.
Copyright © 2009 Cengage Learning
Identifying Factors…
Factors that identify the t-test and estimator of
Copyright © 2009 Cengage Learning
:
Inference about the ratio of two variances
So far we’ve looked at comparing measures of central
location, namely the mean of two populations.
When looking at two population variances, we consider the
ratio of the variances, i.e. the parameter of interest to us is:
The sampling statistic:
is F distributed with
degrees of freedom.
Copyright © 2009 Cengage Learning
Inference about the ratio of two variances
Our null hypothesis is always:
H0:
(i.e. the variances of the two populations will be equal, hence
their ratio will be one)
Therefore, our statistic simplifies to:
Copyright © 2009 Cengage Learning
Example 13.7
IDENTIFY
In Example 12.3, we applied the chi-squared test of a
variance to determine whether there was sufficient evidence
to conclude that the population variance was less than 1.0.
Suppose that the statistics practitioner also collected data
from another container-filling machine and recorded the fills
of a randomly selected sample.
Can we infer at the 5% significance level that the second
machine is superior in its consistency?
Copyright © 2009 Cengage Learning
Example 13.7
IDENTIFY
The problem objective is to compare two populations where
the data are interval.
Because we want information about the consistency of the
two machines, the parameter we wish to test is σ12 / σ22,
where σ12 is the variance of machine 1 and σ22, is the
variance for machine 2.
Copyright © 2009 Cengage Learning
Example 13.7
IDENTIFY
We need to conduct the F-test of to determine whether the
variance of population 2 is less than that of population 1.
Expressed differently, we wish to determine whether there is
enough evidence to infer that is σ12 is larger than σ22. Hence
the hypotheses we test are
H0: σ12 / σ22 = 1
H1: σ12 / σ22 > 1
Copyright © 2009 Cengage Learning
Example 13.7
COMPUTE
For manual calculations click
Example 13.7 Manual Calculations
For Excel skip to next slide.
Copyright © 2009 Cengage Learning
12.106
Example 13.7
COMPUTE
Click Data, Data Analysis, F-Test Two-Sample for
Variances.
Copyright © 2009 Cengage Learning
Example 13.7
A
B
C
1 F-Test Two-Sample for Variances
2
3
Machine 1 Machine 2
4 Mean
999.7
999.8
5 Variance
0.6333
0.4528
6 Observations
25
25
7 df
24
24
8 F
1.40
9 P(F<=f) one-tail
0.2085
10 F Critical one-tail
1.9838
Copyright © 2009 Cengage Learning
COMPUTE
Example 13.7
INTERPRET
There is not enough evidence to infer that the variance of
machine 2 is less than the variance of machine 1.
Copyright © 2009 Cengage Learning
Example 13.8
Determine the 95% confidence interval estimate of the ratio
of the two population variances in Example 13.7.
The confidence interval estimator for σ12 / σ22 is:
Copyright © 2009 Cengage Learning
Example 13.8
COMPUTE
For manual calculations click
Example 13.8 Manual Calculations
For Excel skip to next slide.
Copyright © 2009 Cengage Learning
12.111
Example 13.8
COMPUTE
Open the F-Estimator_2 Variances worksheet in the Estimators
workbook and substitute sample variances, samples sizes, and the
confidence level.
A
B
C
D
1 F-Estimate of the Ratio of Two Variances
2
3
Sample 1 Sample 2 Confidence Interval Estimate
4 Sample variance
0.6333
0.4528 Lower confidence limit
5 Sample size
25
25
Upper confidence limit
6 Confidence level
0.95
E
0.6164
3.1741
That is, we estimate that σ12 / σ22 lies between .6164 and 3.1741
Note that one (1.00) is in this interval.
Copyright © 2009 Cengage Learning
Identifying Factors
Factors that identify the F-test and estimator of
Copyright © 2009 Cengage Learning
:
Difference Between Two Population Proportions
We will now look at procedures for drawing inferences
about the difference between populations whose data are
nominal (i.e. categorical).
As mentioned previously, with nominal data, calculate
proportions of occurrences of each type of outcome. Thus,
the parameter to be tested and estimated in this section is the
difference between two population proportions: p1–p2.
Copyright © 2009 Cengage Learning
Statistic and Sampling Distribution…
To draw inferences about the the parameter p1–p2, we take
samples of population, calculate the sample proportions and
look at their difference.
is an unbiased estimator for p1–p2.
x1 successes in a
sample of size n1
from population 1
Copyright © 2009 Cengage Learning
Sampling Distribution
The statistic
is approximately normally distributed if
the sample sizes are large enough so that:
Since its “approximately normal” we can describe the
normal distribution in terms of mean and variance…
…hence this z-variable will also be approximately standard
normally distributed:
Copyright © 2009 Cengage Learning
Testing and Estimating p1–p2
Because the population proportions (p1 & p2) are unknown,
the standard error:
is unknown. Thus, we have two different estimators for the
standard error of
, which depend upon the null
hypothesis. We’ll look at these cases on the next slide…
Copyright © 2009 Cengage Learning
Test Statistic for p1–p2
There are two cases to consider…
Copyright © 2009 Cengage Learning
Example 13.9
The General Products Company produces and sells a bath
soap, which is not selling well.
Hoping to improve sales General products decided to
introduce more attractive packaging.
The company’s advertising agency developed two new
designs.
Copyright © 2009 Cengage Learning
Example 13.9
The first design features several bright colors to distinguish
it from other brands.
The second design is light green in color with just the
company’s logo on it.
As a test to determine which design is better the marketing
manager selected two supermarkets.
In one supermarket the soap was packaged in a box using the
first design and in the second supermarket the second design
was used.
Copyright © 2009 Cengage Learning
Example 13.9
The product scanner at each supermarket tracked every
buyer of soap over a one week period.
The supermarkets recorded the last four digits of the scanner
code for each of the five brands of soap the supermarket
sold.
Xm13-09
The code for the General Products brand of soap is 9077(the
other codes are 4255, 3745, 7118, and 8855).
Copyright © 2009 Cengage Learning
Example 13.9
After the trial period the scanner data were transferred to a
computer file.
Because the first design is more expensive management has
decided to use this design only if there is sufficient evidence
to allow them to conclude that it is better.
Should management switch to the brightly-colored design or
the simple green one?
Copyright © 2009 Cengage Learning
Example 13.9
IDENTIFY
The problem objective is to compare two populations. The
first is the population of soap sales in supermarket 1 and the
second is the population of soap sales in supermarket 2.
The data are nominal because the values are “buy General
Products soap” and “buy other companies’ soap.”
These two factors tell us that the parameter to be tested is the
difference between two population proportions p1-p2 (where
p1 and p2 are the proportions of soap sales that are a General
Products brand in supermarkets 1 and 2, respectively.
Copyright © 2009 Cengage Learning
Example 13.9
IDENTIFY
Because we want to know whether there is enough evidence to
adopt the brightly-colored design, the alternative hypothesis is
H1: (p1 – p2) > 0
The null hypothesis must be
H0: (p1 – p2) = 0
which tells us that this is an application of Case 1. Thus, the test
statistic is
z
Copyright © 2009 Cengage Learning
(p̂1  p̂ 2 )
 1
1 

p̂(1  p̂) 
 n1 n 2 
Example 13.9
COMPUTE
For manual calculations click
Example 13.9 Manual Calculations
For Excel skip to next slide.
Copyright © 2009 Cengage Learning
12.125
Example 13.9
COMPUTE
Click Add-Ins, Data Analysis Plus, Z-Test: 2 Proportions
Copyright © 2009 Cengage Learning
Example 13.9
COMPUTE
A
B
C
1 z-Test: Two Proportions
2
3
Supermarket 1 Supermarket 2
4 Sample Proportions
0.1991
0.1493
5 Observations
904
1038
6 Hypothesized Difference
0
7 z Stat
2.90
8 P(Z<=z) one tail
0.0019
9 z Critical one-tail
1.6449
10 P(Z<=z) two-tail
0.0038
11 z Critical two-tail
1.96
Copyright © 2009 Cengage Learning
Example 13.9
INTERPRET
The value of the test statistic is z = 2.90; its p-value is .0019.
There is enough evidence to infer that the brightly-colored
design is more popular than the simple design. As a result, it
is recommended that management switch to the first design.
Copyright © 2009 Cengage Learning
Example 13.10
Suppose in our test marketing of soap packages scenario that
instead of just a difference between the two package
versions, the brightly colored design had to outsell the
simple design by at least 3%
Copyright © 2009 Cengage Learning
Example 13.10
IDENTIFY
Our research hypothesis now becomes:
H1: (p1–p2) > .03
And so our null hypothesis is: H0: (p1–p2) = .03
Since the r.h.s. of H0 is
not zero, it’s a “case 2” type problem
Copyright © 2009 Cengage Learning
Example 13.10
COMPUTE
Click Add-Ins, Data Analysis Plus, Z-Test: 2 Proportions
Copyright © 2009 Cengage Learning
Example 13.10
COMPUTE
A
B
C
1 z-Test: Two Proportions
2
3
Supermarket 1 Supermarket 2
4 Sample Proportions
0.1991
0.1493
5 Observations
904
1038
6 Hypothesized Difference
0.03
7 z Stat
1.14
8 P(Z<=z) one tail
0.1261
9 z Critical one-tail
1.6449
10 P(Z<=z) two-tail
0.2522
11 z Critical two-tail
1.96
Copyright © 2009 Cengage Learning
Example 13.10
INTERPRET
There is not enough evidence to infer that the brightly
colored design outsells the other design by 3% or more.
Copyright © 2009 Cengage Learning
Confidence Interval Estimator
The confidence interval estimator for p1–p2 is given by:
and as you may suspect, its valid when…
Copyright © 2009 Cengage Learning
Example 13.11
To help estimate the difference in profitability, the
Marketing manager in Examples 13.9 and 13.10 would
like to estimate the difference between the two
proportions. A confidence level of 95% is suggested.
Copyright © 2009 Cengage Learning
Example 13.11
COMPUTE
For manual calculations click
Example 13.11 Manual Calculations
For Excel skip to next slide.
Copyright © 2009 Cengage Learning
12.136
Example 13.11
COMPUTE
Click Add-Ins, Data Analysis Plus, Z-Estimate: 2 Proportions
Copyright © 2009 Cengage Learning
Example 13.11
COMPUTE
A
B
C
D
1 z-Estimate: Two Proportions
2
3
Supermarket 1 Supermarket 2
4 Sample Proportions
0.1991
0.1493
5 Observations
904
1038
6
7 LCL
0.0159
8 UCL
0.0837
Copyright © 2009 Cengage Learning
Identifying Factors…
Factors that identify the z-test and estimator for p1–p2
Copyright © 2009 Cengage Learning
Download