Chapter 7: The One-Sample t Test and Interval Estimation

advertisement
Chapter 6: The One-Sample t Test and Interval Estimation
1. The t Distributions and the Test for the Mean of One Population When σ Is Not Known
Considering that the whole point of inferential statistics is to make inferences about populations,
it shouldn’t be surprising that you often don’t have the standard deviation of the population to
use in the z-score formula. But fret not, mini-statisticians, since there is a mathematical fix to this
dilemma. Instead of using a population standard deviation (σ), you just need to find the standard
deviation of your sample (s). Once you’ve done that, plug your numbers into the following
formula:
Keep in mind, the distributions that this formula yields won’t be normally distributed, yet will
somewhat resemble that shape—still bell-shaped, but it doesn’t precisely look like the Liberty
Bell anymore. (Though keep in mind, as your sample size (N) increases, the curve will look more
and more similar to the normal distribution.) And just to add another notion to blow your mind,
each sample size has its very own curve. But again, you don’t need to channel that calculus from
yesteryear; some kind mathematicians have created a handy little t distribution table that gives
you the critical values you need. As a reminder, an edited portion of the t table is shown below:
Level of Significance for a Two-Tailed Test
df
...
t.05
t.01
...
.
4
2.78 4.60
.
10
2.23 3.17
11
2.20 3.11
12
2.18 3.06
13
2.16 3.01
14
2.15 2.98
15
2.13 2.95
.
40
2.02 2.70
.
120
1.98 2.62
.
∞
1.96 2.58
You should note that the t table has a column on the left-hand side, corresponding to degrees of
freedom (df). What is the value of df for a one-sample t test, you ask? Just df = N–1. Yup, as
simple as that.
So let’s find a value for t and then look up whether it would be out in the critical range (a good
thing if you want to reject the null) for α = .05 and α = .01 (for your convenience, you can always
assume that two-tailed tests are being performed in this study guide, unless otherwise specified).
Suppose we want to find out if a subgroup of students from Terrific Tower, who are always
known for being the Debbie Downers of the dorm, vary significantly in their happiness ratings
from the rest of the dorm’s population. The statistics are as follows:
μ = 4.6 (The happiness rating of all 120 students.) Please note, we are now using this figure as
the population parameter, because we are measuring the Debbie Downers against the whole
dorm, not the whole world. Also, we are assuming someone misplaced the rest of the dorm’s
data, and thus we don’t know the standard deviation of the population anymore.
= 4.1 (The happiness rating for the 12 Debbie Downeresque students.)
s = .6
N = 12 (df = N – 1 = 11)
Now, plug in the numbers:
(4.1 – 4.6)/(.6/(√12) = -2.887
We are certainly significant at the .05 level (t.05 = ±2.201), but we do not attain significance at
the .01 level (t.01 = ±3.106). So we can reject the null at .05 but must retain it at .01. So yes, those
Debbie Downers are an anomaly, but just how dramatically different are they really? And that’s
where further research and statistics can be used to find out more specific information, which
you’ll learn in later chapters!
Now, you may be surprised that those results didn’t attain significance at .01 when they would
have easily done so in a z-score example (if you recall, z only has to beat 2.58 to be significant at
the .01 level). But don’t forget, those sneaky little degrees of freedom can make a BIG
difference. This is another reason (the size of the SEM is the first) that you must find a
considerably larger difference between a sample and the general population when you have just a
few individuals as part of the study to determine that you’ve truly got something significant
going on!
Keep in mind, the important difference between the t-distributions and the z-distribution is that
the t-distribution tails will be fatter, which implies that the critical values will be a bit higher and
harder to beat. Even if you have a large sample, it is best to report it as beating the t (rather than
the z) critical value, simply so no one thinks you’re trying to give yourself an easier value to
surpass.
Try this example:
1. You want to determine if the number of all-nighters architecture majors pull per year
differs significantly from the general college population (μ = 9). You are given the raw
scores for 15 architecture majors: 25, 42, 35, 10, 2, 40, 12, 16, 10, 4, 3, 0, 29, 14, 6.
2. Interval Estimation and Confidence Intervals versus Null Hypothesis Testing
Sometimes proving that the null should be rejected isn’t enough, as it just tells us that one
particular point is not likely to be the parameter of the data we found. So when you want
additional information, it’s time to turn to confidence intervals (CIs). These bad boys will
provide, with varying amounts of confidence (usually 95% or 99%, your choice), a range of
values within which we believe the parameter we are looking for truly exists. In other words, we
can be pretty darn sure that the actual population mean is in whatever range we determine, and
that is based on this formula:
As an example, let’s suppose that you want to know, with 95% confidence, the mean number of
beers that fraternity brothers force the pledges to drink at all American fraternities during rush
week. We’re going to use the average number of beers consumed by 24 randomly selected
pledges from the U.S. as our sample mean, = 39, s = 8. (Luckily, it’s actually the Beermaster’s
job to take note of this number for each pledge, so we have data to use!) Next, we need to know
our df; since the number of pledges sampled was 24, our df = 24 – 1 = 23. Next, we can look for
the t-value for the CI for 95%, which is 2.069 (the .05, two-tailed critical t for 23 df). Now, we
just plug in the numbers: 39 ± 2.069 (8/√24) = 39 ± 3.38. Usually this is written as 35.62 ≤ μ ≤
42.38, which is saying that we can assume, with 95% accuracy, that the population mean for
number of beers that a pledge drinks in a week during rush falls somewhere between about 35
and 42 beers. And, well, my liver is cringing just thinking about that number!
Keep in mind, the 99% CI would be very similar to calculate; however, you would use the tvalue of 2.787, which would make the range a bit larger. If you think about this for a moment (let
those wheels crank for a bit), it should be fairly intuitive that if you want to be MORE confident
that the mean truly exists within your range, you’ll need to expand it somewhat. So, the 99% CI
comes out to: 34.45 ≤ μ ≤ 43.55. As you can see, the likely beer consumption range is now
roughly 34.5 to 43.5, which means it is now about 9 beers wide, instead of just under 7 beers
wide for the 95% CI.
Now let’s try some examples . . .
2. Every summer, Anthropology 101 takes a class trip to Africa to learn firsthand about the
roots of humankind. Mary really wants to go on the trip but is worried about how much it
will cost her to go. She wants to be 95% confident that the trip will cost her no more than
$975, and she is going to base her decision on the amount of money the 28 students from the
class last year spent ( = 905, s = 145, N = 28). Can she be 95% sure? Could she also be
99% sure?
3) Jake is a biochemistry major, but he really wants to take the course Human Sexuality this
semester. He wants to be 95% confident that he has enough time (only 70 free hours to offer)
to study for the course to ensure he won’t fail (and he’ll even learn a thing or two!). He is
using data from the past year the course was taught ( = 60, s = 12, N = 16) to figure out the
maximum and minimum amount of hours other students have put into the course. Can Jake
take the course?
3. The Standard Error of a Proportion
Let’s imagine that your friend is a math major, and she says that there seems to be an unusually
large number of left-handed students among her fellow math majors (they never have enough
left-handed desks, etc.). She searched for estimates of the percentage of lefties in the U.S.
population, and found them to range from 7% to 13%, so she decided that a population
proportion (π) of .1 was reasonable for purposes of comparison with the math majors. If she
surveys all of the 120 math majors at her university, and finds that the proportion of left-handed
students is .15 (i.e., 18 of the 120), can she conclude that the proportion of lefties among math
majors is actually more than the national average of .10? She can answer this question by
computing a simple z score to compare the math majors with the population – a z score for
proportions, that is. Here is what that z score formula looks like:
P 
z
 (1   )
N
Now let’s fill in the values for this example, and compute the z score.
z
.15  .1

.1(1  .1)
120
.05
.05

 1.83
.00075 .0274
You can see from Table A in your text that the area beyond a z score of 1.83 is 50 – 46.64 =
3.36, so the one-tailed p value is 3.36/100 = .0336. That makes her fellow math majors pretty
unusual as a group, and she could reject the null hypothesis that π = .1 for math majors, if she
could justify a one-tailed test. Unfortunately, her two-tailed p value is 2 * .0336 = .0672, which
is not significant at the .05 level with the more conservative two-tailed test. (And like many math
majors, she’s a bit conservative, though we do not have any data to back up that observation.)
Now, here’s an example for you to work out:
4. A friend thinks that there are an unusual number of women with blond hair among the
psychology majors at his college. He conducts a survey and finds that 36 of the 160 female
psych majors at his school are naturally blond. If the percentage of blondes in the female collegeaged population of the U.S. is 15%, is he correct that there is an unusual percentage of blondes in
his department? Can he reject the null hypothesis that the percentage of blondes in his
department is no different from the national average?
Answers to Exercises
1.
= 16.5333; s = 14.182; μ = 9; N = 15; s /√N = 3.662; t = 7.533/3.662 = 2.06 < t.05 (14) =
2.145, not sig.
2. Mary can be confident that it won’t be more than $975 because the 95% CI upper limit is
961.23: t (27) = 2.052; CI = 905 ± 2.052 (145/√28) = 905 ± 56.2298 = 848.77 ≤ μ ≤ 961.23.
However, she cannot be sure at the 99% level, as CI = 905 ± 2.771 (145/√28) = 905 ± 75.932=
829.07 ≤ μ ≤ 980.93.
3. Jake can take the course, because the 95% CI upper limit is 66.39, which is less than 70
hours: t (15) = 2.131; CI = 60 ±2.131 (12/√16) = 53.61 ≤ μ ≥ 66.39. He can even be 99%
confident, too! (51.16 ≤ μ ≤ 68.84)
4. For this example, π = 15/100 = .15; N = 160; and P = 36/160 = .225.
z
.225  .15
.075
.075


 2.68
.15(.85)
.0008 .028
160
The two-tailed p value corresponding to z = 2.68 is .0074, so p < .01. The null hypothesis can be
rejected.
Download