3680 Lecture 12

advertisement
Math 3680
Lecture #12
The Central Limit Theorem
for Sample Sums and
Sample Averages
In the previous lecture, we introduced the central limit
theorem when drawing from a box containing
exclusively “0”s and “1”s.
We now generalize this technique to other kinds of
populations: heights of college students, incomes, or
anything else which is not dichotomous.
Linear Combinations of
Random Variables
Theorem. For any random variables X and Y and
any constants a, b, and c, we have
E( a X + b Y + c ) = a E( X ) + b E ( Y ) + c.
Note: X and Y do not have to be independent.
Definition: Random variables X and Y are independent if for
all constants a and b,
P(X ≤ a and Y ≤ b) = P(X ≤ a) P(Y ≤ b)
For discrete random variables, this is the same as saying that
for all x and all y,
P(X=x and Y=y) = P(X =x) P(Y=y)
Theorem. If X and Y are independent, then
E( X Y ) = E( X ) E ( Y ).
Proof. We show the discrete case; the continuous case
is analogous (using integrals instead of summations).
E ( X Y )   x y P( X  x, Y  y )
x
y
  x y P( X  x) P(Y  y )
x
y


  x P( X  x)  y P(Y  y ) 
x
 y

  x P( X  x) E (Y )
x
 E ( X ) E (Y ).
Theorem. For any random variable X and any
constant a, we have
Var( a X ) = a2 Var( X ),
SD( a X ) = |a| SD( X ).
(Remember this from earlier?)
Theorem. For any independent random variables X
and Y, we have
Var( X + Y ) = Var( X ) + Var ( Y ).
Proof.
2
2
Var ( X Y )  E ([ X  Y ] )  [ E ( X  Y )]
 E ( X 2  2 XY  Y 2 )  [ E ( X )  E (Y )]2
 E ( X )  2 E ( XY )  E (Y )
2
2
 [ E ( X )]  2 E ( X ) E (Y )  [ E (Y )]
 Var ( X )  Var (Y ).
2
2
Note: The assumption of independence is critical in
the last theorem. For example, if X = Y, then
Var( X + X ) = Var( 2 X ) = 4 Var( X )
 Var( X ) + Var( X )
Example. Let X and Y be independent r.v.’s with
X ~ Binomial(8, 0.4) and Y ~ Binomial(8, 0.4).
Find E( X 2 ) and E( X Y ).
Example. Let S be the sum of 5 thrown dice.
Find E( S ) and SD( S ).
The Central Limit Theorem
(or the Law of Averages
for the Sample Sum)
The normal approximation may be used for other random
variables beside binomial ones.
Theorem. Suppose random variables X1, X2, …, Xn are drawn
with replacement from a large population with mean m and
standard deviation s. Let
SUM = X1 + X2 + …+ Xn.
Then E(SUM) = n m and SD(SUM) = s n .
(Why?)
Furthermore, if n is “large,” then we may accurately
approximate probabilities of SUM by converting to standard
units and using the normal curve.
The normal approximation may be used for other random
variables beside binomial ones.
Theorem. Suppose random variables X1, X2, …, Xn are drawn
without replacement from a population of size N with mean m
and standard deviation s. Let
SUM = X1 + X2 + …+ Xn.
N n
Then E(SUM) = n m and SD(SUM) = s n
. (Why?)
N 1
Furthermore, if n is “large,” then we may accurately
approximate probabilities of SUM by converting to standard
units and using the normal curve.
Question: How large is large “enough” for the normal
curve to be applicable? The answer is, It depends. If
the box itself follows the normal distribution exactly,
then so will SUM, no matter what the value of n is.
However, this trivial case rarely happens in practice.
For a more typical example, let’s look at
P(X = 1) = 1/3
P(X = 2) = 1/3
P(X = 3) = 1/3
Sum of 1 Random Variables
0.3
0.25
0.2
0.15
0.1
0.05
1
2
3
Sum of 2 Random Variables
0.3
0.25
0.2
0.15
0.1
0.05
1
2
3
4
5
6
Sum of 3 Random Variables
0.25
0.2
0.15
0.1
0.05
2
4
6
8
Sum of 4 Random Variables
0.2
0.15
0.1
0.05
2
4
6
8
10
12
Sum of 5 Random Variables
0.2
0.15
0.1
0.05
2.5
5
7.5
10
12.5
15
Sum of 10 Random Variables
0.14
0.12
0.1
0.08
0.06
0.04
0.02
10
15
20
25
30
Sum of 15 Random Variables
0.12
0.1
0.08
0.06
0.04
0.02
25
30
35
40
Sum of 20 Random Variables
0.1
0.08
0.06
0.04
0.02
30
35
40
45
50
55
Sum of 25 Random Variables
0.08
0.06
0.04
0.02
40
45
50
55
60
65
Sum of 30 Random Variables
0.08
0.06
0.04
0.02
50
55
60
65
70
75
Sum of 35 Random Variables
0.08
0.06
0.04
0.02
60
70
80
90
Sum of 40 Random Variables
0.06
0.04
0.02
70
80
90
100
Another example: suppose
P(X = 1) = 1/7
P(X = 2) = 1/7
P(X = 5) = 3/7
P(X = 9) = 1/7
P(X = 20) = 1/7
Sum of 1 Random Variables
0.4
0.3
0.2
0.1
5
10
15
20
Sum of 2 Random Variables
0.2
0.15
0.1
0.05
10
20
30
40
Sum of 3 Random Variables
0.12
0.1
0.08
0.06
0.04
0.02
10
20
30
40
50
60
Sum of 4 Random Variables
0.08
0.06
0.04
0.02
20
40
60
80
Sum of 5 Random Variables
0.06
0.05
0.04
0.03
0.02
0.01
20
40
60
80
100
Sum of 6 Random Variables
0.05
0.04
0.03
0.02
0.01
20
40
60
80
100
120
Sum of 7 Random Variables
0.05
0.04
0.03
0.02
0.01
20
40
60
80
100
120
140
Sum of 8 Random Variables
0.04
0.03
0.02
0.01
25
50
75
100
125
150
Sum of 9 Random Variables
0.035
0.03
0.025
0.02
0.015
0.01
0.005
25
50
75
100
125
150
175
Sum of 10 Random Variables
0.03
0.025
0.02
0.015
0.01
0.005
20
40
60
80
100
120
140
Sum of 15 Random Variables
0.02
0.015
0.01
0.005
25
50
75
100
125
150
175
Sum of 20 Random Variables
0.0175
0.015
0.0125
0.01
0.0075
0.005
0.0025
100
150
200
Sum of 25 Random Variables
0.014
0.012
0.01
0.008
0.006
0.004
0.002
100
150
200
250
Sum of 30 Random Variables
0.012
0.01
0.008
0.006
0.004
0.002
100
150
200
250
300
Sum of 35 Random Variables
0.01
0.008
0.006
0.004
0.002
100
150
200
250
300
350
Sum of 40 Random Variables
0.01
0.008
0.006
0.004
0.002
200
250
300
350
400
Sum of 45 Random Variables
0.01
0.008
0.006
0.004
0.002
200
250
300
350
400
450
Sum of 50 Random Variables
0.008
0.006
0.004
0.002
250
300
350
400
450
500
Example. Two hundred tickets are drawn at random
with replacement from the following box of tickets:
1
2
3
4
5
• What is the smallest possible sum? The biggest?
• What is the expected sum?
• Find the probability that the sum of the tickets is
more than 630.
Example: Thirty-six jets wait to take off from an
airport. The average taxi and take-off time for each
jet is 8.5 minutes, with an SD of 2.5 minutes. What
is the probability that the total taxi and take-off time
for the 36 jets is less than 320 minutes?
Example. A gambler makes 1000 column bets at
roulette. The chance of winning on any one play is
12/38. The gambler can either win $2 or lose $1 on
each play. Find the probability that, in total, the
gambler wins at least $0.
Example. A gambler makes 10,000 column bets at
roulette. The chance of winning on any one play is
12/38. The gambler can either win $2 or lose $1 on
each play. Find the probability that, in total, the
gambler wins at least $0.
The Central Limit Theorem
(or the Law of Averages)
for the Sample Mean
Theorem. Suppose random variables X1, X2, …, Xn are drawn
with replacement from a large population with mean m and
standard deviation s. Let
X1  X 2    X n
X
n
Then E( X ) = m and SD(
)=
s
n
.
(Why?)
Furthermore, if n is “large,” then we may accurately
approximate probabilities of
by converting to standard
units and using the normal curve. (As before, this is exact if
the original population follows the normal curve.)
Theorem. Suppose random variables X1, X2, …, Xn are drawn
without replacement from a population of size N with mean m
and standard deviation s. Let
X1  X 2    X n
X
n
Then E( X ) = m and SD(
)=
s
n
N n
.
N 1
(Why?)
Furthermore, if n is “large,” then we may accurately
approximate probabilities of
by converting to standard
units and using the normal curve. (As before, this is exact if
the original population follows the normal curve.)
Example: The cookie machine at Chips Ahoy adds
a random number of chips to each cookie. The
number of chips is a random number with average
28.5 and SD 5.3. Find the probability that, in a bag
of 50 cookies, the average number of chips per
cookie is at least 30.
Example: A computer generates 100 random
numbers between 0 and 1 (presumably evenly). Find
the probability that the average of these numbers is
between 0.48 and 0.49.
Review of Law of Averages
Population
Statistic
(Parameter)
Use the Law of
Averages for
Sample...
Dichotomous (p)
(0-1 box)
K
Count
p
Proportion
SUM
Sum
Quantitative (m)
Mean
COUNT (0/1 box)
E(K) = n p
SD(K) = np (1  p )
PROPORTION (0/1 box)
E(P) = p
SD(P) = p (1  p )
n
SUM
E(SUM) = n m
SD(SUM) = s n
AVERAGE
E( X ) = m
SD(X ) = s
n
For all four types of problems, multiply the
SD by the finite population correction factor
if drawing without replacement.
N n
N 1
Download