Applied Statistics I Liang Zhang July 8, 2008

advertisement
Applied Statistics I
Liang Zhang
Department of Mathematics, University of Utah
July 8, 2008
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
1 / 15
Distribution for Sample Mean
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
2 / 15
Distribution for Sample Mean
Example (Problem 54)
Suppose the sediment density (g/cm) of a randomly selected specimen
from a certain region is normally distributed with mean 2.65 and standard
deviation .85 (suggested in “Modeling Sediment and Water Column
Interactions for Hydrophobic Pollutants”, Water Research, 1984:
1169-1174).
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
2 / 15
Distribution for Sample Mean
Example (Problem 54)
Suppose the sediment density (g/cm) of a randomly selected specimen
from a certain region is normally distributed with mean 2.65 and standard
deviation .85 (suggested in “Modeling Sediment and Water Column
Interactions for Hydrophobic Pollutants”, Water Research, 1984:
1169-1174).
a. If a random sample of 25 specimens is selected, what is the
probability that the sample average sediment density is at most 3.00?
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
2 / 15
Distribution for Sample Mean
Example (Problem 54)
Suppose the sediment density (g/cm) of a randomly selected specimen
from a certain region is normally distributed with mean 2.65 and standard
deviation .85 (suggested in “Modeling Sediment and Water Column
Interactions for Hydrophobic Pollutants”, Water Research, 1984:
1169-1174).
a. If a random sample of 25 specimens is selected, what is the
probability that the sample average sediment density is at most 3.00?
b. How large a sample size would be required to ensure that the above
probability is at least .99?
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
2 / 15
Distribution for Sample Mean
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
3 / 15
Distribution for Sample Mean
The Central Limit Theorem (CLT)
Let X1 , X2 , . . . , Xn be a random sample from a distribution with mean
value µ and standard deviation σ. Then if n is sufficiently large, X has
approximately a normal distribution with mean value µ and standard
√
deviation σ/ n, and T0 also has approximately a normal distribution with
√
mean value nµ and standard deviation nσ. The larger the value of n, the
better the approximation.
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
3 / 15
Distribution for Sample Mean
The Central Limit Theorem (CLT)
Let X1 , X2 , . . . , Xn be a random sample from a distribution with mean
value µ and standard deviation σ. Then if n is sufficiently large, X has
approximately a normal distribution with mean value µ and standard
√
deviation σ/ n, and T0 also has approximately a normal distribution with
√
mean value nµ and standard deviation nσ. The larger the value of n, the
better the approximation.
Remark:
1. As long as n is sufficiently large, CLT is applicable no matter Xi ’s are
discrete random variables or continuous random variables.
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
3 / 15
Distribution for Sample Mean
The Central Limit Theorem (CLT)
Let X1 , X2 , . . . , Xn be a random sample from a distribution with mean
value µ and standard deviation σ. Then if n is sufficiently large, X has
approximately a normal distribution with mean value µ and standard
√
deviation σ/ n, and T0 also has approximately a normal distribution with
√
mean value nµ and standard deviation nσ. The larger the value of n, the
better the approximation.
Remark:
1. As long as n is sufficiently large, CLT is applicable no matter Xi ’s are
discrete random variables or continuous random variables.
2. How large should n be such that CLT is applicable?
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
3 / 15
Distribution for Sample Mean
The Central Limit Theorem (CLT)
Let X1 , X2 , . . . , Xn be a random sample from a distribution with mean
value µ and standard deviation σ. Then if n is sufficiently large, X has
approximately a normal distribution with mean value µ and standard
√
deviation σ/ n, and T0 also has approximately a normal distribution with
√
mean value nµ and standard deviation nσ. The larger the value of n, the
better the approximation.
Remark:
1. As long as n is sufficiently large, CLT is applicable no matter Xi ’s are
discrete random variables or continuous random variables.
2. How large should n be such that CLT is applicable?
Generally, if n > 30, CLT can be used.
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
3 / 15
Distribution for Sample Mean
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
4 / 15
Distribution for Sample Mean
Example (Problem 49)
There are 40 students in an elementary statistics class. On the basis of
years of experience, the instructor knows that the time needed to grade a
randomly chosen first examination paper is a random variable with an
expected value of 6 min and a standard deviation of 6 min.
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
4 / 15
Distribution for Sample Mean
Example (Problem 49)
There are 40 students in an elementary statistics class. On the basis of
years of experience, the instructor knows that the time needed to grade a
randomly chosen first examination paper is a random variable with an
expected value of 6 min and a standard deviation of 6 min.
a. If grading times are independent and the instructor begins grading at
6:50pm and grades continuously, what is the (approximate) probability
that he is through grading before the 11:00pm TV news begins?
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
4 / 15
Distribution for Sample Mean
Example (Problem 49)
There are 40 students in an elementary statistics class. On the basis of
years of experience, the instructor knows that the time needed to grade a
randomly chosen first examination paper is a random variable with an
expected value of 6 min and a standard deviation of 6 min.
a. If grading times are independent and the instructor begins grading at
6:50pm and grades continuously, what is the (approximate) probability
that he is through grading before the 11:00pm TV news begins?
b. If the sports report begins at 11:10pm, what is the probability that he
misses part of the report if he waits unitl grading is done before
turning on the TV?
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
4 / 15
Distribution for Sample Mean
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
5 / 15
Distribution for Sample Mean
The original version of CLT
The Central Limit Theorem (CLT)
Let X1 , X2 , . . . be a sequence of i.i.d. random variables from a
distribution with mean value µ and standard deviation σ. Define random
variables
Pn
Xi − nµ
Yn = i=1√
for n = 1, 2, . . .
nσ
Then as n → ∞, Yn has approximately a normal distribution.
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
5 / 15
Distribution for Sample Mean
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
6 / 15
Distribution for Sample Mean
Corollary
Let X1 , X2 , . . . , Xn be a random sample from a distribution for which only
positive values are possible [P(Xi > 0) = 1]. Then if n is sufficiently large,
the product Y = X1 X2 · · · Xn has approximately a lognormal distribution.
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
6 / 15
Distribution for Linear Combinations
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
7 / 15
Distribution for Linear Combinations
Proposition
Let X1 , X2 , . . . , Xn have mean values µ1 , µ2 , . . . , µn , respectively, and
variances σ12 , σ22 , . . . , σn2 , respectively.
1.Whether or not the Xi s are independent,
E (a1 X1 + a2 X2 + · · · + an Xn ) = a1 E (X1 ) + a2 E (X2 ) + · · · + an E (Xn )
= a1 µ1 + a2 µ2 + · · · + an µn
2. If X1 , X2 , . . . , Xn are independent,
V (a1 X1 + a2 X2 + · · · + an Xn ) = a12 V (X1 ) + a22 V (X2 ) + · · · + an2 V (Xn )
= a1 σ12 + a2 σ22 + · · · + an σn2
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
7 / 15
Distribution for Linear Combinations
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
8 / 15
Distribution for Linear Combinations
Proposition (Continued)
Let X1 , X2 , . . . , Xn have mean values µ1 , µ2 , . . . , µn , respectively, and
variances σ12 , σ22 , . . . , σn2 , respectively.
3. More generally, for any X1 , X2 , . . . , Xn
V (a1 X1 + a2 X2 + · · · + an Xn ) =
n X
n
X
ai aj Cov (Xi , Xj )
i=1 j=1
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
8 / 15
Distribution for Linear Combinations
Proposition (Continued)
Let X1 , X2 , . . . , Xn have mean values µ1 , µ2 , . . . , µn , respectively, and
variances σ12 , σ22 , . . . , σn2 , respectively.
3. More generally, for any X1 , X2 , . . . , Xn
V (a1 X1 + a2 X2 + · · · + an Xn ) =
n X
n
X
ai aj Cov (Xi , Xj )
i=1 j=1
We call a1 X1 + a2 X2 + · · · + an Xn a linear combination of the Xi ’s.
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
8 / 15
Distribution for Linear Combinations
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
9 / 15
Distribution for Linear Combinations
Example (Problem 64)
Suppose your waiting time for a bus in the morning is uniformly
distributed on [0,8], whereas waiting time in the evening is uniformly
distributed on [0,10] independent of morning waiting time.
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
9 / 15
Distribution for Linear Combinations
Example (Problem 64)
Suppose your waiting time for a bus in the morning is uniformly
distributed on [0,8], whereas waiting time in the evening is uniformly
distributed on [0,10] independent of morning waiting time.
a. If you take the bus each morning and evening for a week, what is your
total expected waiting time?
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
9 / 15
Distribution for Linear Combinations
Example (Problem 64)
Suppose your waiting time for a bus in the morning is uniformly
distributed on [0,8], whereas waiting time in the evening is uniformly
distributed on [0,10] independent of morning waiting time.
a. If you take the bus each morning and evening for a week, what is your
total expected waiting time?
b. What is the variance of your total waiting time?
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
9 / 15
Distribution for Linear Combinations
Example (Problem 64)
Suppose your waiting time for a bus in the morning is uniformly
distributed on [0,8], whereas waiting time in the evening is uniformly
distributed on [0,10] independent of morning waiting time.
a. If you take the bus each morning and evening for a week, what is your
total expected waiting time?
b. What is the variance of your total waiting time?
c. What are the expected value and variance of the difference between
morning and evening waiting times on a given day?
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
9 / 15
Distribution for Linear Combinations
Example (Problem 64)
Suppose your waiting time for a bus in the morning is uniformly
distributed on [0,8], whereas waiting time in the evening is uniformly
distributed on [0,10] independent of morning waiting time.
a. If you take the bus each morning and evening for a week, what is your
total expected waiting time?
b. What is the variance of your total waiting time?
c. What are the expected value and variance of the difference between
morning and evening waiting times on a given day?
d. What are the expected value and variance of the difference between
total morning waiting time and total evening waiting time on a
particular week?
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
9 / 15
Distribution for Linear Combinations
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
10 / 15
Distribution for Linear Combinations
Corollary
E (X1 − X2 ) = E (X1 ) − E (X2 ) and, if X1 and X2 are independent,
V (X1 − X2 ) = V (X1 ) + V (X2 ).
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
10 / 15
Distribution for Linear Combinations
Corollary
E (X1 − X2 ) = E (X1 ) − E (X2 ) and, if X1 and X2 are independent,
V (X1 − X2 ) = V (X1 ) + V (X2 ).
Proposition
If X1 , X2 , . . . , Xn are independent, normally distributed rv’s (with possibly
different means and/or variances), then any linear combination of the Xi s
also has a normal distribution. In particular, the difference X1 − X2
between two independent, normally distributed variables is itself normally
distributed.
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
10 / 15
Distribution for Linear Combinations
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
11 / 15
Distribution for Linear Combinations
Example (Problem 62)
Manufacture of a certain component requires three different maching
operations. Machining time for each operation has a normal distribution,
and the three times are independent of one another. The mean values are
15, 30, and 20min, respectively, and the standard deviations are 1, 2, and
1.5min, respectively.
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
11 / 15
Distribution for Linear Combinations
Example (Problem 62)
Manufacture of a certain component requires three different maching
operations. Machining time for each operation has a normal distribution,
and the three times are independent of one another. The mean values are
15, 30, and 20min, respectively, and the standard deviations are 1, 2, and
1.5min, respectively.
What is the probability that it takes at most 1 hour of machining time to
produce a randomly selected component?
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
11 / 15
Point Estimation
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
12 / 15
Point Estimation
Example (a variant of Problem 62, Ch5)
Manufacture of a certain component requires three different maching
operations. The total time for manufacturing one such component is
known to have a normal distribution. However, the mean µ and variance
σ 2 for the normal distribution are unknown. If we did an experiment in
which we manufactured 10 components and record the operation time,
and the sample time is given as following:
1
2
3
4
5
time 63.8 60.5 65.3 65.7 61.9
6
7
8
9
10
time 68.2 68.1 64.8 65.8 65.4
What can we say about the population mean µ and population variance
σ2?
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
12 / 15
Point Estimation
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
13 / 15
Point Estimation
Example (a variant of Problem 64, Ch5)
Suppose the waiting time for a certain bus in the morning is uniformly
distributed on [0, θ], where θ is unknown. If we record 10 waiting times as
follwos:
1
2
3
4
5
time 7.6 1.8 4.8 3.9 7.1
6
7
8
9
10
time 6.1 3.6 0.1 6.5 3.5
What can we say about the parameter θ?
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
13 / 15
Point Estimation
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
14 / 15
Point Estimation
Definition
A point estimate of a parameter θ is a single number that can be
regarded as a sensible value for θ. A point estimate is obtained by
selecting a suitable statistic and computing its value from the given sample
data. The selected statistic is called the point estimator of θ.
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
14 / 15
Point Estimation
Definition
A point estimate of a parameter θ is a single number that can be
regarded as a sensible value for θ. A point estimate is obtained by
selecting a suitable statistic and computing its value from the given sample
data. The selected statistic is called the point estimator of θ.
P
e.g. X = 10
i=1 Xi /10 is a point estimator for µ for the normal distribution
example.
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
14 / 15
Point Estimation
Definition
A point estimate of a parameter θ is a single number that can be
regarded as a sensible value for θ. A point estimate is obtained by
selecting a suitable statistic and computing its value from the given sample
data. The selected statistic is called the point estimator of θ.
P
e.g. X = 10
i=1 Xi /10 is a point estimator for µ for the normal distribution
example.
The largest sample data X10,10 is a point estimator for θ for the uniform
distribution example.
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
14 / 15
Methods of Point Estimation
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
15 / 15
Methods of Point Estimation
Definition
Let X1 , X2 , . . . , Xn be a random sample from a distribution with pmf or
pdf f (x). For k = 1, 2, 3, . . . , the kth population moment, or kth
moment
of the distribution f (x), is E (X k ). The kth sample moment
P
is n1 ni=1 Xik .
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
15 / 15
Methods of Point Estimation
Definition
Let X1 , X2 , . . . , Xn be a random sample from a distribution with pmf or
pdf f (x). For k = 1, 2, 3, . . . , the kth population moment, or kth
moment
of the distribution f (x), is E (X k ). The kth sample moment
P
is n1 ni=1 Xik .
Definition
Let X1 , X2 , . . . , Xn be a random sample from a distribution with pmf or
pdf f (x; θ1 , . . . , θm ), where θ1 , . . . , θm are parameters whose values are
unknown. Then the moment estimators θ̂1 , . . . , θ̂m are obtained by
equating the first m sample moments to the corresponding first m
population moments and solving for θ1 , . . . , θm .
Liang Zhang (UofU)
Applied Statistics I
July 8, 2008
15 / 15
Download