Definition

advertisement
Normal Distribution
Normal Distribution
Definition
A continuous rv X is said to have a normal distribution with
parameter µ and σ (µ and σ 2 ), where −∞ < µ < ∞ and σ > 0, if
the pdf of X is
f (x; µ, σ) = √
1
2
2
e −(x−µ) /(2σ )
2πσ
We use the notation X ∼ N(µ, σ 2 ) to denote that X is rormally
distributed with parameters µ and σ 2 .
Normal Distribution
Proposition
For X ∼ N(µ, σ 2 ), we have
E (X ) = µ and V (X ) = σ 2
Normal Distribution
Definition
The normal distribution with parameter values µ = 0 and σ = 1 is
called the standard normal distribution. A random variable
having a standard normal distribution is called a standard normal
random variable and will be denoted by Z . The pdf of Z is
1
2
f (z; 0, 1) = √ e −z /2
2π
−∞<z <∞
The graph of f (z; 0, 1) is called
R z the standard normal (or z) curve.
The cdf of Z is P(Z ≤ z) = −∞ f (y ; 0, 1)dy , which we will
denote by Φ(z).
Normal Distribution
Proposition
If X has a normal distribution with mean µ and stadard deviation
σ, then
X −µ
Z=
σ
has a standard normal distribution. Thus
a−µ
b−µ
≤Z ≤
)
σ
σ
b−µ
a−µ
= Φ(
) − Φ(
)
σ
σ
P(a ≤ X ≤ b) = P(
P(X ≤ a) = Φ(
a−µ
)
σ
P(X ≥ b) = 1 − Φ(
b−µ
)
σ
Normal Distribution
Normal Distribution
Proposition
{(100p)th percentile for N(µ, σ 2 )} =
µ + {(100p)th percentile for N(0, 1)} · σ
Normal Distribution
Normal Distribution
Proposition
Let X be a binomial rv based on n trials with success probability p.
Then if the binomial probability histogram is not too skewed, X has
√
approximately a normal distribution with µ = np and σ = npq,
where q = 1 − p. In particular, for x = a posible value of X ,
area under the normal curve
P(X ≤ x) = B(x; n, p) ≈
to the left of x+0.5
x+0.5 − np
= Φ( √
)
npq
In practice, the approximation is adequate provided that both
np ≥ 10 and nq ≥ 10, since there is then enough symmetry in the
underlying binomial distribution.
Normal Distribution
Normal Distribution
A graphical explanation for
P(X ≤ x) = B(x; n, p) ≈
= Φ(
area under the normal curve
to the left of x+0.5
x+0.5 − np
)
√
npq
Normal Distribution
A graphical explanation for
P(X ≤ x) = B(x; n, p) ≈
= Φ(
area under the normal curve
to the left of x+0.5
x+0.5 − np
)
√
npq
Normal Distribution
Normal Distribution
Example (Problem 54)
Suppose that 10% of all steel shafts produced by a certain process
are nonconforming but can be reworked (rather than having to be
scrapped). Consider a random sample of 200 shafts, and let X
denote the number among these that are nonconforming and can
be reworked. What is the (approximate) probability that X is
between 15 and 25 (inclusive)?
Normal Distribution
Example (Problem 54)
Suppose that 10% of all steel shafts produced by a certain process
are nonconforming but can be reworked (rather than having to be
scrapped). Consider a random sample of 200 shafts, and let X
denote the number among these that are nonconforming and can
be reworked. What is the (approximate) probability that X is
between 15 and 25 (inclusive)?
In this problem n = 200, p = 0.1 and q = 1 − p = 0.9. Thus
np = 20 > 10 and nq = 180 > 10
Normal Distribution
Example (Problem 54)
Suppose that 10% of all steel shafts produced by a certain process
are nonconforming but can be reworked (rather than having to be
scrapped). Consider a random sample of 200 shafts, and let X
denote the number among these that are nonconforming and can
be reworked. What is the (approximate) probability that X is
between 15 and 25 (inclusive)?
In this problem n = 200, p = 0.1 and q = 1 − p = 0.9. Thus
np = 20 > 10 and nq = 180 > 10
P(15 ≤ X ≤ 25) = Bin(25; 200, 0.1) − Bin(14; 200, 0.1)
25 + 0.5 − 20
15 + 0.5 − 20
≈ Φ( √
) − Φ( √
)
200 · 0.1 · 0.9
200 · 0.1 · 0.9
= Φ(0.3056) − Φ(−0.2500)
= 0.6217 − 0.4013
= 0.2204
Exponential Distribution
Exponential Distribution
Definition
X is said to have an exponential distribution with parameter
λ(λ > 0) if the pdf of X is
(
λe −λx x ≥ 0
f (x; λ) =
0
otherwise
Exponential Distribution
Exponential Distribution
Proposition
If X ∼ EXP(λ), then
E (X ) =
1
λ
and
V (X ) =
1
λ2
And the cdf for X is
(
1 − e −λx
F (x; λ) =
0
x ≥0
x <0
Exponential Distribution
Exponential Distribution
“Memoryless” Property
P(X ≥ t) = P(X ≥ t + t0 | X ≥ t0 )
for any positive t and t0 .
Exponential Distribution
“Memoryless” Property
P(X ≥ t) = P(X ≥ t + t0 | X ≥ t0 )
for any positive t and t0 .
In words, the distribution of additional lifetime is exactly the same
as the original distribution of lifetime, so at each point in time the
component shows no effect of wear. In other words, the
distribution of remaining lifetime is independent of current age.
Probability Plot
Probability Plot
Example:
There is a machine available for cutting corks intended for use in
wine bottles. We want to find out the distribution of the diameters
of the corks produced by that machine. Assume we have 10
samples produced by that machine and the diameters is recorded
as following:
3.0879
2.6030
3.2546
3.5931
2.8970
3.1253
2.7377
2.4756
2.7740
2.5133
Probability Plot
Probability Plot
Probability Plot
Probability Plot
3.0879
2.6030
3.2546
3.5931
2.8970
3.1253
2.7377
2.4756
2.7740
2.5133
Probability Plot
3.0879
2.6030
3.2546
3.5931
2.8970
3.1253
2.7377
2.4756
2.7740
2.5133
Probability Plot
Probability Plot
Sample Percentile
Probability Plot
Sample Percentile
Recall: The (100p)th percentile of the distribution of a
continuous rv X , denoted by η(p), is defined by
R η(p)
p = F (η(p)) = −∞ f (y )dy .
Probability Plot
Sample Percentile
Recall: The (100p)th percentile of the distribution of a
continuous rv X , denoted by η(p), is defined by
R η(p)
p = F (η(p)) = −∞ f (y )dy .
In words, the (100p)th percentile η(p) is the X value such that
there are 100p% X values below η(p).
Probability Plot
Sample Percentile
Recall: The (100p)th percentile of the distribution of a
continuous rv X , denoted by η(p), is defined by
R η(p)
p = F (η(p)) = −∞ f (y )dy .
In words, the (100p)th percentile η(p) is the X value such that
there are 100p% X values below η(p).
Similarly, we can define sample percentile in the same manner,
i.e. the (100p)th percentile xp is the value such that there are
100p% sample values below xp .
Probability Plot
Sample Percentile
Recall: The (100p)th percentile of the distribution of a
continuous rv X , denoted by η(p), is defined by
R η(p)
p = F (η(p)) = −∞ f (y )dy .
In words, the (100p)th percentile η(p) is the X value such that
there are 100p% X values below η(p).
Similarly, we can define sample percentile in the same manner,
i.e. the (100p)th percentile xp is the value such that there are
100p% sample values below xp .
Unfortunately, xp may not be a sample value for some p.
Probability Plot
Sample Percentile
Recall: The (100p)th percentile of the distribution of a
continuous rv X , denoted by η(p), is defined by
R η(p)
p = F (η(p)) = −∞ f (y )dy .
In words, the (100p)th percentile η(p) is the X value such that
there are 100p% X values below η(p).
Similarly, we can define sample percentile in the same manner,
i.e. the (100p)th percentile xp is the value such that there are
100p% sample values below xp .
Unfortunately, xp may not be a sample value for some p.
e.g. for the previous example, what is the 35th percentile for the
ten sample values?
Probability Plot
Probability Plot
Definition
Assume we have a sample with size n. Order the n sample
observations from smallest to largest. Then the ith smallest
observation in the list is taken to be the [100(i − 0.5)/n]th sample
percentile.
Probability Plot
Definition
Assume we have a sample with size n. Order the n sample
observations from smallest to largest. Then the ith smallest
observation in the list is taken to be the [100(i − 0.5)/n]th sample
percentile.
Remark:
1. Why “i − 0.5”?
Probability Plot
Definition
Assume we have a sample with size n. Order the n sample
observations from smallest to largest. Then the ith smallest
observation in the list is taken to be the [100(i − 0.5)/n]th sample
percentile.
Remark:
1. Why “i − 0.5”? We regard the sample observation as being half
in the lower group and half in the upper group.
Probability Plot
Definition
Assume we have a sample with size n. Order the n sample
observations from smallest to largest. Then the ith smallest
observation in the list is taken to be the [100(i − 0.5)/n]th sample
percentile.
Remark:
1. Why “i − 0.5”? We regard the sample observation as being half
in the lower group and half in the upper group.
e.g. if n = 9, then the sample median is the 5th largest
observation and this observation is regarded as two parts: one in
the lower half and one in the upper half.
Probability Plot
Definition
Assume we have a sample with size n. Order the n sample
observations from smallest to largest. Then the ith smallest
observation in the list is taken to be the [100(i − 0.5)/n]th sample
percentile.
Remark:
1. Why “i − 0.5”? We regard the sample observation as being half
in the lower group and half in the upper group.
e.g. if n = 9, then the sample median is the 5th largest
observation and this observation is regarded as two parts: one in
the lower half and one in the upper half.
2. Once the percentage values 100(i − 0.5)/n(i = 1, 2, . . . , n) have
been calculated, sample percentiles corresponding to intermediate
percentages can be obtained by linear interpolation.
Probability Plot
Probability Plot
Example: for the previous example, the [100(i − 0.5)/n]th sample
percentile is tabulated as following:
2.4756
2.5133
2.6030
100(1-.5)/10 = 5%
100(2-.5)/10 = 15% 100(3-.5)/10 = 25%
2.7377
2.7740
100(4-.5)/10 = 35% 100(5-.5)/10 = 45%
2.8970
3.0879
3.1253
100(6-.5)/10 = 55% 100(7-.5)/10 = 65% 100(8-.5)/10 = 75%
3.2546
3.5931
100(9-.5)/10 = 85% 100(10-.5)/10 = 95%
Probability Plot
Example: for the previous example, the [100(i − 0.5)/n]th sample
percentile is tabulated as following:
2.4756
2.5133
2.6030
100(1-.5)/10 = 5%
100(2-.5)/10 = 15% 100(3-.5)/10 = 25%
2.7377
2.7740
100(4-.5)/10 = 35% 100(5-.5)/10 = 45%
2.8970
3.0879
3.1253
100(6-.5)/10 = 55% 100(7-.5)/10 = 65% 100(8-.5)/10 = 75%
3.2546
3.5931
100(9-.5)/10 = 85% 100(10-.5)/10 = 95%
The 10th percentile would be (2.4756 + 2.5133)/2 = 2.49445
Probability Plot
Probability Plot
Idea for Quantile-Quantile Plot:
1. Determine the “[100(i − 0.5)/n]th sample percentile” for a
given sample.
Probability Plot
Idea for Quantile-Quantile Plot:
1. Determine the “[100(i − 0.5)/n]th sample percentile” for a
given sample.
2. Find the corresponding [100(i − 0.5)/n]th percentile from the
population with the assumed distribution; for example, if the
assumed distribution is standard normal, then find
corresponding [100(i − 0.5)/n]th percentile from the standard
normal distribution.
Probability Plot
Idea for Quantile-Quantile Plot:
1. Determine the “[100(i − 0.5)/n]th sample percentile” for a
given sample.
2. Find the corresponding [100(i − 0.5)/n]th percentile from the
population with the assumed distribution; for example, if the
assumed distribution is standard normal, then find
corresponding [100(i − 0.5)/n]th percentile from the standard
normal distribution.
3. Consider the (population percentile, sample percentile) pairs,
i.e.
[100(i − 0.5)/n]th percentile, ith smallest sample
of the distribution
observation
Probability Plot
Idea for Quantile-Quantile Plot:
1. Determine the “[100(i − 0.5)/n]th sample percentile” for a
given sample.
2. Find the corresponding [100(i − 0.5)/n]th percentile from the
population with the assumed distribution; for example, if the
assumed distribution is standard normal, then find
corresponding [100(i − 0.5)/n]th percentile from the standard
normal distribution.
3. Consider the (population percentile, sample percentile) pairs,
i.e.
[100(i − 0.5)/n]th percentile, ith smallest sample
of the distribution
observation
4. Each pair plotted as a point on a two-dimensional coordinate
system should fall close to a 45◦ line.
Probability Plot
Idea for Quantile-Quantile Plot:
1. Determine the “[100(i − 0.5)/n]th sample percentile” for a
given sample.
2. Find the corresponding [100(i − 0.5)/n]th percentile from the
population with the assumed distribution; for example, if the
assumed distribution is standard normal, then find
corresponding [100(i − 0.5)/n]th percentile from the standard
normal distribution.
3. Consider the (population percentile, sample percentile) pairs,
i.e.
[100(i − 0.5)/n]th percentile, ith smallest sample
of the distribution
observation
4. Each pair plotted as a point on a two-dimensional coordinate
system should fall close to a 45◦ line. Substantial deviations
of the plotted points from a 45◦ line cast doubt on the
assumption that the distribution under consideration is the
correct one.
Probability Plot
Probability Plot
Example 4.29:
The value of a certain physical constant is known to an
experimenter. The experimenter makes n = 10 independent
measurements of this value using a particular measurement device
and records the resulting measurement errors (error = observed
value - true value). These observations appear in the following
table.
Percentage
Sample Observation
Percentage
Sample Observation
5
-1.91
55
0.35
15
-1.25
65
0.72
25
-0.75
75
0.87
35
-0.53
85
1.40
45
0.20
95
1.56
Probability Plot
Example 4.29:
The value of a certain physical constant is known to an
experimenter. The experimenter makes n = 10 independent
measurements of this value using a particular measurement device
and records the resulting measurement errors (error = observed
value - true value). These observations appear in the following
table.
Percentage
Sample Observation
Percentage
Sample Observation
5
-1.91
55
0.35
15
-1.25
65
0.72
25
-0.75
75
0.87
35
-0.53
85
1.40
45
0.20
95
1.56
Is it plausible that the random variable measurement error has
standard normal distribution?
Probability Plot
Probability Plot
We first find the corresponding population distribution percentiles,
in this case, the z percentiles:
Percentage
Sample Observation
z percentile
Percentage
Sample Observation
z percentile
5
-1.91
-1.645
55
0.35
0.126
15
-1.25
-1.037
65
0.72
0.385
25
-0.75
-0.675
75
0.87
0.675
35
-0.53
-0.385
85
1.40
1.037
45
0.20
-0.126
95
1.56
1.645
Probability Plot
Probability Plot
Probability Plot
Probability Plot
What about the first example? We are only interested in whether
the ten sample observations come from a normal distribution.
Probability Plot
What about the first example? We are only interested in whether
the ten sample observations come from a normal distribution.
Recall:
{(100p)th percentile for N(µ, σ 2 )} =
µ + {(100p)th percentile for N(0, 1)} · σ
Probability Plot
What about the first example? We are only interested in whether
the ten sample observations come from a normal distribution.
Recall:
{(100p)th percentile for N(µ, σ 2 )} =
µ + {(100p)th percentile for N(0, 1)} · σ
If µ = 0, then the pairs (σ · [z percentile], observation) fall on a
45◦ line, which has slope 1.
Probability Plot
What about the first example? We are only interested in whether
the ten sample observations come from a normal distribution.
Recall:
{(100p)th percentile for N(µ, σ 2 )} =
µ + {(100p)th percentile for N(0, 1)} · σ
If µ = 0, then the pairs (σ · [z percentile], observation) fall on a
45◦ line, which has slope 1.
Therefore the pairs ([z percentile], observation) fall on a line
passing through (0,0) (i.e., one with y -intercept 0) but having
slope σ rather than 1.
Probability Plot
What about the first example? We are only interested in whether
the ten sample observations come from a normal distribution.
Recall:
{(100p)th percentile for N(µ, σ 2 )} =
µ + {(100p)th percentile for N(0, 1)} · σ
If µ = 0, then the pairs (σ · [z percentile], observation) fall on a
45◦ line, which has slope 1.
Therefore the pairs ([z percentile], observation) fall on a line
passing through (0,0) (i.e., one with y -intercept 0) but having
slope σ rather than 1.
Now for µ 6= 0, the y -intercept is µ instead of 0.
Probability Plot
Probability Plot
Normal Probability Plot
A plot of the n pairs
([100(i − 0.5)/n]th z percentile, ith smallest observation)
on a two-dimensional coordinate system is called a normal
probability plot. If the sample observations are in fact drawn from
a normal distribution with mean value µ and standard deviation σ,
the points should fall close to a straight line with slope σ and
y -intercept µ. Thus a plot for which the points fall close to some
straight line suggests that the assumption of a normal population
distribution is plausible.
Probability Plot
Probability Plot
Percentage
Sample Observation
z percentile
Percentage
Sample Observation
z percentile
5
2.4756
-1.645
55
2.8970
0.126
15
2.5133
-1.037
65
3.0879
0.385
25
2.6030
-0.675
75
3.1253
0.675
35
2.7377
-0.385
85
3.2546
1.037
45
2.7740
-0.126
95
3.5931
1.645
Probability Plot
Probability Plot
A nonnormal population distribution can often be placed in one of
the following three categories:
1. It is symmetric and has “lighter tails” than does a normal
distribution; that is, the density curve declines more rapidly
out in the tails than does a normal curve.
2. It is symmetric and heavy-tailed compared to a normal
distribution.
3. It is skewed.
Probability Plot
Probability Plot
Symmetric and “light-tailed”: e.g. Uniform distribution
Probability Plot
Probability Plot
Symmetric and heavy-tailed: e.g. Cauchy distribution with pdf
f (x) = 1/[π(1 + x 2 )] for −∞ < x < ∞
Probability Plot
Probability Plot
Skewed: e.g. lognormal distribution
Probability Plot
Probability Plot
Some guidances for probability plot for normal distributions
(from the book Fitting Equations to Data (2nd ed.) Daniel,
Cuthbert, and Fed Wood, Wiley, New York, 1980)
Probability Plot
Some guidances for probability plot for normal distributions
(from the book Fitting Equations to Data (2nd ed.) Daniel,
Cuthbert, and Fed Wood, Wiley, New York, 1980)
1. For sample size smaller than 30, there is typically greater
variation in the apperance of the probability plot.
Probability Plot
Some guidances for probability plot for normal distributions
(from the book Fitting Equations to Data (2nd ed.) Daniel,
Cuthbert, and Fed Wood, Wiley, New York, 1980)
1. For sample size smaller than 30, there is typically greater
variation in the apperance of the probability plot.
2. Only for much larger sample sizes does a linear pattern generally
predominate.
Probability Plot
Some guidances for probability plot for normal distributions
(from the book Fitting Equations to Data (2nd ed.) Daniel,
Cuthbert, and Fed Wood, Wiley, New York, 1980)
1. For sample size smaller than 30, there is typically greater
variation in the apperance of the probability plot.
2. Only for much larger sample sizes does a linear pattern generally
predominate.
Therefore, when a plot is based on a small sample size, only a very
substantial departure from linearity should be taken as conclusive
evidence of nonnorality.
Probability Plot
Probability Plot
Definition
Consider a family of probability distributions involving two
parameters, θ1 and θ2 , and let F (x; θ1 , θ2 ) denote the
corresponding cdf’s.
The parameters θ1 and θ2 are said to be location and scale
parameters, respectively, if F (x; θ1 , θ2 ) is a function of
(x − θ1 )/θ2 .
Probability Plot
Definition
Consider a family of probability distributions involving two
parameters, θ1 and θ2 , and let F (x; θ1 , θ2 ) denote the
corresponding cdf’s.
The parameters θ1 and θ2 are said to be location and scale
parameters, respectively, if F (x; θ1 , θ2 ) is a function of
(x − θ1 )/θ2 .
e.g.
1. Normal distributions N(µ, σ): F (x; µ, σ) = Φ( x−µ
σ ).
Probability Plot
Definition
Consider a family of probability distributions involving two
parameters, θ1 and θ2 , and let F (x; θ1 , θ2 ) denote the
corresponding cdf’s.
The parameters θ1 and θ2 are said to be location and scale
parameters, respectively, if F (x; θ1 , θ2 ) is a function of
(x − θ1 )/θ2 .
e.g.
1. Normal distributions N(µ, σ): F (x; µ, σ) = Φ( x−µ
σ ).
2. The extreme value distribution with cdf
F (x; θ1 , θ2 ) = 1 − e −e
(x−θ1 )/θ2
Probability Plot
Probability Plot
For Weibull distribution:
α
F (x; α, β) = 1 − e −(x/β) ,
the parameter β is a scale parameter but α is NOT a location
parameter. α is usually referred to as a shape parameter.
Probability Plot
For Weibull distribution:
α
F (x; α, β) = 1 − e −(x/β) ,
the parameter β is a scale parameter but α is NOT a location
parameter. α is usually referred to as a shape parameter.
Fortunately, if X has a Weibull distribution with shape parameter
α and scale parameter β, then the transformed variable ln(X ) has
an extreme value distribution with location parameter θ1 = ln(β)
and scale parameter θ2 = 1/α.
Probability Plot
Probability Plot
The gamma distribution also has a shape parameter α. However,
there is no transformation h(•) such that h(X ) has a distribution
that depends only on location and scale parameters.
Probability Plot
The gamma distribution also has a shape parameter α. However,
there is no transformation h(•) such that h(X ) has a distribution
that depends only on location and scale parameters.
Thus, before we construct a probability plot, we have to estimate
the shape parameter from the sample data.
Jointly Distributed Random Variables
Jointly Distributed Random Variables
Consider tossing a fair
(1,1) (1,2) (1,3)
(2,1) (2,2) (2,3)
···
···
···
(6,1) (6,2) (6,3)
die twice. Then the outcomes would be
(1,4) (1,5) (1,6)
(2,4) (2,5) (2,6)
···
···
···
(6,4) (6,5) (6,6)
and the probability for each outcome is
1
36 .
Jointly Distributed Random Variables
Consider tossing a fair
(1,1) (1,2) (1,3)
(2,1) (2,2) (2,3)
···
···
···
(6,1) (6,2) (6,3)
die twice. Then the outcomes would be
(1,4) (1,5) (1,6)
(2,4) (2,5) (2,6)
···
···
···
(6,4) (6,5) (6,6)
1
.
and the probability for each outcome is 36
If we define two random variables by X = the outcome of the
first toss and Y = the outcome of the second toss,
Jointly Distributed Random Variables
Consider tossing a fair
(1,1) (1,2) (1,3)
(2,1) (2,2) (2,3)
···
···
···
(6,1) (6,2) (6,3)
die twice. Then the outcomes would be
(1,4) (1,5) (1,6)
(2,4) (2,5) (2,6)
···
···
···
(6,4) (6,5) (6,6)
1
.
and the probability for each outcome is 36
If we define two random variables by X = the outcome of the
first toss and Y = the outcome of the second toss, then
the outcome for this experiment (two tosses) can be describe by
the random pair (X , Y ),
Jointly Distributed Random Variables
Consider tossing a fair
(1,1) (1,2) (1,3)
(2,1) (2,2) (2,3)
···
···
···
(6,1) (6,2) (6,3)
die twice. Then the outcomes would be
(1,4) (1,5) (1,6)
(2,4) (2,5) (2,6)
···
···
···
(6,4) (6,5) (6,6)
1
.
and the probability for each outcome is 36
If we define two random variables by X = the outcome of the
first toss and Y = the outcome of the second toss, then
the outcome for this experiment (two tosses) can be describe by
the random pair (X , Y ), and the probability for any possible value
1
of that random pair (x, y ) is 36
.
Jointly Distributed Random Variables
Jointly Distributed Random Variables
Definition
Let X and Y be two discrete random variables defined on the
sample space S of an experiment. The joint probability mass
function p(x, y ) is defined for each pair of numbers (x, y ) by
p(x, y ) = P(X = x and Y = y )
P P
(It must be the case that p(x, y ) ≥ 0 and x y p(x, y ) = 1.)
For any event A consisting of pairs of (x, y ), the probability
P[(X , Y ) ∈ A] is obtained by summing the joint pmf over pairs in
A:
XX
P[(X , Y ) ∈ A] =
p(x, y )
(x,y )∈A
Jointly Distributed Random Variables
Jointly Distributed Random Variables
Example (Problem 75)
A restaurant serves three fixed-price dinners costing $12, $15, and
$20. For a randomly selected couple dinning at this restaurant, let
X = the cost of the man’s dinner and Y = the cost of
the woman’s dinner. If the joint pmf of X and Y is assumed to
be
y
p(x, y )
12 15 20
12 .05 .05 .10
x
15 .05 .10 .35
20 0 .20 .10
Jointly Distributed Random Variables
Example (Problem 75)
A restaurant serves three fixed-price dinners costing $12, $15, and
$20. For a randomly selected couple dinning at this restaurant, let
X = the cost of the man’s dinner and Y = the cost of
the woman’s dinner. If the joint pmf of X and Y is assumed to
be
y
p(x, y )
12 15 20
12 .05 .05 .10
x
15 .05 .10 .35
20 0 .20 .10
a. What is the probability for them to both have the $12 dinner?
Jointly Distributed Random Variables
Example (Problem 75)
A restaurant serves three fixed-price dinners costing $12, $15, and
$20. For a randomly selected couple dinning at this restaurant, let
X = the cost of the man’s dinner and Y = the cost of
the woman’s dinner. If the joint pmf of X and Y is assumed to
be
y
p(x, y )
12 15 20
12 .05 .05 .10
x
15 .05 .10 .35
20 0 .20 .10
a. What is the probability for them to both have the $12 dinner?
b. What is the probability that they have the same price dinner?
Jointly Distributed Random Variables
Example (Problem 75)
A restaurant serves three fixed-price dinners costing $12, $15, and
$20. For a randomly selected couple dinning at this restaurant, let
X = the cost of the man’s dinner and Y = the cost of
the woman’s dinner. If the joint pmf of X and Y is assumed to
be
y
p(x, y )
12 15 20
12 .05 .05 .10
x
15 .05 .10 .35
20 0 .20 .10
a. What is the probability for them to both have the $12 dinner?
b. What is the probability that they have the same price dinner?
c. What is the probability that the man’s dinner cost $12?
Jointly Distributed Random Variables
Jointly Distributed Random Variables
Definition
Let X and Y be two discrete random variables defined on the
sample space S of an experiment with joint probability mass
function p(x, y ). Then the pmf’s of each one of the variables alone
are called the marginal probability mass functions, denoted by
pX (x) and pY (y ), respectively. Furthermore,
X
X
pX (x) =
p(x, y ) and pY (y ) =
p(x, y )
y
x
Jointly Distributed Random Variables
Jointly Distributed Random Variables
Example (Problem 75) continued:
The marginal probability mass functions for the previous example is
calculated as following:
y
p(x, y ) 12 15 20 p(x, ·)
12
.05 .05 .10
.20
x
15
.05 .10 .35
.50
20
0 .20 .10
.30
p(·, y ) .10 .35 .55
Jointly Distributed Random Variables
Jointly Distributed Random Variables
Definition
Let X and Y be continuous random variables. A joint probability
density function f (x, y ) for
R ∞these
R ∞ two variables is a function
satisfying f (x, y ) ≥ 0 and −∞ −∞ f (x, y )dxdy = 1.
For any two-dimensional set A
ZZ
P[(X , Y ) ∈ A] =
f (x, y )dxdy
A
In particular, if A is the two-dimensilnal rectangle
{(x, y ) : a ≤ x ≤ b, c ≤ y ≤ d}, then
Z bZ
P[(X , Y ) ∈ A] = P(a ≤ X ≤ b, c ≤ Y ≤ d) =
d
f (x, y )dydx
a
c
Jointly Distributed Random Variables
Jointly Distributed Random Variables
Definition
Let X and Y be continuous random variables with joint pdf
f (x, y ). Then the marginal probability density functions of X
and Y , denoted by fX (x) and fY (y ), respectively, are given by
Z ∞
fX (x) =
f (x, y )dy
for − ∞ < x < ∞
Z−∞
∞
fY (y ) =
f (x, y )dx
for − ∞ < y < ∞
−∞
Jointly Distributed Random Variables
Jointly Distributed Random Variables
Example (variant of Problem 12)
Two components of a minicomputer have the following joint pdf
for their useful lifetimes X and Y :
(
xe −(x+y ) x ≥ 0 and y ≥ 0
f (x, y ) =
0
otherwise
Jointly Distributed Random Variables
Example (variant of Problem 12)
Two components of a minicomputer have the following joint pdf
for their useful lifetimes X and Y :
(
xe −(x+y ) x ≥ 0 and y ≥ 0
f (x, y ) =
0
otherwise
a. What is the probability that the lifetimes of both components
excceed 3?
Jointly Distributed Random Variables
Example (variant of Problem 12)
Two components of a minicomputer have the following joint pdf
for their useful lifetimes X and Y :
(
xe −(x+y ) x ≥ 0 and y ≥ 0
f (x, y ) =
0
otherwise
a. What is the probability that the lifetimes of both components
excceed 3?
b. What are the marginal pdf’s of X and Y ?
Jointly Distributed Random Variables
Example (variant of Problem 12)
Two components of a minicomputer have the following joint pdf
for their useful lifetimes X and Y :
(
xe −(x+y ) x ≥ 0 and y ≥ 0
f (x, y ) =
0
otherwise
a. What is the probability that the lifetimes of both components
excceed 3?
b. What are the marginal pdf’s of X and Y ?
c. What is the probability that the lifetime X of the first
component excceeds 3?
Jointly Distributed Random Variables
Example (variant of Problem 12)
Two components of a minicomputer have the following joint pdf
for their useful lifetimes X and Y :
(
xe −(x+y ) x ≥ 0 and y ≥ 0
f (x, y ) =
0
otherwise
a. What is the probability that the lifetimes of both components
excceed 3?
b. What are the marginal pdf’s of X and Y ?
c. What is the probability that the lifetime X of the first
component excceeds 3?
d. What is the probability that the lifetime of at least one
component excceeds 3?
Download