Normal Distribution Normal Distribution Definition A continuous rv X is said to have a normal distribution with parameter µ and σ (µ and σ 2 ), where −∞ < µ < ∞ and σ > 0, if the pdf of X is f (x; µ, σ) = √ 1 2 2 e −(x−µ) /(2σ ) 2πσ We use the notation X ∼ N(µ, σ 2 ) to denote that X is rormally distributed with parameters µ and σ 2 . Normal Distribution Proposition For X ∼ N(µ, σ 2 ), we have E (X ) = µ and V (X ) = σ 2 Normal Distribution Definition The normal distribution with parameter values µ = 0 and σ = 1 is called the standard normal distribution. A random variable having a standard normal distribution is called a standard normal random variable and will be denoted by Z . The pdf of Z is 1 2 f (z; 0, 1) = √ e −z /2 2π −∞<z <∞ The graph of f (z; 0, 1) is called R z the standard normal (or z) curve. The cdf of Z is P(Z ≤ z) = −∞ f (y ; 0, 1)dy , which we will denote by Φ(z). Normal Distribution Proposition If X has a normal distribution with mean µ and stadard deviation σ, then X −µ Z= σ has a standard normal distribution. Thus a−µ b−µ ≤Z ≤ ) σ σ b−µ a−µ = Φ( ) − Φ( ) σ σ P(a ≤ X ≤ b) = P( P(X ≤ a) = Φ( a−µ ) σ P(X ≥ b) = 1 − Φ( b−µ ) σ Normal Distribution Normal Distribution Proposition {(100p)th percentile for N(µ, σ 2 )} = µ + {(100p)th percentile for N(0, 1)} · σ Normal Distribution Normal Distribution Proposition Let X be a binomial rv based on n trials with success probability p. Then if the binomial probability histogram is not too skewed, X has √ approximately a normal distribution with µ = np and σ = npq, where q = 1 − p. In particular, for x = a posible value of X , area under the normal curve P(X ≤ x) = B(x; n, p) ≈ to the left of x+0.5 x+0.5 − np = Φ( √ ) npq In practice, the approximation is adequate provided that both np ≥ 10 and nq ≥ 10, since there is then enough symmetry in the underlying binomial distribution. Normal Distribution Normal Distribution A graphical explanation for P(X ≤ x) = B(x; n, p) ≈ = Φ( area under the normal curve to the left of x+0.5 x+0.5 − np ) √ npq Normal Distribution A graphical explanation for P(X ≤ x) = B(x; n, p) ≈ = Φ( area under the normal curve to the left of x+0.5 x+0.5 − np ) √ npq Normal Distribution Normal Distribution Example (Problem 54) Suppose that 10% of all steel shafts produced by a certain process are nonconforming but can be reworked (rather than having to be scrapped). Consider a random sample of 200 shafts, and let X denote the number among these that are nonconforming and can be reworked. What is the (approximate) probability that X is between 15 and 25 (inclusive)? Normal Distribution Example (Problem 54) Suppose that 10% of all steel shafts produced by a certain process are nonconforming but can be reworked (rather than having to be scrapped). Consider a random sample of 200 shafts, and let X denote the number among these that are nonconforming and can be reworked. What is the (approximate) probability that X is between 15 and 25 (inclusive)? In this problem n = 200, p = 0.1 and q = 1 − p = 0.9. Thus np = 20 > 10 and nq = 180 > 10 Normal Distribution Example (Problem 54) Suppose that 10% of all steel shafts produced by a certain process are nonconforming but can be reworked (rather than having to be scrapped). Consider a random sample of 200 shafts, and let X denote the number among these that are nonconforming and can be reworked. What is the (approximate) probability that X is between 15 and 25 (inclusive)? In this problem n = 200, p = 0.1 and q = 1 − p = 0.9. Thus np = 20 > 10 and nq = 180 > 10 P(15 ≤ X ≤ 25) = Bin(25; 200, 0.1) − Bin(14; 200, 0.1) 25 + 0.5 − 20 15 + 0.5 − 20 ≈ Φ( √ ) − Φ( √ ) 200 · 0.1 · 0.9 200 · 0.1 · 0.9 = Φ(0.3056) − Φ(−0.2500) = 0.6217 − 0.4013 = 0.2204 Exponential Distribution Exponential Distribution Definition X is said to have an exponential distribution with parameter λ(λ > 0) if the pdf of X is ( λe −λx x ≥ 0 f (x; λ) = 0 otherwise Exponential Distribution Exponential Distribution Proposition If X ∼ EXP(λ), then E (X ) = 1 λ and V (X ) = 1 λ2 And the cdf for X is ( 1 − e −λx F (x; λ) = 0 x ≥0 x <0 Exponential Distribution Exponential Distribution “Memoryless” Property P(X ≥ t) = P(X ≥ t + t0 | X ≥ t0 ) for any positive t and t0 . Exponential Distribution “Memoryless” Property P(X ≥ t) = P(X ≥ t + t0 | X ≥ t0 ) for any positive t and t0 . In words, the distribution of additional lifetime is exactly the same as the original distribution of lifetime, so at each point in time the component shows no effect of wear. In other words, the distribution of remaining lifetime is independent of current age. Probability Plot Probability Plot Example: There is a machine available for cutting corks intended for use in wine bottles. We want to find out the distribution of the diameters of the corks produced by that machine. Assume we have 10 samples produced by that machine and the diameters is recorded as following: 3.0879 2.6030 3.2546 3.5931 2.8970 3.1253 2.7377 2.4756 2.7740 2.5133 Probability Plot Probability Plot Probability Plot Probability Plot 3.0879 2.6030 3.2546 3.5931 2.8970 3.1253 2.7377 2.4756 2.7740 2.5133 Probability Plot 3.0879 2.6030 3.2546 3.5931 2.8970 3.1253 2.7377 2.4756 2.7740 2.5133 Probability Plot Probability Plot Sample Percentile Probability Plot Sample Percentile Recall: The (100p)th percentile of the distribution of a continuous rv X , denoted by η(p), is defined by R η(p) p = F (η(p)) = −∞ f (y )dy . Probability Plot Sample Percentile Recall: The (100p)th percentile of the distribution of a continuous rv X , denoted by η(p), is defined by R η(p) p = F (η(p)) = −∞ f (y )dy . In words, the (100p)th percentile η(p) is the X value such that there are 100p% X values below η(p). Probability Plot Sample Percentile Recall: The (100p)th percentile of the distribution of a continuous rv X , denoted by η(p), is defined by R η(p) p = F (η(p)) = −∞ f (y )dy . In words, the (100p)th percentile η(p) is the X value such that there are 100p% X values below η(p). Similarly, we can define sample percentile in the same manner, i.e. the (100p)th percentile xp is the value such that there are 100p% sample values below xp . Probability Plot Sample Percentile Recall: The (100p)th percentile of the distribution of a continuous rv X , denoted by η(p), is defined by R η(p) p = F (η(p)) = −∞ f (y )dy . In words, the (100p)th percentile η(p) is the X value such that there are 100p% X values below η(p). Similarly, we can define sample percentile in the same manner, i.e. the (100p)th percentile xp is the value such that there are 100p% sample values below xp . Unfortunately, xp may not be a sample value for some p. Probability Plot Sample Percentile Recall: The (100p)th percentile of the distribution of a continuous rv X , denoted by η(p), is defined by R η(p) p = F (η(p)) = −∞ f (y )dy . In words, the (100p)th percentile η(p) is the X value such that there are 100p% X values below η(p). Similarly, we can define sample percentile in the same manner, i.e. the (100p)th percentile xp is the value such that there are 100p% sample values below xp . Unfortunately, xp may not be a sample value for some p. e.g. for the previous example, what is the 35th percentile for the ten sample values? Probability Plot Probability Plot Definition Assume we have a sample with size n. Order the n sample observations from smallest to largest. Then the ith smallest observation in the list is taken to be the [100(i − 0.5)/n]th sample percentile. Probability Plot Definition Assume we have a sample with size n. Order the n sample observations from smallest to largest. Then the ith smallest observation in the list is taken to be the [100(i − 0.5)/n]th sample percentile. Remark: 1. Why “i − 0.5”? Probability Plot Definition Assume we have a sample with size n. Order the n sample observations from smallest to largest. Then the ith smallest observation in the list is taken to be the [100(i − 0.5)/n]th sample percentile. Remark: 1. Why “i − 0.5”? We regard the sample observation as being half in the lower group and half in the upper group. Probability Plot Definition Assume we have a sample with size n. Order the n sample observations from smallest to largest. Then the ith smallest observation in the list is taken to be the [100(i − 0.5)/n]th sample percentile. Remark: 1. Why “i − 0.5”? We regard the sample observation as being half in the lower group and half in the upper group. e.g. if n = 9, then the sample median is the 5th largest observation and this observation is regarded as two parts: one in the lower half and one in the upper half. Probability Plot Definition Assume we have a sample with size n. Order the n sample observations from smallest to largest. Then the ith smallest observation in the list is taken to be the [100(i − 0.5)/n]th sample percentile. Remark: 1. Why “i − 0.5”? We regard the sample observation as being half in the lower group and half in the upper group. e.g. if n = 9, then the sample median is the 5th largest observation and this observation is regarded as two parts: one in the lower half and one in the upper half. 2. Once the percentage values 100(i − 0.5)/n(i = 1, 2, . . . , n) have been calculated, sample percentiles corresponding to intermediate percentages can be obtained by linear interpolation. Probability Plot Probability Plot Example: for the previous example, the [100(i − 0.5)/n]th sample percentile is tabulated as following: 2.4756 2.5133 2.6030 100(1-.5)/10 = 5% 100(2-.5)/10 = 15% 100(3-.5)/10 = 25% 2.7377 2.7740 100(4-.5)/10 = 35% 100(5-.5)/10 = 45% 2.8970 3.0879 3.1253 100(6-.5)/10 = 55% 100(7-.5)/10 = 65% 100(8-.5)/10 = 75% 3.2546 3.5931 100(9-.5)/10 = 85% 100(10-.5)/10 = 95% Probability Plot Example: for the previous example, the [100(i − 0.5)/n]th sample percentile is tabulated as following: 2.4756 2.5133 2.6030 100(1-.5)/10 = 5% 100(2-.5)/10 = 15% 100(3-.5)/10 = 25% 2.7377 2.7740 100(4-.5)/10 = 35% 100(5-.5)/10 = 45% 2.8970 3.0879 3.1253 100(6-.5)/10 = 55% 100(7-.5)/10 = 65% 100(8-.5)/10 = 75% 3.2546 3.5931 100(9-.5)/10 = 85% 100(10-.5)/10 = 95% The 10th percentile would be (2.4756 + 2.5133)/2 = 2.49445 Probability Plot Probability Plot Idea for Quantile-Quantile Plot: 1. Determine the “[100(i − 0.5)/n]th sample percentile” for a given sample. Probability Plot Idea for Quantile-Quantile Plot: 1. Determine the “[100(i − 0.5)/n]th sample percentile” for a given sample. 2. Find the corresponding [100(i − 0.5)/n]th percentile from the population with the assumed distribution; for example, if the assumed distribution is standard normal, then find corresponding [100(i − 0.5)/n]th percentile from the standard normal distribution. Probability Plot Idea for Quantile-Quantile Plot: 1. Determine the “[100(i − 0.5)/n]th sample percentile” for a given sample. 2. Find the corresponding [100(i − 0.5)/n]th percentile from the population with the assumed distribution; for example, if the assumed distribution is standard normal, then find corresponding [100(i − 0.5)/n]th percentile from the standard normal distribution. 3. Consider the (population percentile, sample percentile) pairs, i.e. [100(i − 0.5)/n]th percentile, ith smallest sample of the distribution observation Probability Plot Idea for Quantile-Quantile Plot: 1. Determine the “[100(i − 0.5)/n]th sample percentile” for a given sample. 2. Find the corresponding [100(i − 0.5)/n]th percentile from the population with the assumed distribution; for example, if the assumed distribution is standard normal, then find corresponding [100(i − 0.5)/n]th percentile from the standard normal distribution. 3. Consider the (population percentile, sample percentile) pairs, i.e. [100(i − 0.5)/n]th percentile, ith smallest sample of the distribution observation 4. Each pair plotted as a point on a two-dimensional coordinate system should fall close to a 45◦ line. Probability Plot Idea for Quantile-Quantile Plot: 1. Determine the “[100(i − 0.5)/n]th sample percentile” for a given sample. 2. Find the corresponding [100(i − 0.5)/n]th percentile from the population with the assumed distribution; for example, if the assumed distribution is standard normal, then find corresponding [100(i − 0.5)/n]th percentile from the standard normal distribution. 3. Consider the (population percentile, sample percentile) pairs, i.e. [100(i − 0.5)/n]th percentile, ith smallest sample of the distribution observation 4. Each pair plotted as a point on a two-dimensional coordinate system should fall close to a 45◦ line. Substantial deviations of the plotted points from a 45◦ line cast doubt on the assumption that the distribution under consideration is the correct one. Probability Plot Probability Plot Example 4.29: The value of a certain physical constant is known to an experimenter. The experimenter makes n = 10 independent measurements of this value using a particular measurement device and records the resulting measurement errors (error = observed value - true value). These observations appear in the following table. Percentage Sample Observation Percentage Sample Observation 5 -1.91 55 0.35 15 -1.25 65 0.72 25 -0.75 75 0.87 35 -0.53 85 1.40 45 0.20 95 1.56 Probability Plot Example 4.29: The value of a certain physical constant is known to an experimenter. The experimenter makes n = 10 independent measurements of this value using a particular measurement device and records the resulting measurement errors (error = observed value - true value). These observations appear in the following table. Percentage Sample Observation Percentage Sample Observation 5 -1.91 55 0.35 15 -1.25 65 0.72 25 -0.75 75 0.87 35 -0.53 85 1.40 45 0.20 95 1.56 Is it plausible that the random variable measurement error has standard normal distribution? Probability Plot Probability Plot We first find the corresponding population distribution percentiles, in this case, the z percentiles: Percentage Sample Observation z percentile Percentage Sample Observation z percentile 5 -1.91 -1.645 55 0.35 0.126 15 -1.25 -1.037 65 0.72 0.385 25 -0.75 -0.675 75 0.87 0.675 35 -0.53 -0.385 85 1.40 1.037 45 0.20 -0.126 95 1.56 1.645 Probability Plot Probability Plot Probability Plot Probability Plot What about the first example? We are only interested in whether the ten sample observations come from a normal distribution. Probability Plot What about the first example? We are only interested in whether the ten sample observations come from a normal distribution. Recall: {(100p)th percentile for N(µ, σ 2 )} = µ + {(100p)th percentile for N(0, 1)} · σ Probability Plot What about the first example? We are only interested in whether the ten sample observations come from a normal distribution. Recall: {(100p)th percentile for N(µ, σ 2 )} = µ + {(100p)th percentile for N(0, 1)} · σ If µ = 0, then the pairs (σ · [z percentile], observation) fall on a 45◦ line, which has slope 1. Probability Plot What about the first example? We are only interested in whether the ten sample observations come from a normal distribution. Recall: {(100p)th percentile for N(µ, σ 2 )} = µ + {(100p)th percentile for N(0, 1)} · σ If µ = 0, then the pairs (σ · [z percentile], observation) fall on a 45◦ line, which has slope 1. Therefore the pairs ([z percentile], observation) fall on a line passing through (0,0) (i.e., one with y -intercept 0) but having slope σ rather than 1. Probability Plot What about the first example? We are only interested in whether the ten sample observations come from a normal distribution. Recall: {(100p)th percentile for N(µ, σ 2 )} = µ + {(100p)th percentile for N(0, 1)} · σ If µ = 0, then the pairs (σ · [z percentile], observation) fall on a 45◦ line, which has slope 1. Therefore the pairs ([z percentile], observation) fall on a line passing through (0,0) (i.e., one with y -intercept 0) but having slope σ rather than 1. Now for µ 6= 0, the y -intercept is µ instead of 0. Probability Plot Probability Plot Normal Probability Plot A plot of the n pairs ([100(i − 0.5)/n]th z percentile, ith smallest observation) on a two-dimensional coordinate system is called a normal probability plot. If the sample observations are in fact drawn from a normal distribution with mean value µ and standard deviation σ, the points should fall close to a straight line with slope σ and y -intercept µ. Thus a plot for which the points fall close to some straight line suggests that the assumption of a normal population distribution is plausible. Probability Plot Probability Plot Percentage Sample Observation z percentile Percentage Sample Observation z percentile 5 2.4756 -1.645 55 2.8970 0.126 15 2.5133 -1.037 65 3.0879 0.385 25 2.6030 -0.675 75 3.1253 0.675 35 2.7377 -0.385 85 3.2546 1.037 45 2.7740 -0.126 95 3.5931 1.645 Probability Plot Probability Plot A nonnormal population distribution can often be placed in one of the following three categories: 1. It is symmetric and has “lighter tails” than does a normal distribution; that is, the density curve declines more rapidly out in the tails than does a normal curve. 2. It is symmetric and heavy-tailed compared to a normal distribution. 3. It is skewed. Probability Plot Probability Plot Symmetric and “light-tailed”: e.g. Uniform distribution Probability Plot Probability Plot Symmetric and heavy-tailed: e.g. Cauchy distribution with pdf f (x) = 1/[π(1 + x 2 )] for −∞ < x < ∞ Probability Plot Probability Plot Skewed: e.g. lognormal distribution Probability Plot Probability Plot Some guidances for probability plot for normal distributions (from the book Fitting Equations to Data (2nd ed.) Daniel, Cuthbert, and Fed Wood, Wiley, New York, 1980) Probability Plot Some guidances for probability plot for normal distributions (from the book Fitting Equations to Data (2nd ed.) Daniel, Cuthbert, and Fed Wood, Wiley, New York, 1980) 1. For sample size smaller than 30, there is typically greater variation in the apperance of the probability plot. Probability Plot Some guidances for probability plot for normal distributions (from the book Fitting Equations to Data (2nd ed.) Daniel, Cuthbert, and Fed Wood, Wiley, New York, 1980) 1. For sample size smaller than 30, there is typically greater variation in the apperance of the probability plot. 2. Only for much larger sample sizes does a linear pattern generally predominate. Probability Plot Some guidances for probability plot for normal distributions (from the book Fitting Equations to Data (2nd ed.) Daniel, Cuthbert, and Fed Wood, Wiley, New York, 1980) 1. For sample size smaller than 30, there is typically greater variation in the apperance of the probability plot. 2. Only for much larger sample sizes does a linear pattern generally predominate. Therefore, when a plot is based on a small sample size, only a very substantial departure from linearity should be taken as conclusive evidence of nonnorality. Probability Plot Probability Plot Definition Consider a family of probability distributions involving two parameters, θ1 and θ2 , and let F (x; θ1 , θ2 ) denote the corresponding cdf’s. The parameters θ1 and θ2 are said to be location and scale parameters, respectively, if F (x; θ1 , θ2 ) is a function of (x − θ1 )/θ2 . Probability Plot Definition Consider a family of probability distributions involving two parameters, θ1 and θ2 , and let F (x; θ1 , θ2 ) denote the corresponding cdf’s. The parameters θ1 and θ2 are said to be location and scale parameters, respectively, if F (x; θ1 , θ2 ) is a function of (x − θ1 )/θ2 . e.g. 1. Normal distributions N(µ, σ): F (x; µ, σ) = Φ( x−µ σ ). Probability Plot Definition Consider a family of probability distributions involving two parameters, θ1 and θ2 , and let F (x; θ1 , θ2 ) denote the corresponding cdf’s. The parameters θ1 and θ2 are said to be location and scale parameters, respectively, if F (x; θ1 , θ2 ) is a function of (x − θ1 )/θ2 . e.g. 1. Normal distributions N(µ, σ): F (x; µ, σ) = Φ( x−µ σ ). 2. The extreme value distribution with cdf F (x; θ1 , θ2 ) = 1 − e −e (x−θ1 )/θ2 Probability Plot Probability Plot For Weibull distribution: α F (x; α, β) = 1 − e −(x/β) , the parameter β is a scale parameter but α is NOT a location parameter. α is usually referred to as a shape parameter. Probability Plot For Weibull distribution: α F (x; α, β) = 1 − e −(x/β) , the parameter β is a scale parameter but α is NOT a location parameter. α is usually referred to as a shape parameter. Fortunately, if X has a Weibull distribution with shape parameter α and scale parameter β, then the transformed variable ln(X ) has an extreme value distribution with location parameter θ1 = ln(β) and scale parameter θ2 = 1/α. Probability Plot Probability Plot The gamma distribution also has a shape parameter α. However, there is no transformation h(•) such that h(X ) has a distribution that depends only on location and scale parameters. Probability Plot The gamma distribution also has a shape parameter α. However, there is no transformation h(•) such that h(X ) has a distribution that depends only on location and scale parameters. Thus, before we construct a probability plot, we have to estimate the shape parameter from the sample data. Jointly Distributed Random Variables Jointly Distributed Random Variables Consider tossing a fair (1,1) (1,2) (1,3) (2,1) (2,2) (2,3) ··· ··· ··· (6,1) (6,2) (6,3) die twice. Then the outcomes would be (1,4) (1,5) (1,6) (2,4) (2,5) (2,6) ··· ··· ··· (6,4) (6,5) (6,6) and the probability for each outcome is 1 36 . Jointly Distributed Random Variables Consider tossing a fair (1,1) (1,2) (1,3) (2,1) (2,2) (2,3) ··· ··· ··· (6,1) (6,2) (6,3) die twice. Then the outcomes would be (1,4) (1,5) (1,6) (2,4) (2,5) (2,6) ··· ··· ··· (6,4) (6,5) (6,6) 1 . and the probability for each outcome is 36 If we define two random variables by X = the outcome of the first toss and Y = the outcome of the second toss, Jointly Distributed Random Variables Consider tossing a fair (1,1) (1,2) (1,3) (2,1) (2,2) (2,3) ··· ··· ··· (6,1) (6,2) (6,3) die twice. Then the outcomes would be (1,4) (1,5) (1,6) (2,4) (2,5) (2,6) ··· ··· ··· (6,4) (6,5) (6,6) 1 . and the probability for each outcome is 36 If we define two random variables by X = the outcome of the first toss and Y = the outcome of the second toss, then the outcome for this experiment (two tosses) can be describe by the random pair (X , Y ), Jointly Distributed Random Variables Consider tossing a fair (1,1) (1,2) (1,3) (2,1) (2,2) (2,3) ··· ··· ··· (6,1) (6,2) (6,3) die twice. Then the outcomes would be (1,4) (1,5) (1,6) (2,4) (2,5) (2,6) ··· ··· ··· (6,4) (6,5) (6,6) 1 . and the probability for each outcome is 36 If we define two random variables by X = the outcome of the first toss and Y = the outcome of the second toss, then the outcome for this experiment (two tosses) can be describe by the random pair (X , Y ), and the probability for any possible value 1 of that random pair (x, y ) is 36 . Jointly Distributed Random Variables Jointly Distributed Random Variables Definition Let X and Y be two discrete random variables defined on the sample space S of an experiment. The joint probability mass function p(x, y ) is defined for each pair of numbers (x, y ) by p(x, y ) = P(X = x and Y = y ) P P (It must be the case that p(x, y ) ≥ 0 and x y p(x, y ) = 1.) For any event A consisting of pairs of (x, y ), the probability P[(X , Y ) ∈ A] is obtained by summing the joint pmf over pairs in A: XX P[(X , Y ) ∈ A] = p(x, y ) (x,y )∈A Jointly Distributed Random Variables Jointly Distributed Random Variables Example (Problem 75) A restaurant serves three fixed-price dinners costing $12, $15, and $20. For a randomly selected couple dinning at this restaurant, let X = the cost of the man’s dinner and Y = the cost of the woman’s dinner. If the joint pmf of X and Y is assumed to be y p(x, y ) 12 15 20 12 .05 .05 .10 x 15 .05 .10 .35 20 0 .20 .10 Jointly Distributed Random Variables Example (Problem 75) A restaurant serves three fixed-price dinners costing $12, $15, and $20. For a randomly selected couple dinning at this restaurant, let X = the cost of the man’s dinner and Y = the cost of the woman’s dinner. If the joint pmf of X and Y is assumed to be y p(x, y ) 12 15 20 12 .05 .05 .10 x 15 .05 .10 .35 20 0 .20 .10 a. What is the probability for them to both have the $12 dinner? Jointly Distributed Random Variables Example (Problem 75) A restaurant serves three fixed-price dinners costing $12, $15, and $20. For a randomly selected couple dinning at this restaurant, let X = the cost of the man’s dinner and Y = the cost of the woman’s dinner. If the joint pmf of X and Y is assumed to be y p(x, y ) 12 15 20 12 .05 .05 .10 x 15 .05 .10 .35 20 0 .20 .10 a. What is the probability for them to both have the $12 dinner? b. What is the probability that they have the same price dinner? Jointly Distributed Random Variables Example (Problem 75) A restaurant serves three fixed-price dinners costing $12, $15, and $20. For a randomly selected couple dinning at this restaurant, let X = the cost of the man’s dinner and Y = the cost of the woman’s dinner. If the joint pmf of X and Y is assumed to be y p(x, y ) 12 15 20 12 .05 .05 .10 x 15 .05 .10 .35 20 0 .20 .10 a. What is the probability for them to both have the $12 dinner? b. What is the probability that they have the same price dinner? c. What is the probability that the man’s dinner cost $12? Jointly Distributed Random Variables Jointly Distributed Random Variables Definition Let X and Y be two discrete random variables defined on the sample space S of an experiment with joint probability mass function p(x, y ). Then the pmf’s of each one of the variables alone are called the marginal probability mass functions, denoted by pX (x) and pY (y ), respectively. Furthermore, X X pX (x) = p(x, y ) and pY (y ) = p(x, y ) y x Jointly Distributed Random Variables Jointly Distributed Random Variables Example (Problem 75) continued: The marginal probability mass functions for the previous example is calculated as following: y p(x, y ) 12 15 20 p(x, ·) 12 .05 .05 .10 .20 x 15 .05 .10 .35 .50 20 0 .20 .10 .30 p(·, y ) .10 .35 .55 Jointly Distributed Random Variables Jointly Distributed Random Variables Definition Let X and Y be continuous random variables. A joint probability density function f (x, y ) for R ∞these R ∞ two variables is a function satisfying f (x, y ) ≥ 0 and −∞ −∞ f (x, y )dxdy = 1. For any two-dimensional set A ZZ P[(X , Y ) ∈ A] = f (x, y )dxdy A In particular, if A is the two-dimensilnal rectangle {(x, y ) : a ≤ x ≤ b, c ≤ y ≤ d}, then Z bZ P[(X , Y ) ∈ A] = P(a ≤ X ≤ b, c ≤ Y ≤ d) = d f (x, y )dydx a c Jointly Distributed Random Variables Jointly Distributed Random Variables Definition Let X and Y be continuous random variables with joint pdf f (x, y ). Then the marginal probability density functions of X and Y , denoted by fX (x) and fY (y ), respectively, are given by Z ∞ fX (x) = f (x, y )dy for − ∞ < x < ∞ Z−∞ ∞ fY (y ) = f (x, y )dx for − ∞ < y < ∞ −∞ Jointly Distributed Random Variables Jointly Distributed Random Variables Example (variant of Problem 12) Two components of a minicomputer have the following joint pdf for their useful lifetimes X and Y : ( xe −(x+y ) x ≥ 0 and y ≥ 0 f (x, y ) = 0 otherwise Jointly Distributed Random Variables Example (variant of Problem 12) Two components of a minicomputer have the following joint pdf for their useful lifetimes X and Y : ( xe −(x+y ) x ≥ 0 and y ≥ 0 f (x, y ) = 0 otherwise a. What is the probability that the lifetimes of both components excceed 3? Jointly Distributed Random Variables Example (variant of Problem 12) Two components of a minicomputer have the following joint pdf for their useful lifetimes X and Y : ( xe −(x+y ) x ≥ 0 and y ≥ 0 f (x, y ) = 0 otherwise a. What is the probability that the lifetimes of both components excceed 3? b. What are the marginal pdf’s of X and Y ? Jointly Distributed Random Variables Example (variant of Problem 12) Two components of a minicomputer have the following joint pdf for their useful lifetimes X and Y : ( xe −(x+y ) x ≥ 0 and y ≥ 0 f (x, y ) = 0 otherwise a. What is the probability that the lifetimes of both components excceed 3? b. What are the marginal pdf’s of X and Y ? c. What is the probability that the lifetime X of the first component excceeds 3? Jointly Distributed Random Variables Example (variant of Problem 12) Two components of a minicomputer have the following joint pdf for their useful lifetimes X and Y : ( xe −(x+y ) x ≥ 0 and y ≥ 0 f (x, y ) = 0 otherwise a. What is the probability that the lifetimes of both components excceed 3? b. What are the marginal pdf’s of X and Y ? c. What is the probability that the lifetime X of the first component excceeds 3? d. What is the probability that the lifetime of at least one component excceeds 3?