Math 149A HW 5 Solutions 1. Suppose that X is a random variable such that E(X) = 13 and V ar(X) = 25. Using Chebyshev’s inequality, give a bound on P (5 < X < 21). Solution: This is similar to the last example from class on 10/30. The idea is to write P (5 < X < 21) = P (|X − 13| < 8) = 1 − P (|X − 13| ≥ 8). Chebyshev’s inequality states that 1 k2 In this case σ = 5, and we want kσ = 8. So we’ll take k = 1.6, giving a bound P (|X − µ| ≥ kσ) ≤ P (|X − 13| ≥ 8) ≤ 1 ≈ 0.391 1.62 This means that P (5 < X < 21) ≥ 1 − .391 ≈ 0.609 Remark: Notice that the inequality sign flipped when I did the subtraction! 2. Chebyshev’s inequality (with k = 2) says that for any variable X with mean µ and variance σ 2 , we have P(|X − µ| ≥ 2σ) ≤ 1/4. However, ”at most” is not the same as equal, and the actual probability may be very different. For each of the following four distributions, determine the actual probability that |X − µ| ≥ 2σ. Part a: X is continuous, with pdf given by f (x) = 1 2 for 0 ≤ x < 2, and 0 otherwise. Solution: We have Z 2 µ = 0 E(X 2 ) = Z 0 2 x x2 dx = |20 = 1 2 4 2 x x3 4 dx = |20 = 2 6 3 1 3 1 2 So σ = √3 . The event we care about is |X − 1| ≥ √3 , which corresponds to either V ar(X) = E(X 2 ) − [E(X)]2 = 2 2 X ≥ 1 + √ ≈ 1.15 or X ≤ 1 − √ ≈ −0.15 3 3 but X is never in either of these ranges, because X is always between 0 and 2. So the probability is 0 . 1 Remark: One way you could have run into trouble here is to try writing 2 2 2 P (|X − 1| ≥ √ = 1 − P (1 − √ < X < 1 + √ ) 3 3 3 Z 1+ √2 3 1 dx = 1− 2 1− √2 3 The problem with this is that the pdf of X is not 12 everywhere – only between 0 and 2. So you can’t just plug in 21 for the pdf without checking where x is (this also shows up in part b). If you work out numerically the value you get by doing this integral, you’ll end up with a negative number for your answer. This should be a sign that something has gone wrong! Part b: X is continuous, with pdf given by f (x) = 2e2x for 0 ≤ x and 0 otherwise (note: You computed µ and σ for this distribution on the last homework already!) Solution: On the last homework, you found µ = 12 , σ 2 = 41 . So we want P (|X − 21 | ≥ 1). Since X is always non-negative, this corresponds to the event X ≥ 32 . This is 3 3 P (X ≥ ) = 1 − P (X < ) 2 2 Z 3/2 = 1− 2e−2x dx 0 3/2 = 1 − −e−2x |0 = 1 − 1 − e−3 = e−3 ≈ 0.050 Part c: X is discrete, with pmf given by p(1) = p(−1) = 12 , and 0 everywhere else. Solution: We have 1 1 − = 0, 2 2 1 2 1 V ar(X) = (1) + (−1)2 = 1 2 2 But this means X is always within 2σ of the mean, so the probability is 0 . E(X) = Part d: X is discrete, with pmf given by p(4) = p(−4) = everywhere else 1 , p(0) 8 = 3 , 4 and p(x) = 0 Solution: We have 1 1 E(X) = 4( ) + (−4)( ) = 0 8 8 1 1 V ar(X) = (4 − 0)2 ( ) + (−4 − 0)2 ( ) = 4 8 8 σ = 2 So the event ”|X − µ| ≥ 2σ” corresponds to the event ”|X| ≥ 4”. Of the possible values of 1 X, this event holds when x = 4 or x = −4. So the probability we want is p(4) + p(−4) = . 4 2 Note that for all four of these distributions, Chebyshev’s inequality says the same thing: P (|X − µ| ≥ 2σ) ≤ 14 . But the actual probabilities varied quite a bit from part to part. In general, Chebyshev’s inequality is only a bound on the probability X is far away from the mean. For some distributions, the actual probability is much smaller than you’d guess just by looking at Chebyshev’s inequality. But, as part d shows, there are situations where Chebyshev’s inequality is actually tight (an equality). So if you want to beat Chebyshev’s inequality, you need to look beyond the mean and variance, which is where problem 3 comes into play. 3a: Let X be any random variable such that E(X − µ)4 exists. Show that for any c > 0, we have E(X − µ)4 P (|X − µ| > c) ≤ c4 Solution: We follow the same idea as when we proved Chebyshev’s inequality: First write everything in terms of the thing we know the expectation of, then apply Markov’s inequality. So we write P (|X − µ| > c) = P ((X − µ)4 > c4 ) ≤ P ((X − µ)4 ≥ c4 ) Since (X − µ)4 is non-negative, we can use Markov’s inequality, which tells us that P ((X − µ)4 ≥ c4 ) ≤ E(X − µ)4 c4 this is exactly what we want. 3b: Let X be any random variable whose mgf M (t) is defined for all positive t. Show that for any a > 0 and any t, we have P (X > a) ≤ e−at M (t) Solution: The right hand side here is M (t) E(etX ) = eat eat In this form it’s starting to look like part a. So our goal will be to use Markov’s inequality, applied to Y = etX . Again, we start by writing things in terms of Y : P (X > a) = P (etX > eat ) ≤ P (Y ≥ eat ) Here it’s important that t is positive, so that etX is an increasing function of X and doesn’t flip the inequality. By Markov’s inequality, we have E(Y ) M (t) P (Y ≥ eat ) ≤ at = at , e e and again we’re done. Remark: Chebyshev’s inequality says if the variance is small, a variable is usually close to the mean. These inequalities say something similar, but rely on you knowing the fourth moment or the mgf. In some cases Chebyshev’s inequality can be very far off (give you a bound that’s far away from the actual probability), but these bounds can give you a much better estimate. 3 4. Let X and Y be two random variables whose joint pdf is given by 6 2 (x + y) if 0 ≤ x, y < 1 5 f (x, y) = 0 otherwise Part a: Compute P 0 ≤ X ≤ 1 2 and 2 3 ≤Y ≤1 Solution: This is Z 1/2 Z 1 6 2 (x + y) dy dx 0 2/3 5 Z 1/2 6 y2 2 = x y+ 1 dx 5 2 y=2/3 0 Z 1/2 2 5 6 x + = dx 5 3 18 0 1/2 6 x3 5x + = 5 9 18 0 = 11 60 Part b: Compute P (Y ≤ X) Solution: This is Z 1 Z 1 6 2 (x + y) dy dx 0 x 5 Z 1 6 y2 2 = x y+ 1 dx 2 y=x 0 5 Z 1 6 1 x2 2 3 = x + −x − dx 2 2 0 5 1 6 x 3 x x4 = + − 5 6 2 4 0 = 1 2 Part c: Compute the cdf of (X, Y ). 4 Solution: This is Z P (X ≤ x and Y ≤ y) = = = = = x Z y 6 2 (s + t) dt ds 0 0 5 Z x 6 2 t2 s t+ y dx 2 t=0 0 5 Z x 6 2 y2 s y+ dx 2 0 5 x 6 s3 y sy 2 + 5 3 2 s=0 3 2 2x3 y + 3xy 2 6 x y xy + = 5 3 2 5 Part d: Compute the margnial (individual distribution) of X, then do the same for Y . Solution: For X, the pdf is Z 1 6 6 2 y2 2 (x + y)dy = f (x) = x y+ 5 2 0 5 1 = 6 2 3 x + for 0 ≤ x ≤ 1 5 5 = 2 6 + y for 0 ≤ y ≤ 1 5 5 y=0 For Y , the pdf is Z f (y) = 0 1 6 2 6 (x + y)dx = 5 5 x3 + xy 3 1 x=0 (in both cases the pdf is 0 outside this range). 5. Let X and Y be two variables whose pdf is given by −(x+y) e if x, y ≥ 0 f (x, y) = 0 else Part a: Compute P (X + Y ≤ 1). Solution: If X + Y ≤ 1, then 0 ≤ Y ≤ 1 − X. So the probability in question is Z 1 Z 1−x e−(x+y) dy dx 0 0 Z 1 1−x = −e−(x+y) y=0 dx 0 Z 1 = e−x − e−1 dx 0 = −e−x − e−1 x = 1− 1 x=0 ( since e−1 is a constant ) 2 e Part b: Suppose that Z = X + Y . Compute the cdf of Z. 5 Solution: We want P (X + Y ≤ z), so we do the same calculation as part a, but with the 1s in the upper limits replaced by z, getting Z z Z z−x e−(x+y) dy dx P (Z ≤ z) = Z0 z 0 z−x −e−(x+y) y=0 dx = Z0 z e−x − e−z dx = 0 = −e−x − e−z x = z x=0 ( since e−z is a constant ) 1 − e−z − ze−z Part c: What is the pdf of Z? Solution: Now that we have the cdf, we just need to take a derivative using the product rule. f (z) = F 0 (z) = e−z − (e−z − ze−z ) = ze−z 6