Name: ID: 18440: Probability and Random variables Quiz 2 Solutions • You will have 50 minutes to complete this test. • No calculators, notes, or books are permitted. • If a question calls for a numerical answer, you do not need to multiply everything out. (For example, it is fine to write something like (0.9)7!/(3!2!) as your answer.) • Don’t forget to write your name on the top of every page. • Please show your work and explain your answer. We will not award full credit for the correct numerical answer without proper explanation. Good luck! 1 2 Name: ID: Problem 1 (35 points) 3 types of particles (first,second and third type) are created according to N1 (t), . . . , N3 (t), t ≥ 0 independent Poisson processes with λ = 1. P • (5 points) What is the law of 3i=1 Ni (1) ? (give the name and density) • (5 points) What is the probability that N1 (1) + N2 (2) equal 10 ? • (5 points) Write down the density function for the amount of time until a first particle (of any kind) is created. • (10 points) Compute the probability that after one hour 3 particles are created knowing that no particles of the first kind were. • (10 points) What is the conditional probability distribution of (N1 (1) + N2 (2))4 , knowing N1 (1) + N2 (2) = 5N3 (1) ? Proof. (1) The sum of independent Poisson random variables is Poisson with parameter equal to P the sum of parameters. Hence 3i=1 Ni (1) ∼ P oisson(3), and so P( 3 X Ni (1) = k) = i=1 e−3 3k . k! (2) As above we have that N1 (1) + N2 (2) ∼ P oisson(1 + 2), hence P (N1 (1) + N2 (2) = 10) = e−3 310 . 10! (3) Let T1 be the first arrival time of a particle from N1 (t) + N2 (t) + N3 (t). We know that N1 (t) + N2 (t) + N3 (t) is a Poisson process with rate the sum of the rates, in this case 3. Also the first arrival time of a Poisson process with rate λ is distributed according to expon(λ). So we conclude that ( 3e−3t if t > 0 fT1 (t) = 0 otherwise. (4) The question asks to find P (N1 (1) + N2 (1) + N3 (1) = 3|N1 (1) = 0). Using Bayes rule and independence we get P (N1 (1) + N2 (1) + N3 (1) = 3|N1 (1) = 0) = P (N1 (1) + N2 (1) + N3 (1) = 3, N1 (1) = 0) = P (N1 (1) = 0) P (N2 (1) + N3 (1) = 3, N1 (1) = 0) P (N2 (1) + N3 (1) = 3)P (N1 (1) = 0) = = P (N1 (1) = 0) P (N1 (1) = 0) P (N2 (1) + N3 (1) = 3) = e−2 23 . 3! (5) We have that N1 (1) and N2 (2) are non-negative integer random variables, so (N1 (1)+N2 (2))4 is a random variable supported on {k 4 |k ∈ Z≥0 }. So to determine the distribution it suffices to find P ((N1 (1)+N2 (2))4 = k 4 |N1 (1)+N2 (2) = 5N3 (1)) = P (N1 (1)+N2 (2) = k|N1 (1)+N2 (2) = 5N3 (1)). Using Bayes rule and independence we get P (N1 (1)+N2 (2) = k|N1 (1)+N2 (2) = 5N3 (1)) = P (N1 (1) + N2 (2) = k, N1 (1) + N2 (2) = 5N3 (1)) = P (N1 (1) + N2 (2) = 5N3 (1)) P (N1 (1) + N2 (2) = k, k = 5N3 (1)) P (N1 (1) + N2 (2) = k)P (5N3 (1) = k) = . P (N1 (1) + N2 (2) = 5N3 (1)) P (N1 (1) + N2 (2) = 5N3 (1)) 3 Now we estimate the denominator and get B = P (N1 (1) + N2 (2) = 5N3 (1)) = ∞ X P (N1 (1) + N2 (2) = 5m, N3 (1) = m) = m=0 ∞ X P (N1 (1) + N2 (2) = 5m)P (N3 (1) = m) = m=0 ∞ X e−3 m=0 35m −1 1m e (5m)! m! On the other hand we have for the numerator A(k) = P (N1 (1) + N2 (2) = k)P (5N3 (1) = k) = ( 35m −1 1m e m! if k = 5m e−3 (5m)! 0 if 5 - k . We thus conclude that the conditional distribution is P ((N1 (1) + N2 (2))4 = k 4 |N1 (1) + N2 (2) = 5N3 (1)) = A(k) , B and 0 otherwise. 4 Name: ID: Problem 2 (20 points) Let X1 , X2 , X3 , X4 be independent uniform variables on [0, 1]. Write Y = X1 + X2 + X3 and Z = X1 + X2 . • (5 points) Compute the density function of Z. • (5 points) Compute the probability that {Y < 1/2}. • (5 points) Compute Cov(Z, Y ). • (5 points) Are Z and Y independent ? Why ? Proof. (1) Since Xi are supported in [0, 1] we know X1 + X2 is supported in [0, 2]. Next we have that R 1 Ry−1 fX1 (a)fX2 (y − a)da = 2 − y if 2 ≥ y ≥ 1 y fZ (y) = 0 fX1 (a)fX2 (y − a)da = y if y ≤ 1 0 otherwise (2)We have Z 1/2 Z 1/2−x1 Z 1/2−x1 −x2 P (Y < 1/2) = P (X1 + X2 + X3 < 1/2) = dx3 dx2 dx1 = 0 Z 0 1/2 Z 1/2−x1 0 1/2 Z 0 0 (1/2 − x1 − x2 )dx2 dx1 = Z 1 1 1/2 2 1 2 (1/2 − x1 ) dx1 = u du = . 2 2 0 48 0 (3) By linearity of expectation we know E[Z] = E[X1 ] + E[X2 ] = 1/2 + 1/2 = 1 and similarly E[Y ] = 3/2. Next we have Z 1Z E[ZY ] = 0 0 1Z 1 Z 1Z 1 (x1 + x2 )(x1 + x2 + x3 )dx3 dx2 dx1 = 0 0 Z 0 1 0 1 (x1 + x2 )2 + (x1 + x2 )dx2 dx1 = 2 3 7 1 3 7 20 x21 + x1 + = + + = . 2 12 3 4 12 12 Consequently, Cov(Z, Y ) = E[ZY ] − E[Z]E[Y ] = 5 3 1 − = . 3 2 6 (4) It is clear that Y and Z are not independent, as otherwise their covariance would be 0, which we saw to not be the case. Intuitively, we have Y > Z so knowing Z we have a lower bound on Y so the two cannot be independent. 5 Name: ID: Problem 3 (25 points) Consider a sequence of independent tosses of a coin that is biased so that it comes up heads with probability 2/3 and tails with probability 1/3. Let Xi be one if the i-th toss is head, and 0 otherwise. (1) (5 points) Compute E[Xi ] and V ar(Xi ). (2) (5 points) Compute V ar(X1 + 2X2 + 3X3 + 4X4 + 5X5 ) P (3) (5 points) Let Y = 40000 Xi . Approximate the probability that Y ≥ 27667. Pi=1 P 20000 (4) (5 points) Let Z = i=1 Xi − 40000 i=20001 Xi . What can you say about the distribution of Z/200? (5) (5 points) Show that (Y /200, Z/200) are approximately independent. Proof. (1) Set p = 2/3. Then E[Xi ] = 1 × P ( heads on ith toss }) + 0 = p, V ar(Xi ) = E[Xi2 ] − E[Xi ]2 = E[Xi ] − E[Xi ]2 = p(1 − p). (2)We use that for independent random variables X, Y one has V ar(aX + bY ) = a2 V ar(X) + b2 V ar(Y ), to get in view of part (1) V ar(X1 + 2X2 + 3X3 + 4X4 + 5X5 ) = 5 X i2 V ar(Xi ) = p(1 − p)(1 + 4 + 9 + 16 + 25) = 55p(1 − p). i=1 (3) We have using part (1) and linearity of expectation and variance for independent variables 80000 80000 ≈ 26667 as well as V ar(Y ) = 40000 × p(1 − p) = ≈ 8889. E[Y ] = 40000 × p = 3 9 Consequently we have by De Moivre - Laplace Y − 26667 P( √ ≥ a) ≈ 1 − Φ(a), 8889 where Φ denotes the Gaussian cdf, i.e. Z y 1 2 √ e−x /2 . Φ(y) = 2π −∞ Consequently, we have Y − 26667 27667 − 26667 √ P (Y ≥ 27667) = P ( √ ≥ ) ≈ 1 − Φ(a), 8889 8889 where a = √1000 . 8889 P20000 P40000 (4)Let T1 = i=20001 Xi . Similarly to our work above we know that i=1 Xi and T2 = E[Ti ] = 40000/3 = a and V ar(Ti ) = 40000/9 = b. By De Moivre Laplace we know that the distribution of Ti − 40000/3 , 200/3 is approximately Gaussian(0, 1) = N (0, 1). Consequently the distribution of Ti /200 is approximately N (200/3, 1/9). Also it is clear Ti are independent as they depend on disjoint collection of tosses. Thus Z/200 = T1 − T2 ≈ A − B, 6 where A, B are independent N (200/3, 1/9). Using that if X ∼ N (µ1 , σ12 ) and Y ∼ N (µ2 , σ22 ), one has aX + bY ∼ N (aµ1 + bµ2 , a2 σ12 + b2 σ22 ), we conclude that A − B ∼ N (0, 2/9). So the distribution of Z/200 is approximately N (0, 2/9). (5) We have Y /200 = T1 /200+T2 /200 so we have that the distribution of Y /200 is approximately N (400/3, 2/9). We know that two Gaussian random variables are independent if and only if the covariance is 0. Since Z/200 and Y /200 have approximately Gaussian distributions they would be approximately independent if their covariance is 0. But we have 1 1 Cov(Y /200, Z/200) = (E[Y Z]−E[Y ]E[Z]) = (E[(T1 +T2 )(T1 −T2 )]−E[T1 +T2 ]E[T1 −T2 ]) = 40000 40000 1 (E[T12 ] − E[T22 ] − E[T1 ]2 + E[T2 ]2 ) = 0, 40000 where in the last line we used that T1 and T2 have the same distribution and hence the same moments. 7 Name: ID: Problem 4 (10 points) Let X and Y be continuous random variables with joint probability density ( 10a3 + 21b5 on a + b ≤ 1, a ≥ 1, b ≥ 0 fX,Y (a, b) = 0 otherwise. (1) (5 points) Find the probability distribution of Z = 2Y + X. (2) (5 points) Find the probability distribution of W = (X + Y )2 + 1. Proof. (1): We calculate the density of Z. Since X, Y are supported on [0, 1] we know Z is supported on [0, 3]. Then we have Z ∞ Z min(1,x/2) fZ (x) = fX,Y (x − 2b, b)db = 10(x − 2b)3 + 21b5 db. 0 max(0,(x−1)/2) Setting B = max(0, (x − 1)/2) and A = min(1, x/2), above becomes 7A6 − 7B 6 + (−20A4 + 20B 4 ) + 10x3 (A − B) + (−60)x2 (A2 /2 − B 2 /2) + 40x(A3 − B 3 ). 2 (2): We calculate the density of W . Again we observe W is supported on [1, 2]. Then we have for x ≥ 1 Z ∞ √ fW (x) = fX,Y (a, x − 1 − a)da, 0 otherwise fW is 0. We calculate the above to be Z min(1,√x−1) √ 3 10a + 21( x − 1 − a)5 da. √ max(0, x−1−1) √ Setting A = min(1, x − 1) and B = max(0, x − 1 − 1) we get the above equals √ 5 X 10 4 5 (−1)i 21( x − 1)5−i i+1 4 (A − B ) + (A − B i+1 ). 4 i i+1 √ i=0 8 Name: ID: Problem 5 (10 points) A model for the movement of a stock supposes that if the present price of the stock is s, then after one period, it will be either us with probability p, or ds with probability 1 − p. Assuming that successive movements are independent, approximate the probability that the stock’s price will be up at least 30 percent after the next 1000 periods if u = 1.012, d = 0.990, p = .52. Proof. Let X be the number of up jumps of the stock price and 1000 − X the number of down jumps. X is Binomial(1000,p). Then we have that the price of the stock after 1000 periods is Sf inal = uX d1000−X s. Consequently, we are interested in P (Sf inal ≥ 1.3s) = P (uX d1000−X ≥ 1.3), which upon taking logs is just P (X log(u) + (1000 − X) log(d) ≥ log(1.3)). We fix log(u) = a log(d) = b. Then our probability becomes P (Xa + (1000 − X)b ≥ log(1.3)) = P (X ≥ log(1.3) − 1000b ). a−b Using De Moivre - Laplace formula we know that X − 1000p P(p ≥ y) ≈ 1 − Φ(y), 1000p(1 − p) where Z Φ(y) = Consequently we have that if C = y 1 2 √ e−x /2 . 2π −∞ log(1.3)−1000b then a−b X − 1000p C − 1000p P (X ≥ C) = P ( p ≥p ) ≈ Φ(C 0 ), 1000p(1 − p) 1000p(1 − p) where C − 1000p . C0 = p 1000p(1 − p) This suffices for the proof.