PTE 5TH EDITION SOLUTIONS MANUAL HOIL LEE AND WONJUN SEO Contents 1. Measure Theory 3 1.1. Probability Theory 1.2. Distributions 4 1.3. Random Variables 5 1.4. Integration 7 1.5. Properties of Integral 9 1.6. Expected Value 12 1.7. Product Measures, Fubini’s Theorem 16 2. 3 Laws of Large Numbers 18 2.1. Independence 18 2.2. Weak Laws of Large Numbers 21 2.3. Borel-Cantelli Lemmas 26 2.4. Strong Law of Large Numbers 3. Central Limit Theorems 36 37 3.1. The De Moivre-Laplace Theorem 37 3.2. Weak Convergence 37 3.3. Characteristic Functions 47 3.4. Central Limit Theorems 58 3.5. Local Limit Theorems 64 3.6. Poisson Convergence 64 3.7. Poisson Processes 64 ∗ 3.8. Stable Laws 3.9. Infinitely Divisible Distributions∗ 3.10. 4. Limit Theorems in R 64 64 d 65 Martingales 67 4.1. Conditional Expectation 67 4.2. Martingales, Almost Sure Convergence 70 4.3. Examples 72 4.4. p Doob’s Inequality, Convergence in L , p > 1 Date: September 9, 2021. 1 76 4.6. Uniform Integrability, Convergence in L1 4.7. Backwards Martingales 80 4.8. Optional Stopping Theorems 81 4.9. Combinatorics of Simple Random Walk 84 5. Markov Chains 78 84 5.1. Examples 84 5.2. Construction, Markov Properties 86 5.3. Recurrence and Transience 89 5.4. Recurrence of Random Walks 89 5.5. Stationary Measures 89 5.6. Asymptotic Behavior 89 5.7. Periodicity, Tail σ-Field 89 5.8. 6. General State Space 89 Ergodic Theorems 89 6.1. Definitions and Examples 89 6.2. Birkhoff’s Ergodic Theorem 92 6.3. Recurrence 93 6.5. Applications 7. Brownian Motion 95 95 7.1. Definition and Construction 95 7.2. Markov Property, Blumenthal’s 0-1 Law 95 7.3. Stopping Times, Strong Markov Property 96 7.4. Path Properties 96 7.5. Martingales 96 7.6. 8. Itô’s Formula 96 Applications to Random Walk 96 8.3. Donsker’s Theorem 96 8.4. Empirical Distributions, Brownian Bridge 96 8.5. 9. Laws of Iterated Logarithm 96 Multidimensional Brownian Motion 96 9.4. Martingales 96 9.5. Dirichlet Problem 96 2 1. Measure Theory 1.1. Probability Theory. Exercise 1.1.1. (i) F is a σ-algebra. From the definition, ∅, Ω are in F, and F is closed under operation of complement. Now, suppose A1 , A2 , · · · are in F. If all An ’s are countable, then ∪∞ n=1 An , which is countable union of countable sets, is also countable, and thus is in F. Otherwise, c ∞ ∪∞ n=1 An is countable, and thus ∪n=1 An is in F. (ii) P is a probability measure. By definition, P (A) ≥ P (∅) = 0, for any A ∈ F. For disjiont A1 , A2 , · · · ∈ F, if all An ’s are countable, then P (∪∞ n=1 An ) = 0 = ∞ X P (An ). n=1 If one of An is uncountable(say Ak ), then since Ack is countable and Aj ⊂ Ack for j 6= k, c c (∪∞ n=1 An ) ⊂ Ak and Aj ’s are countable for j 6= k. Thus, X P (∪∞ An . n=1 An ) = 1 = P (Ak ) + n6=k Therefore, (Ω, F, P ) is a probability space. Exercise 1.1.2. Let Od = {(a1 , b1 ) × · · · × (ad , bd ) : ai , bi ∈ R}. It suffices to show that σ(Od ) = σ(Sd ). Since, (a1 , b1 ] × · · · × (ad , bd ] = (a1 , b1 ) × · · · × (ad , bd ) = ∞ \ (a1 , b1 + n=1 ∞ [ (a1 , b1 − n=1 1 1 ) × · · · × (ad , bd + ). n n 1 1 ) × · · · × (ad , bd − ) n n Exercise 1.1.3. C = {(a1 , b1 ) × · · · × (ad , bd ) : rational numbers a < b}. Exercise 1.1.4. Let F = ∪i Fi . (i) Clearly, ∅, Ω ∈ F. If A ∈ F, then there exists j such that A ∈ Fj . Then, Ac ∈ Fj , and thus Ac ∈ Fj . Let A and B are in F. Then, there exist j1 , j2 such that A ∈ Fj1 and B ∈ Fj2 . WLOG, let j1 ≤ j2 . Since Fn are increasing sequence, A is also in Fj2 . Thus, A ∪ B ∈ Fj ⊂ F. Therefore, F is an algebra. j n ∀ (ii) Let Ω = (0, 1] and In = {( j−1 2n , 2n ], for j = 1, . . . , 2 } for n. Define Fn ≡ σ(In ) for n = 1, 2, · · · 2(j−1) 2j−1 j 2j−1 2j Since ( j−1 2n , 2n ] = ( 2n+1 , 2n+1 ] ∪ ( 2n+1 , 2n+1 ], Fn ⊂ Fn+1 . Now, (0, 1) = ∞ [ 2n − 2 2n − 1 ], ( n , 2 2n n=1 3 n n 2 −1 ∀ / Fn for ∀ n, (0,1) ∈ / F. where ( 2 2−2 n , 2n ] ∈ Fn for n. However, since (0,1) ∈ Therefore, F is not a σ-algebra. Exercise 1.1.5. (Counterexample) Let A be the set of even numbers. Then, A ∈ A with θ = 21 . Let I0 = {1}, Ik = {n : 2k−1 < n ≤ 2k }, for k = 1, · · · . Construct B as follows. c B = ∪∞ k=0 {(I2k ∩ A ) ∪ (I2k+1 ∩ A)} Then, B ∈ A with θ = 12 . However, A ∩ B ∈ / A for following reasons. (1) [n = 22m ] |(A ∩ B) ∩ {1, 2, . . . , 22m }| = Pm 1 k=1 2 |I2k−1 |. |(A ∩ B) ∩ {1, 2, . . . , 22m }| = 22m = = Pm 1 k=1 2 |I2k−1 | 22m 1+ Pm k=2 22m 22m−1 +1 3 22m 22k−3 → 1 , 6 (2) [n = 22m+1 ] Similarly, |(A ∩ B) ∩ {1, 2, . . . , 22m+1 }| = 22m+1 = = By (1), (2), lim sup |(A∩B)∩{1,2,...,n}| ≥ n Pm+1 1 k=1 2 |I2k−1 | 22m+1 1+ Pm+1 2k−3 k=2 2 22m+1 22m+1 +1 3 22m+1 → 1 3 |(A∩B)∩{1,2,...,n}| 1 3 , lim inf n ≤ 1 6. Thus, A ∩ B ∈ / A. Therefore, A is neither an algebra nor a σ-algebra. 1.2. Distributions. Exercise 1.2.1. To show that Z is a r.v., it suffices to show that for any B ∈ B(R), Z −1 (B) ∈ F. Z −1 (B) = [Z −1 (B) ∩ A] ∪ [Z −1 (B) ∩ Ac ] = [X −1 (B) ∩ A] ∪ [Y −1 (B) ∩ Ac ] ∈ B(R), since X, Y are r.v.s, and A ∈ F. Exercise 1.2.2. Omit. Exercise 1.2.3. Let A be the discontinuity set of distribution function F . Since F is rightcontinuous, for x ∈ A, F (x− ) < F (x) = F (x+ ) holds. For x < y who are in A, we have 4 F (x− ) < F (x) ≤ F (y − ) < F (y). Pick rationals qx and qy such that qx ∈ (F (x− ), F (x)) and qy ∈ (F (y − ), F (y)). Then, qx < qy , and thus |A| ≤ ℵ0 . Therefore, there exist at most countably many discontinuities. Exercise 1.2.4. Let F −1 (y) = inf{x : y ≤ F (x)}. Then followings hold. (i) F (F −1 (y)) = y for y ∈ (0, 1). (ii) F −1 is strictly increasing. (iii) {F −1 (F (X)) ≤ F −1 (y)} = {X ≤ F −1 (y)} (i) comes from continuity of F , and (ii) comes from monotonicity of F . For (iii), {F −1 (F (X)) ≤ F −1 (y)} = {F (X) ≤ y} = {X ≤ F −1 (by (ii)) (y)}. (by definition) Thus, for 0 < y < 1, P (F (X) ≤ y) = P (F −1 (F (X)) ≤ F −1 (y)) (by (ii)) = P (X ≤ F −1 (y)) = F (F −1 (by (iii)) (y)) =y (by (i)) For y = 0, P (F (X) ≤ 0) = limn P (F (X) ≤ 1 n) = limn For y = 1, P (F (X) ≤ 1) = limn P (F (X) ≤ 1 − 1 n) 1 n = 0. = limn 1 − 1 n = 1. Thus, F (X) = Y ∼ U nif (0, 1). Exercise 1.2.5. Omit. Exercise 1.2.6. Omit. Exercise 1.2.7. Omit. 1.3. Random Variables. Exercise 1.3.1. We want to show that σ(X) = σ(X −1 (A)). (⊃) By definition of σ(X) = {{X ∈ B} : B ∈ S}, it is clear. (⊂) It suffices to show that X is σ(X −1 (A))-measurable. Firstly, for any σ-field D on Ω, T = {B ∈ S : {X ∈ B} ∈ D} is also a σ-field on S. Construct T with D = σ(X −1 (A)). Then, T is a σ-field containing A. Since A generates S, S = T , which means that X is σ(X −1 (A))-measurable. Exercise 1.3.2. {X1 + X2 < x} = {X1 < x − X2 } [ = {X1 < 1 < x − X2 } q∈Q = [ [{X1 < q} ∩ {X2 < x − q}] ∈ F. q∈Q 5 Exercise 1.3.3. Since f is continuous, f (Xn ) and f (X) are r.v. Moreover, on Ω0 = {ω : Xn (ω) → X(ω)}, f (Xn )(ω) → f (X)(ω), which means that Ω0 ⊂ Ω00 = {ω : f (Xn )(ω) → f (X)(ω)}, since f is continuous. Thus, P (Ω0 ) ≥ P (Ω) = 1, that is, f (Xn ) → f (X) a.s. Exercise 1.3.4. (i) Let f be a continuous function from Rd → R and let O, Od be col- lections of open sets in R, Rd , respectively. Since pre-image of open set in R under f is also open(by continuity), σ(f −1 (O)) ⊂ σ(Od ) = B(Rd ). Also, by Exercise 1.3.1, σ(f −1 (O)) = f −1 (B(R)). Thus, f is B(Rd ), B(R) -measurable. (ii) Let A be the smallest σ-field that makes all the continuous functions measurable. By (i), we showed that A ⊂ B(Rd ). To show the converse, define fi ((x1 , . . . , xd )) = xi , for i = 1, · · · , d, who are continuous. Then, fi−1 ((ai , bi )) ∈ B(Rd ), and thus, f1−1 ((a1 , b1 )) ∩ · · · ∩ fd−1 ((ad , bd )) = (a1 , b1 ) × · · · × (ad , bd ), for any ai , bi ∈ R. This means that any open boxes in Rd is in A. Therefore, B(Rd ) ⊂ A. Exercise 1.3.5. Let f be a lower semicontinuous function. f is l.s.c if and only if for any > 0, there exists δ > 0 such that f (y) > f (x) − for y ∈ B(x, δ). We want to show that f is l.s.c if and only if {x : f (x) ≤ a} is closed for each a ∈ R. (⇒) Suppose not. Then, there exists a0 ∈ R such that {x : f (x) ≤ a} is not closed, that is, there exists a limit point x0 of {x : f (x) ≤ a} with f (x0 ) > a0 . For = there exists δ > 0 such that f (y) > f (x0 ) − = f (x0 )+a0 2 f (x0 )−a0 , 2 > a0 for y ∈ B(x0 , δ). This contradicts that x0 is a limit point of {x : f (x) ≤ a}. (⇐) Fix x0 . Let’s show that f is l.s.c. at x0 . For any > 0, let a = f (x0 ) − . Then, x0 ∈ {x : f (x) > a}. Since {x : f (x) > a} = {x : f (x) ≤ a}c is an open set, there exists δ > 0 such that B(x0 , δ) ⊂ {x : f (x) > a}. That is, f (y) > f (x) − for y ∈ B(x0 , δ). Thus, f is l.s.c. at x0 . Since x0 is arbitrary, f is l.s.c. For measurability, by Exercise 1.3.1, it suffices to show that f −1 ((a, ∞)) is in B(R). If f is l.s.c, then f −1 ((−∞, a]) = {x : f (x) ≤ a} is closed. Thus, f ∈ B(R). Exercise 1.3.6. Let’s show that f δ = sup{f (y) : |y−x| < δ} is l.s.c. By Exercise 1.3.5, it suffices to show that Aa ≡ {x : f δ (x) ≤ a} is closed for any a ∈ R. Let x0 be a limit point of Aa . For any y ∈ B(x0 , δ), there exists zy ∈ Aa different from x0 such that |x0 − zy | < δ − |x0 − y|. Since |zy − y| ≤ |zy − x0 | + |x0 − y| < δ and z ∈ Aa , f (y) ≤ a. Since this holds for all y ∈ B(x0 , δ), sup{f (y) : |y − x0 | < δ} ≤ a. That is, x0 ∈ Aa , which implies that Aa is closed. Similar argument gives us that fδ is u.s.c. Since semicontinuous functions are measurable (Exercise 1.3.5), f 0 and f0 , limits of semicontinuous functions, are also measurable. Thus, ∞ [ 1 0 0 {f 6= f0 } = x : f (x) − f0 (x) > n n=0 is measurable. (Clearly, f 0 ≥ f0 ) 6 Exercise 1.3.7. (1) Obviously, simple functions are F-measurable. And, for any fn ∈ F, lim sup fn is also F-measurable. (2) Any F-measurable function is a pointwise limit of simple functions. By (1) and (2), class of F-measurable functions is the smallest class containing the simple functions and closed under pointwise limit. Exercise 1.3.8. f X (Ω, σ(X)) → (R, B(R)) → (R, B(R)) (⇐) Since X is σ(X)-measurable, Y = f (X) is σ(X)-measurable. (Composition of measurable functions is measurable.) (⇒) If one can show that T = {Y = f (X) : f ∈ B(R)} contains simple functions with respect to (Ω, σ(X)) and is closed under pointwise limit, then by previous Exercise 1.3.7, {Y : Y ∈ σ(X)} ⊂ T . (a) Let ϕ be a simple function with respect to (Ω, σ(X)). That is, ϕ(ω) = n X cm IAm (ω) where disjoint Am ∈ σ(X). m=1 Since Am ∈ σ(X), there exist disjoint Bm ∈ B(R)X such that Am = X −1 (Bm ). Thus, ϕ(ω) = = = n X m=1 n X m=1 n X cm IAm (ω) cm IX −1 (Bm ) (ω) cm IBm (X(ω)) m=1 Therefore, T contains simple functions with respect to (Ω, σ(X)). (b) Now, let Y be a pointwise limit of Yn = fn (X) where f ∈ B(R). That is, (?) Y (ω) = limn fn (X(ω)) for all ω ∈ Ω. Define f (x) = lim supn fn (x). Since fn ∈ B(R), f ∈ B(R). Also, Y (ω) = f (X(ω)) for all ω ∈ Ω. Thus, T is closed under pointwise limit. (?) : From the definition of fn and f , we can ensure that fn converges to f only on x ∈ X(Ω). Since we don’t know whether convergence also holds on R\X(Ω) or not, we define f (x) by limsup of fn (x). Exercise 1.3.9. Omit. 1.4. Integration. 7 Exercise 1.4.1. Suppose not. Then, there exists n such that µ(An ) > 0, where An = {f (x) > 1 n }. Since f ≥ 0, Z Z f dµ ≥ f dµ ≥ An 1 µ(An ) > 0, n which is a contradiction. Exercise 1.4.2. Define fn (x) = P∞ m m=1 2n IEn,m (x). By MCT, R fn (x) = P∞ m m=1 2n µ(En,m ). Note that if x ∈ En,m , then x ∈ En+1,2m ∪ En+1,2m+1 . Thus, fn (x) ≤ fn+1 (x). And, since R R f (x) − fn (x) < 21n , fn (x) → f (x) for all x. Thus, by MCT, fn dµ = f dµ. Combining two results, we get the desired result. R R (i) Let g ≥ 0. Since gdm = sup{ sdm : 0 ≤ s ≤ f, s simple}, there R R R R exists simple 0 ≤ s ≤ g such that gdm − < sdm. Thus, |g − s|dm = (g − s)dm < Exercise 1.4.3. . For integrable g = g + − g − , find simple s+ and s− such that Z Z |g − − s− |dm < . |g + − s+ |dm < 2 2 Then, Z |g − (s+ − s− )|dm ≤ Z |g − s+ |dm + Z |g − s− |dm < . s+ − s− is the desired simple function. (ii) Firstly, let ϕ = IA , where A is measurable. By outer regularity of Lebesgue measure, U∞ there exists open G ⊃ A such that m(G − A) > 2 . Since G = i=1 (ai , bi ), by continuity Un of measure, there exists n such that m(G\ i=1 (ai , bi )) < 2 . By relabeling (ai , bi ) for i = 1, . . . , n, we can construct a step function q such that Z Z Z |ϕ − q|dm ≤ |IA − IG |dm + |IG − q|dm = m(G − A) + m(G\ ]ni=1 (ai , bi )) < . Now, let ϕ = Pn k=1 bk IAk . For each k, there exists step function qk such that Z . |IAk − qk |dm < Pn k=1 |bk | Then, Z |ϕ − n X bk qk |dm = Z X n k=1 |bk ||IAk − qk |dm k=1 ≤ n X k=1 8 Z |bk | |IAk − qk |dm < . P And we can construct a step function from (iii) For given , pick sufficiently small positive δ k bk qk . < Pk |c | . j j=1 Define a continuous function r as follows. On x ∈ [aj−1 + δ, aj − δ], r(x) = cj . On x ∈ [aj − δ, aj ], connect (aj − δ, cj ) and (aj , 0) continuously. On x ∈ [aj , aj + δ], connect (aj , 0) and (aj + δ, cj+1 ) continuously. Then, Z |q − r|dµ < δ n X |cj | < . j=1 Exercise 1.4.4. Suppose g is a step function. Then, Z aj Z k X cj cos nxdx g(x) cos nxdx = aj−1 j=1 = ≤ k X 1 cj (sin(naj ) − sin(naj−1 )) n j=1 k X 2|cj | j=1 n → 0. For integrable g, previous exercise says there exists a step function q such that R |g − q|dx < , for given > 0. Then, Z Z |g(x) cos nx|dx = |g(x) cos nx − q(x) cos nx + q(x) cos nx|dx Z Z ≤ |g − q|| cos nx|dx + |q(x) cos nx|dx Z Z ≤ |g − q|dx + |q(x) cos nx|dx Z < + |q(x) cos nx|dx → . Since q is a step function, it goes to 0 as n increases. Letting → 0, we get the desired result. 1.5. Properties of Integral. Exercise 1.5.1. Our claim is that if ||g||∞ = M , then |g| ≤ M a.e. Suppose not. Then, µ({x : |g(x)| > M }) = µ(∪∞ n=1 {x : |g(x)| > M + Thus, there exists N ∈ N such that µ({x : |g(x)| > M + 1 N }) 1 }) > 0. n > 0, which contradicts on definition of ||g||∞ . Therefore, |f g| ≤ |f |||g||∞ a.e. By integrating, we get the desired result. Exercise 1.5.2. As shown in the previous exercise, we know that |f | ≤ ||f ||∞ a.e. Then, R |f |p dµ ≤ ||f ||p∞ , and thus lim supp ||f ||p ≤ ||f ||∞ . For the converse, let B = {x : |f (x)| > 9 ||f ||∞ − }. Then, for sufficiently small > 0, 0 < µ(B ≤ 1. Note that Z Z p |f (x)| dµ ≥ |f (x)|p dµ B Z > (||f ||∞ − )p dµ B = µ(B (||f ||∞ − )p Letting p → ∞ and letting → 0+ give us lim inf p ||f ||p ≥ ||f ||∞ Exercise 1.5.3. (i) Z Z ||f + g||pp = |f + g|p dµ ≤ 2p (|f |p + |g|p )dµ = 2p ||f ||pp + 2p ||g||pp < ∞. Thus, ||f + g||p < ∞. If ||f + g||p = 0, then clearly ||f + g||p ≤ ||f ||p + ||g||p . Suppose not. Applying Holder Inequality to |f ||f + g|p−1 and |g||f + g|p−1 with p, q = Z |f ||f + g|p−1 dµ ≤ ||f ||p Z p p−1 , p−1 p p (|f + g|p−1 ) p−1 dµ = ||f ||p ||f + g||pp−1 p−1 Z Z p p |g||f + g|p−1 dµ ≤ ||g||p (|f + g|p−1 ) p−1 dµ = ||g||p ||f + g||p−1 p Then, ||f + g||pp = Z Z ≤ |f + g|p dµ (|f | + |g|)|f + g|p−1 dµ ≤ (||f ||p + ||g||p )||f + g||pp−1 . Since ||f + g||p < ∞, we get ||f + g||p ≤ ||f ||p + ||g||p . (ii) (p = 1) |f + g| ≤ |f | + |g| (p = ∞) Let ||f ||∞ = M1 and ||g||∞ = M2 . Since {x : |f (x)| + |g(x)| > M1 + M2 } ⊂ {x : |f (x)| + |g(x)| > M1 + M2 } ⊂ {x : |f (x)| > M1 } ∪ {x : |g(x)| > M2 }, µ({x : |f (x)| + |g(x)| > M1 + M2 }) ≤ 0. Thus, ||f + g||∞ ≤ M1 + M2 . Pn n Exercise 1.5.4. Let fn := m=0 f 1Em = f 1]m=0 Em . Then, fn → f 1E and |fn | ≤ f : integrable. By DCT, we get the desired result. Exercise 1.5.5. Let fn := gn + g1− . Since gn ↑ g, −gn− ≥ −g1− . Thus, fn = gn+ − gn− + g1− ≥ 0. By MCT, Z Z fn dµ = Z gn dµ + g1− dµ → 10 Z Z gdµ + g1− dµ. Since R g1− dµ < ∞, by substracting, we get the desired result. Exercise 1.5.6. MCT Exercise 1.5.7. (i) MCT (ii) Step 1 g ≥ 0. By (i), for given , there exists N such that R R gdµ < (g ∧ N )dµ + 2 . Pick δ = 2N . Then, for A with µ(A) < δ, Z Z gdµ < (g ∧ N )dµ + ≤ N µ(A) + < . 2 2 A A Step 2 g : integrable. |g| = g + + g − . For given , there exists N such that Z Z + g dµ < (g + ∧ N )dµ + 4 Z Z − − g dµ < (g ∧ N )dµ + . 4 Pick δ = 4N . Then, for A with µ(A) < δ, Z Z Z + |g|dµ = g dµ + g − dµ A A A Z Z < (g + ∧ N )dµ + (g − ∧ N )dµ + 2 A A ≤ 2N µ(A) + < . 2 Exercise 1.5.8. Let c ∈ (a, b). Since f is integrable, by previous exercise(1.5.7), for given R > 0, there exists δ > 0 such that if µ(A) < δ, then A |f |dµ < . For |x − c| < δ, Z x Z x |g(x) − g(c)| = f (y)dy ≤ |f (y)|dy < . c c Since c is arbitrary in (a, b), g is continuous on (a, b). Exercise 1.5.9. (i) (1 ≤ p < ∞) Step 1 f ≥ 0. There exists (ϕn )n∈N such that 0 ≤ ϕn ≤ f and ϕn ↑ f . Let gn = |f − ϕn |p . gn → 0 and R R |gn | ≤ (|f | + |ϕn |)p ≤ 2p |f |p : integrable. By DCT, gn dµ = |f − ϕn |p → 0. Thus, ||ϕn − f ||p → 0. Step 2 Integrable f = f + − f − . − There exist (ϕ+ n )n∈N , (ϕn )n∈N such that + + 0 ≤ ϕ+ and ϕ+ n ≤f n ↑f − − 0 ≤ ϕ− and ϕ− n ≤f n ↑f . − For ϕn ≡ ϕ+ n − ϕn , + − − ||ϕn − f ||p ≤ ||ϕ+ n − f ||p + ||ϕn − f ||p → 0. 11 R P∞ P∞ R Pm Exercise 1.5.10. By MCT, n=1 |fn |dµ = n=1 |fn |dµ < ∞. For gm ≡ n=1 fn , P∞ Pm P∞ gm → g ≡ n=1 fn and |gm | ≤ n=1 |fn | ≤ n=1 |fn | : integrable. By DCT, Z Z X Z Z X m m Z ∞ X gm dµ = fn dµ = fn dµ → gdµ = fn dµ. n=1 n=1 n=1 1.6. Expected Value. Exercise 1.6.1. For a support line l(x) = a(x − x0 ) + ϕ(x0 ) at x0 of ϕ(x) (l(x) ≤ ϕ(x)), strict convexity implies that (∗) l(x) < ϕ(x) if x 6= x0 . For x0 = EX, Eϕ(X) − El(X) = Eϕ(X) − l(EX) = Eϕ(X) − ϕ(EX) = 0 implies ϕ(X) = l(X) a.s. since ϕ(X) ≥ l(X). Thus, X = EX a.s. (∗) : Suppose l(y) = ϕ(y) for y 6= x0 . Then, with λ = 21 , x y 1 1 0 ϕ + > ϕ(x0 ) + ϕ(y) 2 2 2 2 1 1 = l(x0 ) + l(y) 2 2 x y 0 =l + 2 2 which contradicts that l(x) is a support line of ϕ(x). The last equality comes from the linearity of l(x). Exercise 1.6.2. Let l(x1 , . . . , xn ) = ϕ(EX1 , . . . , EXn )+gx> (x1 −EX1 , . . . , xn −EXn ), where gx ∈ ∂f (x). Then, ϕ(x) ≥ l(x) for any x = (x1 , . . . , xn ) ∈ Rn . Taking expectation gives us the desired result. Eϕ(X) ≥ El(X) = ϕ(EX). Exercise 1.6.3. (i) Let ϕ(x) = x2 and A = (−∞, −a] ∪ [a, ∞). Then, it is clear that iA = inf{ϕ(y) : y ∈ A} = a2 . By Chebyshev’s inequality, we have b2 . a2 If some random variable X attains equality above, then X has to satisfy following a2 P (|X| ≥ a) ≤ EX 2 = b2 ⇒ P (|X| ≥ a) ≤ identities. b2 dP = 2 , a A Z b2 dP = 1 − 2 , a Ac Z It is clear that X with P (X = a) = b2 a2 , Z x2 dP = b2 . 2 P (X = 0) = 1− ab 2 satisfies identities. Therefore, inequality is sharp. (ii) For a > 0, it’s clear that if x ∈ {x : |x| ≥ a}, then x2 ≥ a2 . Thus, a2 1(|X|≥a) ≤ X 2 1(|X|≥a) = X 2 − X 2 1(|X|<a) . Since EX 2 is finite, both functions on each sides are integrable. Thus, linearity of integration and order-preserving property of integration give us Z Z Z a2 1(|X|≥a) dP ≤ X 2 dP − X 2 1(|X|<a) dP. 12 We know that LHS = a2 P (|X| ≥ a). Let’s focus on RHS. The first term in RHS is clearly EX 2 . For the second term, we can apply MCT. (It is clear that conditions of MCT hold.) By MCT, Z lim a2 P (|X| ≥ a) ≤ EX 2 − lim X 2 1(|X|<a) dP a→∞ a→∞ Z = EX 2 − lim X 2 1(|X|<a) dP a→∞ = EX 2 − EX 2 = 0 The last equality holds since EX 2 < ∞. Therefore, we have lim a2 P (|X| ≥ a)/EX 2 = 0. a→∞ However, Chebyshev’s inequality with ϕ(x) = x2 and A = (−∞, −a] ∪ [a, ∞), so that iA = a2 , gives us following inequality. a2 P (|X| ≥ a)/EX 2 ≤ 1, which is not sharp. Exercise 1.6.4. (i) Let A = [a, ∞), ϕ(y) = (y + b)2 where 0 < b < a. It is clear that iA = inf{ϕ(y) : y ∈ A} = (a + b)2 . Then by Chebyshev’s inequality, (a + b)2 P (Y ≥ a) ≤ E[ϕ(Y )] = E[Y 2 + 2bY + b2 ] = E[X 2 + 2bX + b2 ] = E[ϕ(X)]. The second equality holds since EY = EX, V ar(Y ) = V ar(X). Note that E[(X + b)2 ] = p(a + b)2 . Thus, P (Y ≥ a) ≤ p. If Y = X, then equality holds. (ii) Let p = σ2 (a2 +σ 2 ) , b= p 1−p a, A = [a, ∞), and ϕ(y) = (y + b)2 . Note that 0 < p < 1 and b > 0. Then, it is clear that iA = (a + b)2 . By Chebyshev’s inequality, (a + b)2 P (Y ≥ a) ≤ E[ϕ(Y )] = E[Y 2 + 2bY + b2 ] = σ 2 + b2 , since E(Y ) = 0, V ar(Y ) = E(Y 2 ) − E(Y )2 = E(Y 2 )σ 2 . Thus, we have P (Y ≥ a) ≤ Since b = p 1−p a = σ2 a , σ 2 + b2 . (a + b)2 we have 4 σ 2 + σa2 σ 2 + b2 P (Y ≥ a) ≤ = = 2 (a + b)2 (a + σa )2 σ2 2 2 a2 (a + σ ) 1 2 2 2 a2 (a + σ ) = σ2 . (a2 + σ 2 ) And equality holds when Y has P (Y = a) = p and P (Y = −b) = 1 − p, which satisfies E(Y ) = ap − b(1 − p) = 0 13 V ar(Y ) = E[(Y − 0)2 ] = a2 p + b2 (1 − p) = Exercise 1.6.5. P (|X| > ) ≤ σ4 a2 σ 2 + 2 = σ2 . 2 +σ a + σ2 a2 (i) It suffices to show that ∀ n, ∃ X such that EX = 0, V ar(X) = 1 and 1 n. Construct X as follows. 1 − 1 , n P (X = x) = 1, if x = 0 √ if x = ± n 2n Then, EX = 0, V ar(X) = 1 and P (|X| > ) ≤ P (X 6= 0) = 1 n. (ii) Similarly, 1 − 1 , n P (X = x) = 1, 2n if x = 1 if x = 1 ± √ nσ Then, EX = 1, V ar(X) = σ 2 and P (|X| > y) ≤ P (X 6= 1) = 1 n. Exercise 1.6.6. Let f = Y and g = I(Y >0) . Then, E|f g| = E|Y I(Y >0) = EY I(Y >0) = EY E|f |2 = EY 2 E|g|2 = EI(Y >0) = P (Y > 0). By Cauchy-Schwarz Ineq., (E|f g|)2 ≤ E|f |2 E|g|2 . Thus, (EY )2 ≤ EY 2 P (Y > 0). Exercise 1.6.7. Checking (i), (ii) in Theorem 1.6.8 is obvious. For (iii), Z 2 Eg(Xn ) = |Xn | α dm Z 1 1 dm = n2 I( n+1 ,n ) = n < 1 for all n. n+1 Thus, Theorem 1.6.8 can be applied. Suppose that Y ≥ |Xn | = Xn . Then, Z ∞ X Y dm ≥ nα = n=1 ∞ X 1 n(n + 1) nα−1 =∞ n+1 n=1 The last equality holds since 0 < α−1 < 1. Therefore, Xn are not dominated by an integrable function. Exercise 1.6.8. Step 1 g : Indicator function. Let g(x) = IA (x) where A ∈ B(R). Then, Z Z g(x)µ(dx) = IA (x)µ(dx) 14 Z = µ(A) Z g(x)f (x)dx = IA (x)f (x)d(x) Z = f (x)dx A = µ(A). Step 2 g : simple function. Linearity of Integral. Step 3 g ≥ 0. There exists non-negative simple function sn ↑ g. Then, Z Z g(x)µ(dx) = lim sn (x)µ(dx) n→∞ Z (MCT) = lim sn (x)µ(dx) n→∞ Z Step 2 = lim sn (x)f (x)dx n→∞ Z (MCT) = g(x)f (x)dx. Step 4 g : Integrable. g = g + − g − . By Step 3, R g(x)µ(dx) = R g(x)f (x)dx. Exercise 1.6.9. Induction. Exercise 1.6.10. Induction. (i) Let A = (|X| > 1). Then, |X|j < |X|k for 0 < j < k on A. Z Z Z |X|j dP = |X|j dP + |X|j dP c A ZA k ≤ |X| dP + 1 A Z ≤ |X|k dP + 1 < ∞. Exercise 1.6.11. k (ii) Let ϕ(x) = x j I(x≥0) . It is easy to check that ϕ is convex. Let Y = |X|j . Then, |EY | = E|X|j < ∞(by (i)) and |Eϕ(Y )| = E|X|k < ∞. By Jensen’s Ineq., k E|X j | j = ϕ(EY ) ≤ Eϕ(Y ) = E|X|k . j Thus, we get E|X|j ≤ (E|X|k ) k . Exercise 1.6.12. P eX = ym = P (X = log ym ) = p(m). EeX = n X p(m)ym m=1 15 e EX n X = exp ! p(m) log ym m=1 p(m) = exp log Πnm=1 ym p(m) = Πnm=1 ym . By Jensen’s Ineq., EeX ≥ exp(EX). And, we get the desired result. Exercise 1.6.13. Exercise 1.5.5 Exercise 1.6.14. (i) E(1/X; X > y) = E(1/X; 1/X < 1/y) ≤ 1 1 P (1/X < 1/y) = P (X > y). y y Thus, limy→∞ yE(1/X; X > y) ≤ limy→∞ P (X > y) = 0. (ii) Let 0 < y < . yE(1/X; X > y) = E(y/X; X > y and 0 ≤ X < ) + E(y/X; X > y and ≤ X). ≤ P (X > y, 0 ≤ X < ) + E(y/X; ≤ X). ≤ P (0 < X < ) + E(y/X; ≤ X). For the second term on right hand side, y X I≤X ≤ y I≤X ≤ 1 and y X I≤X → 0 as y → 0. Thus, by DCT, lim sup yE(1/X; X > y) ≤ P (0 < X < ). y→0 Letting → 0, we get the desired result. Exercise 1.6.15. Exercise 1.5.6 Exercise 1.6.16. Let Xk = XIUk n=0 An . Then, Xk → XIA and |Xk | ≤ |X| : integrable. By DCT, EXk → E(X; A). Note that Z k Z X EXk = U XdP = k n=0 An Thus, E(X; A) = limk→∞ EXk = n=0 P∞ n=0 XdP = An k X E(X; An ). n=0 EXk . 1.7. Product Measures, Fubini’s Theorem. Exercise 1.7.1. Since |f (x, y)| ≥ 0, by Fubini’s Theorem, R X×Y |f |dµ = R R X Y |f (x, y)|µ2 (dy)µ1 (dx) < ∞. And by Fubini’s Theorem again, we get the desired result. Exercise 1.7.2. Construct a probability space (Y, B, ν) = ([0, ∞), B([0, ∞)), λ), where λ denotes a Lebesgue measure. Let f (x, y) = I{(x,y) : 0≤y<g(x)} . Then, Z Z Z f dλdµ = gdµ X Y X Z f d(µ × λ) = (µ × λ)({(x, y) : 0 ≤ y < g(x)}) X×Y 16 Z Z ∞ Z f dµdλ = Y µ({x : g(x) > y}dy. X 0 Since f ≥ 0, Fubini’s Theorem implies that Z Z Z Z Z f dλdµ = f d(µ × λ) = f dµdλ. X Y X×Y Y X And we get the desired result. ExerciseZ1.7.3. (i) Fubini’s Z Theorem with f (x, y) = I{(x,y):a<x≤y≤b} ≥ 0. (ii) F (y)dG(y) + G(x)dF (x) (a,b] (a,b] Z Z (a,b] (i) = Z {F (y) − F (a)}dG(y) + = {G(x) − G(a)}dF (x) + (a,b] Z Z F (a)dG(y) + (a,b] G(a)dF (x) (a,b] Z I{(x,y):a<x≤y≤b} d(µ × ν) + I{(x,y):a<y≤x≤b} d(µ × ν) + F (a)(G(b) − G(a)) + G(a)(F (b) − F (a)) = (µ × ν)((a, b] × (a, b]) + (µ × ν)({(x, y) : a < x = y ≤ b} + F (a)G(b) + F (b)G(a) − 2F (a)G(a) X = (F (b) − F (a))(G(b) − G(a)) + µ({x})ν({x}) + F (a)G(b) + F (b)G(a) − 2F (a)G(a) x∈(a,b] = F (b)(G(b) − F (a)G(a) + X µ({x})ν({x}). x∈(a,b] (iii) If F=G, then the third term in right hand side of (ii) equals zero. Thus, Z 2F (y)dF (y) = F 2 (b) − F 2 (a). (a,b] Exercise 1.7.4. Let f = I{(x,y):x<y≤x+c} . Then, Z Z Z f dµ(y)dx = (F (x + c) − F (x))dx ZR ZR ZR f dxdµ(y) = cdµ(y) R R R = cµ(R). Since f ≥ 0, Fubini’s Theorem gives us that Z Z Z Z f dµ(y)dx = f dxdµ R R R R a ∞ and we get the desired result. Exercise 1.7.5. Z 0 a Z ∞ |e−xy sin x|dydx = 0 Z 0 Z 0 ≤ | sin x| dx x a 1dx < ∞. 0 17 e−xy | sin x|dydx 0 a = Z Z By Exercise 1.7.1, e−xy sin xI{(x,y):0<x<a,0<y} is integrable and we can apply Fubini’s Theorem. Z 0 a sin x dx = x Z a ∞ Z e−xy sin xdydx = 0 0 ∞ Z e−xy sin xdxdy 0 0 ∞ Z a Z = 0 e−xy (−y sin x − cos x) 1 + y2 a dy 0 ∞ 1 − e−ay (y sin a + cos a) dy 1 + y2 0 Z ∞ −ay Z ∞ −ay ye π e = − cos a dy − sin a dy. 2 2 1 + y 1 + y2 0 0 R 1 −1 The last equality comes from the fact that 1+y (y). Moreover, 2 dy = tan Z ∞ −ay Z a Z ∞ −ay e sin x π ye = cos a dx − dy + sin a dy 2 x 2 1 + y 1 + y2 0 0 0 Z ∞ | cos a| + y| sin a| ≤ e−ay dy 1 + y2 0 Z ∞ 1+y ≤ e−ay dy 1 + y2 0 Z ∞ 2 ≤ e−ay · 2dy = . a 0 Z = Letting a → ∞, we get Z ∞ 0 Note that since sin x x π sin x dx = . x 2 is not integrable on R≥0 , we can’t calculate its integration directly. 2. Laws of Large Numbers 2.1. Independence. Exercise 2.1.1. By Thm 2.1.8, it suffices to show that P (X1 ≤ z1 , . . . , Xn ≤ zn ) = R∞ Πni=1 P (Xi ≤ zi ). Let ci = −∞ gi (xi )dxi , for i = 1, . . . , n. Then, we have c1 · · · cn = 1. Our claim is that P (Xi ≤ zi ) = 1 ci Z zi gi (xi )dxi . −∞ WLOG, let i = 1. P (X1 ≤ z1 ) = P (−∞ < X1 ≤ z1 , −∞ < X2 < ∞, . . . , −∞ < Xn < ∞) Z z1 = (c2 · · · cn ) g1 (x1 )dx1 −∞ 1 = c1 Z z1 g1 (x1 )dx1 . −∞ Now, Z z1 P (X1 ≤ z1 , . . . , Xn ≤ zn ) = Z g1 (x1 )dx1 · · · −∞ zn −∞ 18 gn (xn )dxn Z zn Z 1 1 z1 g1 (x1 )dx1 · · · gn (xn )dxn c1 −∞ cn −∞ n Y = P (Xi ≤ zi ) = i=1 Exercise 2.1.2. Let Ai = {Xi−1 (xi ) : xi ∈ Si } ∪ {∅, Ω}. Let I = {n1 , . . . , nk } ⊂ {1, . . . , n} and J = {1, . . . , n} \I = {m1 , . . . , mn−k }. Then, X P (Xn1 = xn1 , · · · , Xnk = xnk ) = ··· xm1 ∈Sm1 = Y X P (X1 = x1 , · · · , Xn = xn ) xmn−1 ∈Smn−k P (Xi = xi ). i∈I By Lemma 2.1.5, Ai ’s are independent. By Theorem 2.1.7, σ(Ai )’s are independent. Sine Xi−1 (A) ∈ σ(Ai ) for all A ∈ 2Si , σ(Xi ) ⊂ σ(Ai ). Therefore, X1 , . . . , Xn are independent. Exercise 2.1.3. (1) h(ρ(x, y)) → [0, ∞) (i) h(ρ(x, y)) = 0 ⇐⇒ ρ(x, y) = 0 ⇐⇒ x = y. First ⇐⇒ holds since h(0) = 0 and h0 (x) > 0 for x > 0. (ii) h(ρ(x, y)) = h(ρ(y, x)) by symmetry of metric ρ. Z ρ(x,z) (iii) h(ρ(x, z)) = h0 (u)du 0 Z ρ(x,y)+ρ(y,z) h0 (u)du ≤ 0 Z ≤ ρ(x,y) 0 Z ρ(y,z) h0 (u)du h (u)du + 0 0 = h(ρ(x, y)) + h(ρ(y, z)). The first inequality holds since ρ is metric. The second inequality holds since h0 is decreasing. (2) By simple calculus, one can show that h(x) = x x+1 satisfies (1). Exercise 2.1.4. It is easy to check the uncorrelatedness. To show that Xn ’s are not independent, it suffices to show that there exist A and B such that P (Xn ∈ A, Xm ∈ B) 6= P (Xn ∈ A)P (Xm ∈ B) for different m and n. A = X −1 (0, ) n For 0 < < 1, let . Then, A ⊂ B. Note that we have 2n intervals in B = X −1 X (A) m m 1 . This is possible since A. Let l be the length of one interval. Now, pick so that 2nl < m P (A) & 0 as & 0. Then, it is clear that P (B) < 1. Thus, P (A ∩ B) = P (A) > P (A)P (B). Therefore, Xn and Xm are not independent. Exercise 2.1.5. (i) Let Dµ be a set of discontinuities of µ. By Exercise 1.2.3, Dµ is at most countable. Z Z P (X + Y = 0) = 1(x+y=0) dµdν 19 Z = µ({−y})dν Z Z = µ({−y})dν + µ({−y})dν. c Dµ Dµ Note that for −y ∈ Dµc , µ({−y}) = P (X ≤ −y) − P (X < −y) = 0. Thus, Z P (X + Y = 0) = µ({−y})1Dµ (y)dν R X = µ({−y})ν({y}). y∈Dµ The second equality holds by Exercise 1.5.6. (ii) With Y = −Y in (i), P (X = Y ) = 0 since Dµ = ∅. Exercise 2.1.6. Note that σ(f (X)) = {X −1 (f −1 (B)) : B ∈ B(R)} σ(f (Y )) = {Y −1 (g −1 (B)) : B ∈ B(R)} Since f and g are measurable, f −1 (B) and g −1 (B) are in B(R) for any B ∈ B(R). Thus, σ(f (X)) ⊂ σ(X) and σ(g(Y )) ⊂ σ(Y ). Therefore, independence of X and Y implies the independence of f (X) and g(Y ). Exercise 2.1.7. By Exercise 2.1.2, it suffices to show that P (Zi = α, Zj = β) = P (Zi = α)P (Zj = β) for all α, β ∈ {0, 1, . . . , K − 1}. Firstly, for given Y = y, P (Zi = α|Y = y) = P (Zi = α) = K−1 X 1 K since K is a prime. Thus, P (Zi = α|Y = y)P (Y = y) y=0 1 1 = . 2 K K Now, for given Y = y, Zi = α and Zj = β means that X + iy ≡ α and X + jY ≡ β. That is, =K· (i − j)y ≡ α − β (mod K). Since gcd(i − j, K) = 1, there exists unique y0 ∈ {0, 1, . . . , K − 1} satisfying Zi = α and Zj = β. And, with y0 , x0 ∈ {0, 1, . . . , K − 1} uniquely determined. Thus, P (Zi = α, Zj = β) = 1 . K2 Therefore, P (Zi = α, Zj = β) = 1 1 1 = · = P (Zi = α)P (Zj = β) for all α, β ∈ {0, 1, . . . , K − 1} . 2 K K K i.i.d Exercise 2.1.8. X1 , X2 , X3 ∼ 21 1{1} + 12 1{−1} and X4 = X1 X2 X3 . n o n o Exercise 2.1.9. A1 = {1, 2}, Ω and A2 = {1, 3}, {1, 4}, {2, 3}, {2, 4}, Ω 20 Exercise 2.1.10. P (X + Y = n) = X P (X = m, Y = n − m) = m X P (X = m)P (Y = n − m) m The last equality holds by the independence. Exercise 2.1.11. Omit. Exercise 2.1.12. Omit. Exercise 2.1.13. (a) F (z) = 91 1[0,1) (z) + 92 1[1,2) (z) + 31 1[2,3) (z) + 29 1[3,4) (z) + 19 1[4,∞) (z) (b) For some c ∈ 0, 29 0 0 1 2 2 9 1 2 1 9 c −c 1 9 c −c 1 9 c 2 9 2 9 −c Exercise 2.1.14. Z Z P (XY ≤ z) = 1(xy≤z) dF dG Z Z 1(x≤z/y,y6=0) + 1(0≤z,y=0) dF dG Z z = F dG + P (Y = 0)I(z ≥ 0). y R\{0} = Exercise 2.1.15. For simplicity, let Ω = [0, 1). We have 0, if ω ∈ 2m , 2m+1 : m = 0, 1, . . . , 2n−1 − 1 2n 2n Yn (ω) = 1, if ω ∈ 2m+1 , 2m+2 : m = 0, 1, . . . , 2n−1 − 1 2n . 2n By construction, it is obvious that P (Yn = 0) = P (Yn = 1) = 1 2. To show that Y1 , Y2 , . . . are independent, we have to show that Y1 , . . . , Yn are independent for all n ∈ N. By Exercise Qn 2.1.2, it suffices to show that P (Y1 = y1 , . . . , Yn = yn ) = i=1 P (Yi = yi ), where yi ∈ {0, 1}. By induction, one can show that {ω : Y1 (ω) = y1 , . . . , Yn (ω) = yn } = 2ln , l+1 for some l ∈ 2n Qn 1 n {0, 1, . . . , 2 − 1} for any yi ’s. Therefore, P (Y1 = y1 , . . . , Yn = yn ) = 2n = i=1 P (Yi = yi ), where yi ∈ {0, 1}. 2.2. Weak Laws of Large Numbers. Exercise 2.2.1. We assume that EXi2 < ∞ so that Var(Xi ) < ∞ for all i. Since L2 convergence implies convergence in probability, it suffices to show that E|Sn /n − νn |2 → 0 as n → ∞. Note that E|Sn /n − νn |2 = 1 E|Sn − ESn |2 n2 21 1 Var(Sn ) n2 n 1 X = 2 Var(Xi ) n i=1 = n ≤ 1 X Var(Xi ) n i=1 i n) The last equality holds since Xi ’s are uncorrelated. Now, fix > 0. Since Var(X → 0, there PN Var(Xi ) n Var(Xn ) exists N ∈ N such that < for all n > N . Let M = i=1 < ∞. Then, for n i n > N, ! n n X 1 1 X Var(Xi ) < M+ n i=1 i n i=N +1 1 ≤ ·M + n By taking limsup and letting → 0, we get E|Sn /n − νn |2 → 0, which is the desired result. Exercise 2.2.2. We assume that r(0) is finite. Let Sn = X1 + · · · + Xn . Since ESn = 0, " # 2 Sn Sn E = Var n n 1 X EXi Xj = 2 n 1≤i,j≤n n n−1 n−2 2 X X X 1 X ≤ 2 EXi2 + 2 EXi Xi+1 + 2 EXi Xi+2 + · · · + 2 EXi Xi+n−2 + 2EX1 Xn n i=1 i=1 i=1 i=1 Now, since r(k) → 0 as k → ∞, for given > 0, ∃ N such that for n ≥ N , |r(n)| < . Moreover, since EXi2 ≤ r(0) < ∞, by Cauchy-Schwartz Inequality, E|Xi Xj | ≤ (EXi2 )1/2 (EXj2 )1/2 ≤ r(0) < ∞. Then, for n ≥ N + 1, we have " # 2 Sn E n n n−1 n−N n−N X X+1 X 1 X 2 ≤ 2 EXi + 2 E|Xi Xi+1 | + · · · + 2 E|Xi Xi+N −1 | + 2 EXi Xi+N + · · · + 2EX1 Xn n i=1 i=1 i=1 i=1 ≤ n−N n−N X X−1 1 nr(0) + 2(n − 1)r(0) + · · · + 2(n − N + 1)r(0) + 2 r(N ) + 2 r(N + 1) + · · · + 2r(n − 1) 2 n i=1 i=1 ≤ n−N n−N X X−1 1 2N nr(0) + 2 |r(N )| + 2 |r(N + 1)| + · · · + 2|r(n − 1)| 2 n i=1 i=1 22 ≤ n−N X 1 2N nr(0) + 2 i · n2 i=1 ≤ n X 1 2N nr(0) + 2 i· 2 n i=1 ≤ 2N r(0) n(n + 1) + n n2 Since > 0 is arbitrary, E h Sn 2 n i → 0, which means that Sn /n → 0 in L2 . Therefore, Sn /n → 0 in probability, which is the desired result. Exercise 2.2.3. (i) Since Ui ’s are i.i.d. and f is a measurable function, f (Ui )’s are also R1 i.i.d. Note that since Ui follows Uniform distribution on [0,1], Ef (U1 ) = 0 f (x)dx. By R1 assumption, E|f (U1 )| = 0 |f (x)|dx < ∞. By WLLN, Z 1 f (U1 ) + · · · + f (Un ) In = → Ef (U1 ) = f (x)dx = I in prob. n 0 R1 (ii) Since 0 |f (x)|2 dx < ∞, Z 1 2 Z 1 2 2 2 Var(f (U1 )) = Ef (U1 ) − (Ef (U1 )) = f (x) dx − f (x)dx < ∞. 0 0 2 By Chebyshev’s inequality with ϕ(x) = x and A = {x ∈ R : |x| > a/n1/2 }, we have iA = inf{ϕ(y) : y ∈ A} = a2 /n and n E(In − I)2 a2 n = 2 Var(In ) a 1 = 2 Var(f (U1 )) a "Z Z 1 2 # 1 1 2 = 2 f (x) dx − f (x)dx a 0 0 P (|In − I| > a/n1/2 ) ≤ Exercise 2.2.4. (1) E|Xi | = ∞. ∞ R∞ 1 P∞ P∞ 1 Since k=2 k log k=2 k ≥ 2 x log x dx = log(log x) 2 = ∞, E|Xi | = C k log k = ∞. (2) There exists µ < ∞ so that Sn /n → µ in probability. Since nP (|X1 | > n) ≤ n ∞ X k=n+1 Z ∞ ∞ X n C n C C C ≤ ≤ dx = , 2 2 2 k log k log n k log n n x log n k=n+1 nP (|X1 | > n) → 0 as n → ∞. Also, there exists µ < ∞ such that µn := E(X1 1(|X1 |≤n) ) = ∞ X ∞ (−1)k · k · k=2 X C C = (−1)k →µ k 2 log k k log k k=2 since it is alternating series. Therefore, by Theorem 2.2.12, Sn /n → µ in probability. 23 Exercise 2.2.5. By definition, Xi takes only positive values. (1) There exists a sequence µn → ∞ such that Sn /n − µn → 0 in probability. Firstly, for x ≥ 2, xP (|Xi | > x) = xP (Xi > x) = 2 log x → 0 as x → ∞. Let µn = E(X1 1(|X1 |≤n) ). Then, for n ≥ 3, Z n e µn = dx = [e log(log x)]ne = e log(log n) → ∞. x log x e Therefore, by Theorem 2.2.12, Sn /n − µn → 0 in probability. (2) E|Xi | = ∞. Since E|Xi | = EXi ≥ µn for all n, E|Xi | = ∞. P∞ Exercise 2.2.6. (i) X = n=1 1(X≥n) . ∞ X x2 P (X = x) (ii) EX 2 = x=1 = P (X = 1) + (1 + 3)P (X = 2) + · · · + (1 + 3 + · · · + (2n − 1))P (X = n) + · · · = P (X ≥ 1) + 3P (X ≥ 2) + · · · + (2n − 1)P (X ≥ n) + · · · = ∞ X (2n − 1)P (X ≥ n). n=1 Exercise 2.2.7. Z ∞ Z ∞ Z h(y)P (X ≥ y)dy = −∞ h(y)1(X≥y) (ω)dP dy −∞ Z Z Ω ∞ = h(y)1(X≥y) (ω)dP dy Ω −∞ Z = H(X)dP = EH(X). Ω The second equality holds by Fubini. Exercise 2.2.8. Let’s use Theorem 2.2.11 in Textbook with Xn,k = Xk for k = 1, . . . , n and bn = 2m(n) where m(n) = min m : 2−m m−3/2 ≥ n−1 . Let’s check the conditions. (0) bn > 0 with bn → ∞. By definition of m(n). Pn (i) k=1 P (|Xn,k | > bk ) → 0. n X ∞ X P (|Xn,k | > bk ) = nP (|X1 | > 2m(n) ) = n k=1 pj j=m(n)+1 = ∞ X j=m(n)+1 n ≤ 2j · j(j + 1) ∞ X j=m(n)+1 1 n = m(n) ≤ m(n)− 2 → 0 2 m(n)2 24 n 2j · m(n)2 (ii) b−2 n Pn k=1 2 E X̄n,k → 0. b−2 n n X 2 E X̄n,k = k=1 n E X12 1{|X1 |≤2m(n) } 22m(n) m(n)3/2 E X12 1{|X1 |≤2m(n) } m(n) 2 m(n) 2 m(n)3/2 X k 2 − 1 pk + p0 . = m(n) 2 k=1 ≤ (1) Note that m(n) X 2k − 1 2 m(n) pk + p0 ≤ 1 + X k=1 22k pk k=1 m(n) X =1+ 1 k(k + 1) 2k · k=1 X =1+ 1≤k< m(n) 2 X ≤1+ 1≤k< =1+2 ≤ 2k · 1 + k(k + 1) 2k + m(n) 2 m(n) 2 +1 2m(n) ·C m(n)2 −2+ X 2k · m(n) 2 ≤k≤m(n) 4 m(n) · (m(n) + 2) 1 k(k + 1) X 2k · m(n) 2 ≤k≤m(n) m(n) 4 2m(n)+1 − 2 2 +1 m(n) · (m(n) + 2) for large n. Thus, 3 1 m(n) 2 2m(n) · · C = C · m(n)− 2 → 0 2 m(n) m(n) 2 Pn p n By (0)-(ii), Theorem 2.2.11 gives us that Snb−a → 0, where an = k=1 E X̄n,k . n (1) ≤ Now, an = nE(X1 1{|X1 |≤2m(n) } m(n) X = n (2k − 1)pk − p0 k=1 1 + m(n) + 1 ∞ X 1 2k · k(k + 1) k=m(n)+1 ∞ X 1 1 1 ≤ n − + m(n) + 1 m(n)2 2k = n − k=m(n)+1 25 1 k(k + 1) ∼− n . m(n) Since n ≤ 2m(n) m(n)3/2 ≤ Cn, log n ∼ m(n). Thus, n an ∼ log n bn = 2m(n) ∼ n(m(n))−3/2 ∼ n(log n)−3/2 . Thus, Sn + logn n p Sn − an ∼ 3 → 0. bn n(log n)− 2 1 Since (log n)− 2 → 0, Sn + n(log n)−1 p → 0. n(log n)−1 Finally, Sn p → 0. n/ log n Exercise 2.2.9. 2.3. Borel-Cantelli Lemmas. Exercise 2.3.1. Since ∪k≥n Ak ⊃ Am for all m ≥ n, P (∪k≥n Ak ) ≥ supk≥n P (Ak ). Thus, P (lim sup An ) = P (∩∞ n=1 ∪k≥n Ak ) = lim P (∪k≥n Ak ) ≥ lim sup P (Ak ) = lim sup P (An ). n→∞ n→∞ k≥n By applying above result to Bn = Acn , we get P (lim inf An ) ≤ lim inf P (An ). Exercise 2.3.2. > 0 given. Since f is uniformly continuous on [−M − 1, M + 1], there exists δ ∈ (0, 1) such that if x, y ∈ [−M − 1, M + 1] and |x − y| < δ, then |f (x) − f (y)| < . Then, |f (x) − f (y)| ≥ implies |x − y| ≥ δ or |x| ≥ M . Thus, P (|f (Xn ) − f (X)| ≥ ) ≤ P (|Xn − X| ≥ δ) + P (|X| ≥ M ). Since Xn → X in probability, lim sup P (|f (Xn ) − f (X)| ≥ ) ≤ P (|X| ≥ M ). n By letting M → ∞, we get lim P (|f (Xn ) − f (X)| ≥ ) = 0. n Exercise 2.3.3. We know that P (ln = k) = 1 , 2k+1 1, 2n k = 0, 1, . . . , n − 1 k = n. (1) lim sup logln n = 1 a.s. 2 26 (i) lim sup logln n ≤ 1 a.s. 2 Fix > 0. Let An = {ln > (1 + ) log2 n} and m = b(1 + ) log2 nc. Then, P (An ) = 1 − P (ln ≤ (1 + ) log2 n m+1 n X 1 1 =1− = 2k+1 2 k=1 (1+) log2 n 1 = n−(1+) . ≤ 2 P∞ Since 1 + > 1, we get n=1 P (An ) < ∞. By Borel-Cantelli Lemma 1, ln P (An i.o.) = P > 1 + i.o. = 0, log2 n that is, P ln ≤ 1 + eventually log2 n = 1. Since > 0 was arbitrarily chosen, we get lim sup logln n ≤ 1 a.s. 2 (ii) lim sup logln n ≥ 1 a.s. 2 Construct a sequence (rn )n∈N by r1 = 1, r2 = 2, r3 = rn−1 + [log2 n] for r ≥ 3. Let Bn = {Xm = 1, for all m such that rn−1 ≤ m ≤ rn }. Then, rn −rn−1 [log2 n] log2 n 1 1 1 1 = ≥ = . P (Bn ) = 2 2 2 n P∞ Since n=1 P (Bn ) = ∞ and Bn ’s are independent, by Borel-Cantelli Lemma 2, P (Bn i.o.) = 1. Since Bn ⊂ {lrn ≥ [log2 n]}, P (lrn ≥ [log2 n] i.o.) = 1. Note that rn ≤ n log2 o n fornn ≥ 2. (It can be shown easily by induction.) Then, since o n lrn lrn [log2 n] [log2 n] [log2 n] ≥ ⊂ ≥ , we get = log2 rn log2 rn log2 rn log2 (n log2 n) log2 n+log2 log2 n l rn [log2 n] P ≥ i.o. = 1. log2 rn log2 n + log2 log2 n Since limn→∞ [log2 n] log2 n+log2 log2 n = 1, lim sup (2) lim inf ln log2 n ln ≥ 1 a.s. log2 n = 0 a.s. It suffices to show that P (ln = 0i.o.) = 1. Note that {ln = 0} = {Xn = 0} are independent and P (Xn = 0) = 21 . Thus, by Borel-Cantelli Lemma 2, P (ln = 0 i.o.) = 1. Exercise 2.3.4. Firstly, find a subsequence (Xn(m) )m∈N such that EXn(m) → lim inf EXn . Since Xn → X in probability, there exists a further subsequence Xn(mk ) such that Xn(mk ) → X a.s. by Theorem 2.3.2 in Textbook. Then, by Fatou’s Lemma, EX = E(lim inf Xn(mk ) ) ≤ lim inf EXn(mk ) = lim inf EXn . 27 The inequality comes from Fatou’s Lemma. Exercise 2.3.5. (a) For all subsequence Xnk , there exists a sub-subsequence Xm(nk ) converging almost surely to X. Since |Xm(nk ) | ≤ Y : integrable, by DCT, EXm(nk ) → EX. This means that every subsequence (EXnk ) of (EXn ) has a sub-subsequence EXm(nk ) converging to EX. Therefore, EXn → EX. (b) See Theorem 1.6.8. Exercise 2.3.6. (a) Note that f (x) = f (x) = 0 if and only if x = 0. x 1+x is non-decreasing, 0 ≤ f (x) ≤ 1 for all x ≥ 0 and Firstly, d(X, Y ) = E(f (|X − Y |)) ≥ [0, ∞). (i) d(X, Y ) = E(f (|X − Y |)) = 0 ⇐⇒ f (|X − Y |) = 0 a.s. ⇐⇒ |X − Y | = 0 a.s. (ii) Clear. (iii) Since |X − Z| ≤ |X − Y | + |Y − Z|, |X − Z| = f (|X − Z|) 1 + |X − Z| ≤ f (|X − Y | + |Y − Z|) |X − Y | + |Y − Z| 1 + |X − Y | + |Y − Z| |X − Y | |Y − Z| = + 1 + |X − Y | 1 + |Y − Z| = Therefore, d(X, Z) = E(f (|X − Z|)) ≤ E(f (|X − Y |)) + E(f (|Y − Z|)) = d(X, Y ) + d(Y, Z). (b) (⇒) Fix > 0. Let ϕ(x) = A} = x 1+x ≥ 0 and A = (, ∞). Then, we have iA = inf{ϕ(y) : y ∈ 1+ . By Chebyshev’s inequality, we get 1+ |Xn − X| 1+ P (|Xn − X| > ) ≤ ·E ≤ · d(Xn , X) → 0. 1 + |Xn − X| Thus, Xn → X in probability. (⇐) For any > 0, |Xn − X| d(Xn , X) = E 1 + |Xn − X| |Xn − X| |Xn − X| =E ; |Xn − X| > + E ; |Xn − X| ≤ 1 + |Xn − X| 1 + |Xn − X| ≤ E(1; |Xn − X| > ) + E ; |Xn − X| ≤ 1+ ≤ P (|Xn − X| < ) + . 1+ 28 Since Xn → X in probability, lim supn d(Xn , X) = 1+ . Since > 0 was arbitrarily chosen, by letting → 0, we get limn d(Xn , X) = 0. Exercise 2.3.7. By Theorem 2.3.2 in Textbook, it suffices to show that there exists a r.v. X∞ such that for any subsequence Xn(m) , there exists a further subsequence Xn(mk ) that converges almost surely to X∞ . Since d(Xm , Xn ) → 0, there exists Nk ∈ N such that d(Xm , Xn ) < 2−k if m, n > Nk . Now, fix a subsequence Xn(m) . Choose increasing subsequence (mk ) so that y n(mk ) ≥ Nk . Using Chebyshev’s Inequality with ϕ(y) = 1+y and A = k12 , ∞ , we get 1 k2 + 1 . P Xn(mk ) − Xn(mk−1 ) > 2 ≤ k 2k P∞ 2 P∞ Since k=1 = P Xn(mk ) − Xn(mk−1 ) > k12 ≤ k=1 k 2+1 < ∞, by Borel-Cantelli Lemma k 1, 1 Xn(mk ) − Xn(mk−1 ) > 2 i.o. = 0, k 1 that is, P Xn(mk ) − Xn(mk−1 ) ≤ k2 eventually = 1. For ω ∈ Ω0 := Xn(mk ) − Xn(mk−1 ) ≤ Xn(mk ) (ω) k∈N is a Cauchy sequence in R. Thus, there exists X∞ (ω) := limk Xn(mk ) (ω). P Since P (Ω0 ) = 1, Xnm(k) → X∞ a.s. It remains to check the well-definedness of X∞ . Consider other subsequence Xn0 (m0 ) such 0 0 ) = 0. By Exercise that Xn0 (m0k ) → X∞ . Since d(Xn(mk ) , Xn0 (m0k ) → 0, by DCT, d(X∞ , X∞ 0 . 2.3.6, it’s equivalent to that X∞ = X∞ Exercise 2.3.8. Since P (∪n An ) = 1, P (∩n Acn ) = 0. Moreover, since An ’s are independent events, so are Acn . And this implies that ! ∞ ∞ \ Y c P An = P (Acn ) n=1 = n=1 ∞ Y (1 − P (An )) = 0. n=1 Since P (An ) < 1, Qm−1 n=1 (1 − P (An )) > 0 for m = 2, 3, · · · . Thus, we have for all m ∈ N, ! ∞ \ c P An = 0. n=m Since {Acn eventually} = S∞ m=1 P (Acn T∞ c n=m An , eventually) = P ∞ \ ∞ [ ! Acn m=1 n=m ≤ ∞ X m=1 P ∞ \ ! Acn = 0. n=m Then, P (An i.o.) = 1. Thus, by Borel-Cantelli Lemma 1(Theorem 2.3.1), 29 P∞ n=1 P (An ) = ∞. 1 k2 eventually , (i) Let Bn = Acn ∩ An+1 . Then, Exercise 2.3.9. P (∪m≥n Am ) ≤ P (An ) + ∞ X P (Bm ) → 0 as n → ∞. m=n Thus, P (An i.o.) = limn→ P (∪m≥n Am ) = 0. (ii) An = 0, n1 with Lebesgue measure. Exercise 2.3.10. 10 (i) For 0 < < 1, P (|Xn | > ) = pn → 0 ⇐⇒ pn → 0. P∞ P∞ (ii) Let An = {|Xn | > }. Then, P (An i.o.) = 0 ⇐⇒ n=1 pn < ∞. Note n=1 P (An ) = Exercise 2.3.11. that (⇒) holds since Xn ’s are independent. Exercise 2.3.12. Let’s enumerate Ω = {ω1 , ω2 , · · · }. Let Ω0 = {ωi : P (ωi ) > 0}. Since P (Ωc0 ) = 0, if one can show that Xn (ω) → X(ω) for ω ∈ Ω0 , then it’s done. Fix > 0. Since Xn → X in probability, for each ωi ∈ Ω0 , there exists Ni ∈ N such that P (|Xn − X|) > ) < 1 2 P (ωi ) if n > Ni . Since 12 P (ωi ) < P (ωi ), ωi ∈ / {ω : |Xn − X| > for n > Ni } . Thus, Xn (ωi ) → X(ωi ) as n → ∞. P∞ 1 n > Exercise 2.3.13. Pick cn such that P |Xn | > |cnn | ≤ 21n . Since n=1 P X cn n < ∞, Xn 1 1 n by Borel-Cantelli Lemma 1, P X > ≤ i.o. = 0. That is, P eventually = 1. cn n cn n Now, for given > 0, choose N ∈ N such that 1 N < . Then, since {|Xn | ≤ |cn |} ⊃ {|Xn | ≤ |cn |/N }, Xn 1 Xn ≤ eventually ≥ P ≤ eventually = 1. P cn cn N Since is arbitrarily chosen, Xn /cn → 0 a.s. Exercise 2.3.14. (⇒) Suppose P n P (Xn > A) = ∞ for all A. Then, by Borel-Cantelli Lemma 2, P (Xn > A i.o.) = 1 for all A. That is, sup Xn = ∞ a.s. which is the contradiction. P (⇐) Since n P (Xn > A) < ∞ for some A, by Borel-Cantelli Lemma 1, P (Xn > A i.o.) = 0. Thus, sup Xn ≤ A < ∞ a.s. Exercise 2.3.15. Xn =1 log n Xn P log n > (i) lim supn→∞ a.s. (≤) It suffices to show that 1 + i.o. = 0 for any > 0. By definition, P Xn > (1 + ) log n = n−(1+) . Since 1 + > 1, we have ∞ ∞ X X P Xn > (1 + ) log n = n=1 n=1 1 n1+ < ∞. By Borel-Cantelli Lemma 1, P Xn > (1 + ) log n i.o. = 0. 30 Xn 0 (≥) It suffices to show that P log n > 1 i.o. = 1. Since Xn s are independent and P∞ P∞ 1 n=1 P (Xn > log n) = n=1 n = ∞, by Borel-Cantelli Lemma 2, we get P (Xn > log n i.o.) = 1. (ii) If one can show that (a) 1 ≤ lim inf n→∞ Mn Mn (b) ≤ lim sup ≤ 1 log n n→∞ log n a.s., then it’s done. Mn (a) It suffices to show that P log < 1 − i.o. = 0 for any > 0. Note that n Mn P <1− =P max Xm < (1 − ) log n 1≤m≤n log n n Y = P (Xm < (1 − ) log n) m=1 = 1− 1 n n1− ≤ e−n . The second equality holds since Xi ’s are i.i.d. By L’Hopital’s rule, e−n / n12 → 0. Thus, by comparison test, ∞ X Mn < 1 − < ∞. log n n=1 Mn By Borel-Cantelli Lemma 1, P log < 1 − i.o. = 0. n Mn (b) It suffices to show that P log n < 1 + eventually = 1 for any > 0. P Fix > 0. Let Ω0 = {ω : lim supn→∞ Xn (ω) log n = 1}. (?) For ω ∈ Ω0 , there exists N1 (ω) ∈ N such that Xn (ω) log n < 1 + if n > N1 (ω). Now, for n > N1 (ω), Mn (ω) Xn (ω) = max 1≤m≤n log n log n = max max Xn (ω) Xm (ω) , max 1≤m≤N1 (ω) log n N1 (ω)<m≤n log n - There exists N2 (ω) ∈ N such that max1≤m≤N1 (ω) - By (?), Xm (ω) log n ≤ Xm (ω) log m Xm (ω) log n < 1 + for N1 (ω) < m ≤ n. Thus, for n > N (ω) = max{N1 (ω), N2 (ω)}, we have Mn (ω) < 1 + . log n 31 . < 1 + if n > N2 (ω). n Mn log n P n (1 That is, Ω0 ⊂ Exercise 2.3.16. (1) o < 1 + eventually . Since P (Ω0 ) = 1 by (i), Mn < 1 + eventually = 1. P log n − F (λn )) < ∞. Let Bn = {Xn > λn }. Since 1 − F (λn ) = P (Xn > λn ), by Borel-Cantelli Lemma 1, P (Xn > λn i.o.) = 0. That is, P (Xn ≤ λn eventually) = 1. Note that (?) for ω ∈ Ω0 := {ω : Xn (ω) ≤ λn eventually}, there exists N1 (ω) ∈ N such that Xn (ω) ≤ λn if n > N1 (ω). Now, for n > N1 (ω), max Xm (ω) = max max 1≤m≤n 1≤m≤N1 (ω) Xm (ω), max N1 (ω)<m≤n Xm (ω) . - Since λn ↑ ∞, there exists N2 (ω) ∈ N such that max1≤m≤N1 (ω) Xm (ω) ≤ λn if n > N2 (ω). - By (?), Xm (ω) ≤ λm ≤ λn for N1 (ω) < m ≤ n. Thus, for n > N (ω) = max{N1 (ω), N2 (ω)}, we have max Xm (ω) ≤ λn . 1≤m≤n That is, Ω0 ⊂ {max1≤m≤n Xm ≤ λn eventually}. Since P (Ω0 ) = 1, P (max1≤m≤n Xm ≤ λn eventually) = 1, which means that P max Xm > λn i.o. = 0. 1≤m≤n (2) P n (1 − F (λn )) = ∞. By Borel-Cantelli Lemma 2, P (Xn > λn i.o.) = 1. Then, clearly, P (max1≤m≤n Xn > λn i.o.) = 1. Exercise 2.3.17. The answers are (i) E|Y1 | < ∞, (ii) EY1+ < ∞, (iii) nP(|Y1 | > n) → 0, and (iv) P(|Y1 | < ∞) = 1. (i) If E|Y1 | < ∞, then X Z ∞ ∞ ∞ X Y1 > = P (|Y1 | > n) ≤ P (|Y1 | > t) = E|Y1 | < ∞ P n 0 n=1 n=1 since P(|Y1 | > ·) decreases. By Borel-Cantelli, we have the convergence. Conversely, if E|Y1 | = ∞, then Z ∞ = E|Y1 | = ∞ P(|Y1 | > t) dt ≤ 0 ∞ X P(|Y1 | > n) n=0 since P(|Y1 | > ·) decreases. By the second Borel-Cantelli, the convergence fails. 32 (ii) If EY1+ < ∞, then Yn+ /n → 0 a.s. from (i). Thus for a given > 0, there exists N such that Yn /n+ < for all n > N . Then, maxm≤n Ym+ ≤ max n YN+ Y1+ ,··· , , → 0 n n almost surely as n → ∞. That is, maxm≤n Ym+ =0 n→∞ n lim almost surely. Now we have lim sup n→∞ maxm≤n Ym maxm≤n Ym+ ≤ lim sup = 0. n n n→∞ We claim that lim sup n→∞ maxm≤n Ym ≥ 0. n Note that P maxm≤n Ym < − n = F (−n)n where F is the distribution of Y1 . Let M be large so that F (−n) < δ < 1 for all n > M . Then we have X ∞ M ∞ X X maxm≤n Ym maxm≤n Ym P < − ≤ P < − + δn n n n=1 n=1 n=M +1 M +1 ≤M+ δ 1−δ < ∞. Therefore, by Borel-Cantelli, lim sup n→∞ maxm≤n Ym ≥ − n almost surely. This proves our claim. Conversely, if EY1+ = ∞, then Z ∞ ∞ X ∞ = EY1+ = P(Y1 > t) dt ≤ P(Y1 > n), 0 n=0 and thus Yn /n > 1 infinitely often. (iii) If (iv) By definition, Yn /n → 0 in probability if and only if P(|Yn | > n) → 0 for any > 0. This is equivalent to P(|Y1 | = ∞) = 0. Exercise 2.3.18 and 2.3.19 will be solved by the way we did in Theorem 2.3.9 in Textbook. 33 Exercise 2.3.18. Step 1. Xn anα → 1 in probability. Fix > 0. Since β < 2α, Xn − EXn Var(Xn ) Bnβ B P ≤ ∼ 2 2 · nβ−2α → 0. > ≤ 2 2 2 2 EXn (EXn ) (EXn ) a Thus, Xn −EXn EXn → 0 in probability. Therefore, Xn EXn → 1 in probability. Step 2. Find subsequence (nk ). Let nk = k [2/(2α−β)]+1 (+1 to avoid the case where h 2 2α−β i = 0.) and Tk = Xnk . Then, Bnβk Tk − ETk Var(Tk ) B > ≤ 2 ≤ ∼ 2 2 · nkβ−2α P 2 2 2 ETk (ETk ) (ETk ) a 2 B ≤ 2 2 · k ( 2α−β )·(β−2α) a = Ck −2 P∞ −ETk The last inequality holds by construction of nk . Since k=1 P TkET > < ∞, k by Borel-Cantelli Lemma 1, Tk − ETk > i.o. = 0. P ETk That is, Step 3. Xn anα Tk ETk → 1 a.s. → 1 a.s. For sufficiently large n such that nk ≤ n < nk+1 and ω ∈ Ω0 := {ω : Tk ETk → 1}, we have Tk (ω) Xn (ω) Tk+1 (ω) Tk+1 (ω) ETk+1 ETk Tk (ω) · = ≤ ≤ = · . ETk+1 ETk ETk+1 EXn ETk ETk+1 ETk Note that anα ETk+1 k+1 ∼ = ETk anα k Therefore, we have a.s. Step 4. Xn nα Xn (ω) EXn " 2 +1 #α 1 [ 2α−β ] → 1. 1+ k → 1. Since P (Ω0 ) = 1 by Step 2, we conclude that Xn EXn →1 → a a.s. Xn =a· nα EXn anα Xn · →a EXn a.s. Exercise 2.3.19. Suppose Y ∼ P oisson(λ), where λ > 1. Then, we can decompose Y into Y = Y1 + · · · + Yn , where Yi ’s are independent and Yi ∼ P oisson(λi . Thus, without loss of generality, we may assume that Xn are independent Poisson r.v.’s with EXn = λn , where λn ≤ 1. 34 Step 1. Sn ESn → 1 in probability. P Fix > 0. Since n λn = ∞, P λn Sn − ESn Var(Sn ) 1 P > ≤ 2 = 2 Pn = 2P → 0. 2 2 ESn (ESn ) ( n λn ) n λn Thus, Sn −ESn ESn → 0 in probability. Therefore, Sn ESn → 1 in probability. Step 2. Find subsequence (nk ). P Pn Since n λn = ∞, one can find subsequence (nk ) such that nk = inf{n : i=1 λi ≥ k 2 }. Let Tk = Snk . Then, Tk − ETk Var(Tk ) 1 1 > ≤ 2 = 2 Pnk ≤ 2 2. P ETk (ETk )2 k λ i=1 i P∞ −ETk Since k=1 P TkET > < ∞, by Borel-Cantelli Lemma 1, k Tk − ETk P > i.o. = 0. ETk That is, Step 3. Sn ESn Tk ETk → 1 a.s. → 1 a.s. For n ∈ N such that nk ≤ n < nk+1 , we have k2 ≤ nk X λi = ETk ≤ k 2 + 1 ≤ (k + 1)2 ≤ ETk+1 ≤ (k + 1)2 + 1. i=1 Note that the assumption that λn ≤ 1 is used here. With such n and ω ∈ Ω0 := {ω : Tk (ω) ETk → 1}, ETk Tk (ω) Tk (ω) Sn (ω) Tk+1 (ω) Tk+1 (ω) ETk+1 · = ≤ ≤ = · . ETk+1 ETk ETk+1 ESn ETk ETk+1 ETk Note that 2(k + 1) ETk+1 (k + 1)2 + 1 =1+ → 1. ≤ ETk k2 k2 Therefore, we have Sn (ω) ESn → 1. Since P (Ω0 ) = 1 by Step 2, we conclude that Sn ESn →1 a.s. Exercise 2.3.20. It suffices to show that for any M > 0, P Xn n log2 n ≥ M i.o. = 1. Fix M > 0. Let kn = blog2 (M · n log2 n)c + 1. Then, for n ≥ 2, P (X1 ≥ M · n log2 n) = ∞ X 1 1 = kn −1 2j 2 j=kn = 1 1 ≥ . M · n log2 n 2blog2 (M ·n log2 n)c 35 Then, by Exercise 2.2.4, ∞ X X1 ≥M n log2 n ∞ X 1 = ∞. M · n log2 n n=2 n=2 Xn Since Xn ’s are independent, by Borel-Cantelli Lemma 2, we get P n log P ≥ 2n ≥ M i.o. = 1. 2.4. Strong Law of Large Numbers. i.i.d i.i.d Exercise 2.4.1. Xi , Yi ≥ 0. Xi ∼ F , Yi ∼ G. EX1 , EY1 < ∞. Let Tn := (X1 + Y1 ) + · · · + (Xn + Yn ) and Nt := min {n : Tn ≤ t}. Note that Nt → ∞ as t → ∞. Then, PNt i=1 Xi Rt ≤ ≤1− t PNt i=1 t By SLLN and Theorem 2.4.7, we have PNt PNt i=1 Yi a.s. i=1 Xi a.s. → EX1 , → EY1 , Nt Nt t Yi . Nt a.s. 1 → . t EX1 + EY1 Thus, we get Rt a.s. EX1 → . t EX1 + EY1 Exercise 2.4.2. Let Yn+1 = Xn+1 /|Xn |. Since Yn is independent of X1 , · · · , Xn , (Yn )n is an iid sequence. Note that log |Xn | log |Y1 | + · · · + log |Yn | = n n converges to E log |Y1 | whenever it exists by the Strong Law of Large Numbers. Now to find the expectation, Z E log |Y1 | = B1 1 log |x| dx = π Z 0 1 Z ∂Br 1 log r dS dr = π Z 0 1 1 2r log r dr = − . 2 Remark. For d-dimensional vectors, one can derive that it converges to −1/d by a similar argument. Exercise 2.4.3. (i) Wn = (ap + (1 − p)Vn−1 )Wn−1 = (ap + (1 − p)Vn−1 )(ap + (1 − p)Vn−2 )Wn−2 .. . = n−1 Y (ap + 1 − p)Vi ). i=1 Pn−1 log(ap + (1 − p)Vi ). By SLLN, Pn−1 log(ap + (1 − p)Vi ) a.s. log Wn n−1 = · i=1 → E log(ap + (1 − p)V1 ) =: c(p). n n n−1 Thus, log Wn = i=1 36 (ii) c(p) = R log(ap + (1 − p)ν)µ(dν), where µ : distribution of V . By Theorem A.5.1, 2 Z a−ν 00 c (p) = − µ(dν) < 0. ap + (1 − p)ν (iii) a−ν µ(dν) = −1 + aE(V1−1 ) ν Z 1 a−ν µ(dν) = 1 − EV1 . c0 (1) = a a 0 Z c (0) = To guarantee that the optimal choice of p lies in (0,1), E(V1−1 ) > should hold. (iv) p = 8−5a 2a2 −10a+8 . 3. Central Limit Theorems 3.1. The De Moivre-Laplace Theorem. 3.2. Weak Convergence. Exercise 3.2.1. Omit. Exercise 3.2.2. (i) P Mn /n1/α ≤ y = P Mn ≤ yn1/α h in = P Xi ≤ yn1/α −α n = 1 − yn1/α n y −α = 1− → exp −y −α . n (ii) P n1/β Mn ≤ y = P Mn ≤ yn−1/β h in = P Xi ≤ yn−1/β n |y|β = 1− → exp −|y|β . n (iii) P (Mn − log n ≤ y) = P (Mn ≤ y + log n) n = [P (Xi ≤ y + log n)] n = 1 − e−y−log n n e−y = 1− → exp −e−y . n 37 1 a and E(V1 ) > a Z Exercise 3.2.3. (i) ∞ e− y2 2 x The first equality gives us ∞ Z ∞ 1 y2 1 − y2 dy = − e− 2 e 2 dy − y y2 x x ∞ ∞ Z ∞ 1 − y2 3 − y2 1 − y2 2 2 2 e = − e dy . − − 3e − y y y4 x x x Z ∞ e− y2 2 dy ≤ x 1 − x2 e 2. x The second equality gives us Z ∞ 2 y2 1 1 − x2 e− 2 dy. ≤ − 3 e x x x Thus, P (Xi > x) ∼ 1 x · 1 2 √1 e− 2 x . 2π P (Xi > x + θ/x) ∼ P (Xi > x) Then, √ 2 1 exp − 21 x + xθ θ 2π(x+ x ) √1 exp − 12 x2 2πx → e−θ (ii) It’s clear that bn → ∞ as n → ∞. By (i), P Xi > bn + bxn · n → e−x . x P (bn (Mn − bn ) ≤ x) = P Mn ≤ bn + bn n x = P Xi ≤ bn + bn n x = 1 − P Xi > bn + bn n P Xi > bn + bxn = 1 − P (Xi > bn ) · P (Xi > bn ) n x P X > b + i n bn 1 = 1 − · n P (Xi > bn ) → exp −e−x . √ (iii) Claim. bn ∼ 2 log n. √ 2 log n) ∼ n√4π1 log n . Thus, for large n, bn ≤ √ (≥) Similar argument with 2 log n − 2 log log n Proof. (≤) By (i), P (Xi > √ 2 log n. Now, by (iii), P (bn (Mn − bn ) ≤ x) → exp (−e−x ). Let xn and yn be xn → ∞ and yn → −∞. Then, xn P Mn − bn ≤ →1 bn 38 yn M n − bn ≤ →0 bn yn xn ⇒P ≤ M n − bn ≤ →1 bn bn Mn bn xn y √ n ≤√ −√ ≤ √ → 1. ⇒P bn 2 log n 2 log n 2 log n bn 2 log n By choosing xn , yn with o b2n , we can show that P √ Mn P → 1. 2 log n Exercise 3.2.4. By Skorohod’s Representation Theorem(Thm 3.2.8 in Textbook), there d d a.s. exists Yn and Y such that Yn = Xn and Y = X with Yn −−→ Y . By Continuous Mapping a.s. Theorem (Exercise 1.3.3), Eg(Yn ) −−→ Eg(Y ). By Fatou’s Lemma, we get lim inf Eg(Xn ) = lim inf Eg(Yn ) ≥ Eg(Y ) = Eg(X). n n Exercise 3.2.5. By Skorohod’s Representation Theorem, there exists Yn and Y with distria.s. bution Fn and F , respectively, such that Yn −−→ Y . By Theorem 1.6.8 in Textbook, Z h(x)dFn (x) →∈ h(x)dF (x). Exercise 3.2.6. Define L(F, G) = {c : F (x − c) − c ≤ G(x) ≤ F (x + c) + c for all x}. Since any distribution is non-decreasing, (a) if a ∈ L(F, G), then for any b ≥ a are also in L(F, G) (b) L(F, G) contains only non-negative constants. (1) ρ is a metric. (i) ρ(F, G) = 0 ⇐⇒ F = G. (⇐) Trivial. (⇒) For any > 0, F (x − ) − ≤ G(x) ≤ F (x + ) + for all x. Letting & 0, we get F (x− ) ≤ G(x) ≤ F (x) for all x. Thus, for x ∈ C(F ), F (x) = G(x). For x ∈ D(F ), since D(F ) is at most countable(Exercise 1.2.3), we can find (xn )n∈N such that xn ∈ C(F ) with xn ↓ x. Since F (xn ) = G(xn ), F (x) = limn F (xn ) = limn G(xn ) = G(x). Therefore, F = G (ii) ρ(F, G) = ρ(G, F ). Suppose a ∈ L(F, G). Then, for all x, F (x) − a ≤ G(x + a) ≤ F (x + 2a) + a F (x − 2a) − a ≤ G(x − a) ≤ F (x) + a. Thus, G(x − a) − a ≤ F (x) ≤ G(x + a) + a for all x. That is, a ∈ L(G, F ). Therefore, ρ(G, F ) ≤ ρ(F, G). By symmetry, it also holds that ρ(F, G) ≤ ρ(G, F ). (iii) ρ(F, H) ≤ ρ(F, G) + ρ(G, H). Suppose a ∈ L(F, G) and b ∈ L(G, H). Then, for all x, F (x − a − b) − a − b ≤ G(x − b) − b ≤ H(x) ≤ G(x + b) + b ≤ F (x + a + b) + a + b. 39 That is, a + b ∈ L(F, H). Thus, ρ(F, H) ≤ ρ(F, G) + ρ(G, H). By (i)-(iii), ρ is a metric. (2) ρ(Fn , F ) → 0 if and only if Fn ⇒ F . (⇒) Since ρ(Fn , F ) → 0, for given > 0, there exists N ∈ N such that ρ(Fn , F ) < for n > N . By (a), ∈ L(Fn , F ) for n > N . That is, if n > N , F (x − ) − ≤ Fn (x) ≤ F (x + ) + , ∀x. Letting n → ∞, F (x − ) − ≤ lim inf Fn (x) ≤ lim sup Fn (x) ≤ F (x + ) + , ∀x. For x ∈ C(F ), letting & 0, F (x) ≤ lim inf Fn (x) ≤ lim sup Fn (x) ≤ F (x). Thus, Fn (x) → F (x). (⇐) Since Fn ⇒ F , limn→∞ Fn (x) = F (x) for x ∈ C(F ). Fix > 0. Since D(F ) is at most countable, we can find {x1 , . . . , xk } ∈ C(F ) such that F (x1 ) < , F (xk ) > 1 − and |xi − xi−1 | < for i = 2, . . . , k . For each xi , there exists Ni () ∈ N such that F (xi ) − < Fn (xi ) < F (xi ) + if n > Ni (). Let N () = max {N1 (), . . . , Nk ()}. Then, for n > N (), (i) if x < x1 , F (x − 2) − 2 ≤ F (x1 ) − 2 < 0 ≤ Fn (x) ≤ Fn (x1 ) < F (x1 ) + < 2 ≤ F (x + 2) + 2. (ii) if x > xk , F (x − 2) − 2 ≤ 1 − 2 < F (xk ) − < Fn (xk ) ≤ Fn (x) ≤ 1 < F (xk ) + 2 ≤ F (x + 2) + 2. (iii) if xi−1 ≤ x ≤ xi+1 for i = 2, . . . , k , F (x − ) − ≤ F (xi−1 ) − < Fn (xi−1 ) ≤ Fn (x) ≤ Fn (xi ) < F (xi ) + ≤ F (x + ) + . By (i)-(iii), 2 ∈ L(F, Fn ) for n > N (). Since > 0 is arbitrarily chosen, we can conclude that ρ(Fn , F ) → 0. Exercise 3.2.7. X ∼ F and Y ∼ G. Recall L(F, G) defined in Exercise 3.2.6. Define K(X, Y ) = { ≥ 0 : P (|X − Y | > ) ≤ }. If one can show that K(X, Y ) ⊂ L(F, G), then it’s done. Suppose a ∈ K(X, Y ). Then, P (|X − Y | > a) ≤ a. Since P (X ≤ x − a) − P (|X − Y | > a) ≤ P (Y ≤ x) ≤ P (X ≤ x + a) + P (|X − Y | > a) ∀x, we get P (X ≤ x − a) − a ≤ P (Y ≤ x) ≤ P (X ≤ x + a) + a, ∀x. Thus, a ∈ L(F, G) 40 Exercise 3.2.8. Recall K(X, Y ) defined in Exercise 3.2.7. It is easy to check that if a ∈ K(X, Y ), then any b > a is also in K(X, Y ). If a = 0, then X = Y , and thus β(X, Y ) = 0. Now, suppose a > 0. (1) a2 1+a ≤ β(X, Y ). Fix 0 < < a. Since ∈ / K(X, Y ), < P (|X − Y | > ). By Exercise 2.3.6 (b) (⇒), we get 1+ < P (|X − Y | > ) ≤ E Thus, 2 1+ |X − Y | 1 + |X − Y | . < β(X, Y ). Letting % a, we get a2 ≤ β(X, Y ). 1 + a2 (2) β(X, Y ) ≤ 2a 1+a . For all b > a, |X − Y | dP 1 + |X − Y | Z Z |X − Y | |X − Y | = dP + dP |X−Y |>b 1 + |X − Y | |X−Y |≤b 1 + |X − Y | Z Z b ≤ 1dP + dP 1 + b |X−Y |>b |X−Y |≤b Z β(X, Y ) = b = P (|X − Y | > b) + · P (|X − Y | ≤ b) 1+b b b + = P (|X − Y | > b) · 1 − 1+b 1+b b b 2b ≤ + = . 1+b 1+b 1+b The last inequality holds since b ∈ K(X, Y ). Letting b & a, we get β(X, Y ) ≤ 2a 1+a . Exercise 3.2.9. Since Fn ⇒ F and F is continuous, Fn (x) → F (x), ∀x. Fix k ∈ N. Pick xj,k = F −1 kj for j = 1, . . . , k − 1 with x0,k = −∞ and xk,k = ∞. For each j, there exists Nk (j) ∈ N such that |Fn (xj,k ) − F (xj,k )| < 1 k if n > Nk (j). Let Nk = max {Nk (1), . . . , Nk (k − 1)}. Then, for n > Nk , if xj−1,k < x < xj,k , then F (x) − 2 1 1 2 ≤ F (xj−1,k) − < Fn (xj−1,k ) ≤ Fn (x) ≤ Fn (xj,k ) < F (xj,k ) + ≤ F (x) + . k k k k That is, |Fn (x) − F (x)| < k2 . Thus, supx |Fn (x) − F (x)| → 0 as n → ∞. Pn iid Exercise 3.2.10. Let Xi ∼ F and Fn (ω, x) = n1 i=1 I(Xi (ω) ≤ x). By Glivenko-Cantelli Theorem(Thm 2.4.9), for each x ∈ R, Fn (ω, x) → F (x) a.s. Let Ωx be a set where Fn (ω, x) → F (x) and let Ω0 = ∩q∈Q Ωq . Then, P (Ω0 ) = 1 and for each ω0 ∈ Ω0 , we have Fn (ω, q) → F (q), ∀q ∈ Q. Now, for x ∈ C(F ), choose r, s ∈ Q s.t. r < x < s and |F (r) − F (x)| < 41 |F (s) − F (x)| < . Then, Fn (x) − F (x) ≤ Fn (s) − F (r) → F (s) − F (r) < 2 Fn (x) − F (x) ≥ Fn (r) − F (s) → F (r) − F (s) > −2. Therefore, −2 ≤ lim inf(Fn (x) − F (x)) ≤ lim sup(Fn (x) − F (x)) ≤ 2. Exercise 3.2.11. Since Xn , 1 ≤ n ≤ ∞, are integer valued, R\Z ⊂ C(Fn ), where Fn is df of Xn . (⇒) For 1 ≤ n < ∞, (i) m ∈ R\Z P (Xn = m) = 0 = P (X∞ = m). (ii) m ∈ Z 1 1 P (Xn = m) = Fn m + − Fn m − 2 2 1 1 − F∞ m − → F∞ m + 2 2 = P (X∞ = m). By (i), (ii), P (Xn = m) → P (X∞ = m), ∀m. (⇐) Want to show that Fn (x) → F∞ (x), ∀x ∈ C(F∞ ). Fix > 0. Since F∞ can have positive probability only on Z, which is countable, we can pick K() := {k1 , · · · , kl } ⊂ Z such that l X (11a) P (X∞ = ki ) > 1 − . i=1 By assumption, for each ki , there exists N (i) ∈ N such that for n > N (i), |P (Xn = ki ) − P (X∞ = ki )| < . l Let N = max {N (1), · · · , N (l )}. Then, for n < N , l l X X [P (Xn = ki ) − P (X∞ = ki )] ≤ |P (Xn = ki ) − P (X∞ = ki )| =: P0 < i=1 i=1 Also, by (11a), we have P1 := X P (X∞ = k) < . k∈Z\K() Thus, P2 : = X P (Xn = k) k∈Z\K() 42 = 1− l X P (Xn = ki ) i=1 = 1− l X P (X∞ = ki ) + i=1 X ≤ l X P (X∞ = ki ) − i=1 P (X∞ = k) + l X P (Xn = ki ) i=1 l X P (Xn = ki ) − P (X∞ = ki ) i=1 k∈Z\K() < 2. Now, for x ∈ C(F∞ ), let Zx = Z ∩ (−∞, x]. X P (Xn = k) Fn (x) = k∈Zx X = X P (Xn = k) + k∈Zx ∩K() P (Xn = k). k∈Zx \K() Similarly, X F∞ (x) = X P (X∞ = k) + k∈Zx ∩K() P (X∞ = k). k∈Zx \K() For n > N , X |Fn (x) − F∞ (x)| = k∈Zx ∩K() X ≤ X P (Xn = k) − P (X∞ = k) + X P (Xn = k) − P (X∞ = k) k∈Zx \K() k∈Zx ∩K() ≤ X P (Xn = k) − P (X∞ = k) + P (Xn = k) + k∈Zx \K() X |P (Xn = k) − P (X∞ = k)| + k∈Zx ∩K() P (Xn = k) + k∈Zx \K() Z ∩ K() ⊂ K() = {k , · · · , k } x 1 l Note that Z \K() ⊂ Z\K() . Thus, x |Fn (x) − F∞ (x)| ≤ P0 + P1 + P2 < 4. Therefore, Fn (x) → F∞ (x), and thus Xn ⇒ X∞ . Exercise 3.2.12. (1) If Xn → X in probability, then Xn ⇒ X. p Since Xn − → X, for given > 0, P (|Xn − X| > ) → 0. (i) P (Xn ≤ x) = P (Xn ≤ x, |Xn − X| ≤ ) + P (Xn ≤ x, |Xn − x| > ) ≤ P (X ≤ x + ) + P (|Xn − X| > ). Then, lim supn P (Xn ≤ x) ≤ P (X ≤ x + ). By letting & 0, we get lim sup P (Xn ≤ x) ≤ P (X ≤ x). n 43 X P (X∞ = k) k∈Zx \K() X k∈Zx \K() P (X∞ = k) . (ii) P (Xn > x) = P (Xn > x, |Xn − X| ≤ ) + P (Xn > x, |Xn − X| > ) ≤ P (X > x − ) + P (|Xn − X| > ). Then, lim sup P (Xn > x) ≤ P (X > x − ) n ⇒ lim sup [1 − P (Xn ≤ x)] ≤ 1 − P (X ≤ x − ) n ⇒ lim inf P (Xn ≤ x) ≥ P (X ≤ x − ). n For x ∈ C(FX ), by letting & 0, we get lim inf P (Xn ≤ x) ≥ P (X ≤ x). n By (i), (ii), for x ∈ C(FX ), Fn (x) = P (Xn ≤ x) → P (X ≤ x) = F (x). Thus, Xn ⇒ X. 0, (2) Since Xn ⇒ c, P (Xn ≤ y) → 1, y<c . Thus, for any > 0, y>c P (|Xn − c| > ) = P (Xn > c + ) + P (Xn < c − ) = 1 − P (Xn ≤ c + ) + P (Xn < c − ) → 0. X ⇒ X, P (X ≤ x) → P (X ≤ x), ∀x ∈ C(F ). n n X Exercise 3.2.13. Since P Y → c, P (|Yn − c| > ) → 0, ∀ > 0. n (i) P (Xn + Yn ≤ z) = P (Xn + Yn ≤ z, |Yn − c| ≤ ) + P (Xn + Yn ≤ z, |Yn − c| > ) ≥ P (Xn ≤ z − c + ) + P (|Yn − c| > ). Since D(FX ) is at most countable, we may assume that z − c + ∈ C(FX ). Then, lim supn P (Xn + Yn ≤ z) ≤ P (X ≤ z − c + ). Letting & 0, lim sup P (Xn + Yn ≤ z) ≤ P (X ≤ z − c) = P (X + c ≤ z). n (ii) P (Xn + Yn > z) = P (Xn + Yn > z, |Yn − c| ≤ ) + P (Xn + Yn > z, |Yn − c| > ) ≤ P (Xn > z − c − ) + P (|Yn − c| > ). By similar argument, we get lim inf P (Xn + Yn ≤ z) ≥ P (X ≤ z − c − ). n For z − c ∈ C(FX ), by letting → 0, lim inf P (Xn + Yn ≤ z) ≥ P (X ≤ z − c) = P (X + c ≤ z). n 44 Note that z ∈ C(FX+c ) if and only if z − c ∈ C(FX ). By (i), (ii), we get P (Xn + Yn ≤ z) → P (X + c ≤ z). X ⇒ X, P (X ≤ x) → P (X ≤ x), ∀x ∈ C(F ). n n X Exercise 3.2.14. Since P Y → c, P (|Yn − c| > ) → 0, ∀ > 0. n For simple proof, let’s assume 0 < < c. (i) P (Xn Yn ≤ z) = P (Xn Yn ≤ z, |Yn − c| ≤ ) + P (Xn Yn ≤ z, |Yn − c| > ) z ≤ P Xn ≤ + P (|Yn − c| > ). c− z Let c− ∈ C(FX ). Then, z lim sup P (Xn Yn ≤ z) ≤ P X ≤ . c− n Letting → 0, z lim sup P (Xn Yn ≤ z) ≤ P X ≤ = P (cX ≤ z). c n (ii) P (Xn Yn > z) = P (Xn Yn > z, |Yn − c| ≤ ) + P (Xn Yn > z, |Yn − c| > ) Similarly, z ≤ P Xn > + P (|Yn − c| > ). c+ z . lim inf P (Xn Yn ≤ z) ≥ P X ≤ n c+ For z c ∈ C(FX ), by letting → 0, z lim inf P (Xn Yn ≤ z) ≥ P X ≤ = P (cX ≤ z). n c Note that z ∈ C(FcX ) if and only if z c ∈ C(FX ). By (i),(ii), we get P (Xn Yn ≤ z) → P (cX ≤ z). Pn d Exercise 3.2.15. Step 1. Xni ≡ Yi /( m=1 Ym2 /n)1/2 . Note that uniformly distributed over the surface of the unit sphere is the only distribution which is invariant to rotation and belongs to unit sphere. iid Since Yi ∼ N (0, 1), Y = (Y1 , · · · , Yn ) ∼ N (0, In ). For any orthogonal matrix A, AY ∼ Pn Pn d d AY Y 2 2 N (0, In ) ≡ Y . Then, ||YY||2 ≡ ||AY m=1 Ym , · · · , Ym / m=1 Ym ) is ||2 . Thus, ||Y ||2 = (Y1 / √ d uniformly distributed over the surface of the unit sphere. Then, Xn ≡ n ||YY||2 , and thus d Xni ≡ Yi n X m=1 45 !1/2 Ym2 /n . Step 2. Yi /( Pn m=1 Ym2 /n)1/2 ⇒ N (0, 1). For Yi ∼ N (0, 1), EYi2 = 1 < ∞. Thus, by WLLN, n X P Ym2 n → 1. m=1 Pn By Exercise 3.2.12(1), m=1 Ym2 /n ⇒ 1. Note that f (x) = √1x I(x>0) is continuous on R\{0} Pn and P ( m=1 Ym2 /n = 0) = P (Y1 = · · · = Yn = 0) = 0. By Continuous Mapping Theorem (Theorem 3.2.10), we get n X Ym2 !−1/2 ⇒ 1. n m=1 By Exercise 3.2.14, Yi · n X Ym2 !−1/2 ⇒ N (0, 1). n m=1 Thus, Xn1 ⇒ standard normal. Exercise 3.2.16. For 1 < M < ∞, since EYnβ → 1, there exists N ∈ N such that for n > N , EYnβ ≤ M . Therefore, we can drop finitely many Yn ’s and say {Fn } is tight. (with ϕ(y) = |y|β ) For every subsequence (Fnk ), by Helly and Tightness Theorem, there exists (Fn(mk ) ) such that Fn(mk ) ⇒ F , where F is a df. Let Y be a r.v. with F. For g(y) = |y|β and h(y) = |y|α , we have (0) g, h are continuous. (i) g(y) > 0 (ii) |h(y)|/g(y) = |y|α−β → 0 as |y| → ∞ (iii) Fn(mk ) ⇒ F R β (iv) |y|β dFn(mk ) (y) = EYn(m ≤ M < ∞. k) By Exercise 3.2.5, Z α |y|α dFn(mk ) (y) = EYn(m → EY α = k) Z |y|α dF (y). Since convergent real sequence has the unique limit, EY α = 1. Also, similar argument for h(y) = |y|γ with γ ∈ (α, β) gives us that γ EYn(m → EY γ . k) From Jensen’s Inequality with following functions, we get EYnγ → 1. γ (1) ϕ1 (y) = |y| α . γ γ (EYnα ) α ≤ E (Ynα ) α = EYnγ 46 ⇒ 1 ≤ lim inf EYnγ . n β (2) ϕ2 (y) = |y| γ . β β (EYnγ ) γ ≤ E (Ynγ ) γ = EYnγ ⇒ lim sup EYnγ ≤ 1. n Therefore, EY γ = 1. γ γ γ For strictly convex ϕ(y) = |y| α , we have (EY α ) α = 1 = E(Y α ) α = EY γ . By Exercise 1.6.1, P Y α = 1 a.s. Since α > 0, Y = 1 a.s. Since Yn(mk ) → 1, by Theorem 2.3.2, P Yn → 1. Exercise 3.2.17. Suppose not. Then, for some K < ∞ and y < 1, there exists a r.v. X with EX 2 = 1 and EX 4 ≤ K such that P (|X| > y) < c for all c > 0. With c = we can construct a sequence (Xn ) such that P (|Xn | > y) → 0, EXn2 1 n, n ∈ N, = 1 and EXn4 ≤ K. Theorem 3.2.14 with ϕ(x) = x2 gives us that {Xn } is tight. Then, by Helly and Tightness Theorem, there exists (Xnm ) such that Xnm ⇒ X∞ . By Generalized DCT with h(x) = x2 2 and g(x) = x4 , we get EX∞ = limm→∞ EXn2m = 1. However, since P (|Xn | > y) → 0, 2 P (|Xn | ≤ y) → 1. Then, EX∞ ≤ y 2 < 1, which is a contradiction. 3.3. Characteristic Functions. Exercise 3.3.1. Let X be a r.v. with ch.f ϕ and X 0 be an independent copy of −X. Then, its ch.f ϕ0 (t) = Eeit(−X) = ϕ(t). By Theorem 3.3.2 and Lemma 3.3.9 in Textbook, Re ϕ is ch.f of 12 (X + X 0 ) and |ϕ|2 is ch.f of X + X 0 . Exercise 3.3.2. (i) Let T 1 2T Z 1 = 2T Z Z −T Z T −∞ Z ∞ −T −∞ IT = = 1 2T −T T e−ita ϕ(t) dt ∞ e−ita eitx µ(dx) dt eit(x−a) µ(dx) dt Since eit(x−a) ≤ 1 and [−T, T ] is a finite interval, by Fubini, Z ∞Z T 1 cos(t(x − a)) + i sin(t(x − a)) dt µ(dx). IT = 2T −∞ −T Since sin t is symmetric about the origin, Z ∞Z T 1 IT = cos(t(x − a)) dt µ(dx) 2T −∞ −T Rx RT 1 Since 0 cos(y)dy = | sin(y)| ≤ |x|, 2T cos(t(x − a)) dt ≤ 1 : integrable. By −T DCT, Z ∞ lim IT = T →∞ −∞ " 1 lim T →∞ 2T Z T # cos(t(x − a)) dt µ(dx). −T 47 (1) x 6= a 1 lim T →∞ 2T Z T sin((x − a)T ) = 0. (x − a)T cos(t(x − a)) dt = lim T →∞ −T (2) x = a 1 T →∞ 2T Z T cos(t(x − a)) dt = 1. lim −T By (1) and (2), we get lim IT = µ {a} . T →∞ (ii) Let µ be the distribution of X. ϕ 2π +t h Z = 2π h +t)x ei( Z 2π h +t)x ei( = µ(dx) µ(dx) hZ Z eitx µ(dx) = hZ Z (2a) = eitx µ(dx) = ϕ(t). By (i), 1 P (X = x) = µ({x}) = lim T →∞ 2T Let Tn = π h + 2π h Z T e−itx ϕ(t)dt. −T · n, with n ≥ 0 . Then, since limit of RHS above exists, Z Tn 1 e−itx ϕ(t)dt. P (X = x) = lim n→∞ 2Tn −T n Now, for 1 ≤ k ∈ N, Z Tk e−itx ϕ(t)dt = Z π h e−i(t+ 2kπ h )x −π h Tk−1 = e−i( 2kπx h ) π h Z 2kπ ·ϕ t+ dt h e−it ϕ(t)dt −π h by (2a). Since x ∈ hZ, 2kπx ∈ N. Thus, e−i( h Z Tk Z e−itx ϕ(t)dt = 2kπx h ) π h = 1. Therefore, e−it ϕ(t)dt −π h Tk−1 Z T0 = −T0 48 e−itx ϕ(t)dt. R −T RT This also holds for −Tkk−1 e−itx ϕ(t)dt = −T0 0 e−itx ϕ(t)dt. Then, Z Tn Z T0 1 1 e−itx ϕ(t)dt = · (2n + 1) e−itx ϕ(t)dt 2Tn −Tn 2 · (2n+1) π −T 0 h Z T0 h e−itx ϕ(t)dt = 2π −T0 Z πh h = e−itx ϕ(t)dt. 2π − πh Therefore, h 2π P (X = x) = Z h π e−itx ϕ(t)dt. h −π (iii) X = Y + b. If P (X ∈ b + hZ) = 1, then P (Y ∈ hZ) = 1. For x ∈ b + hZ, P (X = x) = P (Y = x − b). Then, h 2π Z h = 2π Z h = 2π Z P (Y = x − b) = π h e−it(x−b) ϕY (t)dt −π h π h −π h π h e−itx · eitb · e−itb ϕX (t)dt e−itx ϕX (t)dt. −π h Exercise 3.3.3. Since X and Y are independent, X and −Y are also independent. Then, ch.f of X − Y is given by ϕ(t) · ϕ(t) = |ϕ(t)|2 . By Exercise 3.3.2, Z T 1 P (X − Y = 0) = lim |ϕ(t)|2 dt. T →∞ 2T −T By Exercise 2.1.5, P (X − Y = 0) = X µ {x} 2 . x Exercise 3.3.4. For exponential distribution, f (y) = e−y I(y>0) ϕ(t) = 1 1 − it And, Z Z |ϕ(t)|dt = 1 dt = 1 − it Z √ 1 dt = ∞. 1 + t2 Exercise 3.3.5. ch.f of X1 is given by ϕ1 (t) = EeitX 1 Z 1 1 itx itx 1 e = e · dx = 2 2it −1 −1 1 = eit − e−it 2it 49 1 (cos t + i sin t − cos t + i sin t) 2it sin t = . t n Thus, ch.f of X1 + · · · + Xn is given by ϕ(t) = sint t . = Note that Z ∞ Z sin t t |ϕ(t)|dt = 2 0 n dt. For n ≥ 2, Z ∞ 0 sin t t n Z 1 dt = 0 Z ≤ sin t t 1 n Z dt + 1 ∞ Z 1dt + 0 1 ∞ sin t t n dt 1 dt < ∞. tn Therefore, by Thm 3.3.14, n sin t dt t −∞ n n Z ∞ sin t sin t 1 = · cos tx − i · sin tx dt 2π −∞ t t n Z 1 ∞ sin t cos txdt. = π 0 t f (x) = 1 2π Z ∞ e−itx · Exercise 3.3.6. In Example 3.3.16, we showed that ch.f of Cauchy distribution is given by exp(−|t|). Since X1 , · · · , Xn are independent, by Thm 3.3.2, ϕX1 +···+Xn (t) = n Y ϕXi (t) = i=1 n Y exp(−|t|) = exp(−n|t|) i=1 Bu Thm 3.3.1(e), t ϕ X1 +···+Xn (t) = ϕX1 +···+Xn = exp(−|t|), n n which is ch.f of Cauchy. Thus, by Inversion Formula, X1 +···+Xn d ≡ n X1 . 2 2 σ t Exercise 3.3.7. ϕXn (t) = exp − n2 . 2 σ Since Xn ⇒ X, ϕXn (t) → ϕ∞ (t), ∀t. For t = 1, ϕXn (1) = exp − 2n . Thus, ϕXn (1) converges ⇐⇒ log ϕXn (1) converges ⇐⇒ − σn2 converges. 2 Therefore, there exists σ 2 such that σn2 → σ 2. Sinceσn2 ∈ (0, ∞), we can say that σ 2 ∈ [0, ∞]. σ 2 t2 Now, suppose σ 2 = ∞. Then ϕXn (t) = exp − n2 → 0 if t 6= 0. However, ϕXn (0) = 1, ∀n. Thus, continuity of ϕX at 0 is violated, which means that Xn ; X. Thus, σn2 → σ 2 ∈ [0, ∞). 50 Exercise 3.3.8. Let ϕn (t) be the ch.f of Xn and ψn (t) be the ch.f of Yn , for 1 ≤ n ≤ ∞. By Theorem 3.3.17(Continuity Theorem), ϕn (t) → ϕ∞ (t), ∀t ψn (t) → ψ∞ (t), ∀t. Since Xn and Yn are independent, ch.f of Xn + Yn = ϕn (t)ψn (t) , for 1 ≤ n ≤ ∞. Then, ϕn (t)ψn (t) → ϕ∞ (t)ψ∞ (t), ∀t. Since ϕ∞ , ψ∞ are ch.f, they are continuous at 0. Thus, ϕ∞ ψ∞ is also continuous at 0. By Thm 3.3.17, Xn + Yn ⇒ X∞ + Y∞ . Exercise 3.3.9. By Thm 3.3.2, ch.f of Sn is given by ϕSn (t) = n Y ϕj (t). j=1 Since a.s. convergence implies weak convergence, Sn ⇒ S∞ . By Thm 3.3.17(i), ϕS∞ (t) = lim ϕSn (t) = n→∞ ∞ Y ϕj (t), ∀t. j=1 Exercise 3.3.10. By Exercise 3.3.5, sint t is ch.f of U nif (−1, 1). Let Y1 , . . . , Yn be i.i.d sym1 metric Bernoulli random variables. P (Y1 = 1) = P (Y1 = −1) = 2 . Define Xi = Y2ii . Then, P∞ Pn Q∞ ch.f of Xi = cos 2ti , and thus ch.f of i=1 Xi = i=1 cos 2ti . Claim. i=1 Xi ⇒ U nif (−1, 1). P n X 2k + 1 Xi = 2n i=1 ! = 1 for k = −2n−1 , · · · , −1, 0, 1, · · · , 2n−1 − 1. 2n Then, P n X ! Xi ≤ x b2n xc − 1 + 2n−1 , 0 2 n 1 1 b2 xc − 1 = max + n ,0 2 2 2 = max i=1 1 2n Now, for x ∈ (−1, 1), letting n → ∞ gives us P ∞ X ! Xi ≤ x = i=1 1 1 + x. 2 2 Note that for x ∈ (−1, 1), P (U nif (−1, 1) ≤ x) = 21 + 21 x. Therefore, Pn i=1 Xi ⇒ U nif (−1, 1). By Theorem 3.3.17, we get the desired result. Exercise 3.3.11. ch.f of 2Xj 3j = Eeit 2Xj 3j = by ϕX (t) = 1 2 + 1 2 ∞ Y 1 j=1 51 2it · e 3j . By Exercise 3.3.9, ch.f of X is given 1 2it + e 3j . 2 2 Then, k ∞ Y 2i3 π 1 1 ϕX 3k π = + exp 2 2 3j j=1 = ∞ Y 1 k Y 1 1 1 −l k−j + exp 2π · 3 i · + exp 2π · 3 i 2 2 2 2 j=1 l=1 Since 3 k−j ∈ N ∪ {0}, we get k ϕX 3 π = ∞ Y 1 1 −l + exp 2π3 i . 2 2 j=1 R∞ R∞ 1 2 Exercise 3.3.12. We have −∞ |x|e dx = 2 0 xe− 2 x dx < ∞. By induction, for n ≥ 2, Z ∞ Z ∞ 1 2 n − 21 x2 xn e− 2 x dx dx = 2 |x| e − 12 x2 −∞ 0 Z ∞ h i 1 2 ∞ 1 2 +2 xn−1 e− 2 x dx = 2 −xn−1 e− 2 x 0 0 Z ∞ 1 2 =0+2 xn−1 e− 2 x dx < ∞. 0 Note that 1 2 ϕ(n) (t) = exp− 2 t 2 3 1 1 2 1 1 2 1 2 + · − t + ··· =1− t + · − t 2 2! 2 3! 2 ∞ X 1 =1+ t2k . k! · (−2)k k=1 By Thm 3.3.18, we have in EX n = ϕ(n) (t)|t=0 . Therefore, EX 2n = Exercise 3.3.13. (i) Fix > 0. For 0 < h < (2n)! . 2n n! 2 4 , |ϕi (t + h) − ϕi (t)| ≤ E|eihXi − 1| ≤ E min(|hXi |, 2) 2 2 ≤ E |hXi |; |Xi | ≤ √ + E 2; |Xi | > √ h h √ 2 ≤ 2 h + 2P |Xi | > √ h √ ≤2 h+ ≤ 2. The first inequality comes from Theorem 3.3.1(d). The second inequality comes from (3.3.3). The fifth inequality comes from tightness. 52 (ii) Since µn ⇒ µ∞ , {µn }n∈N is tight, and thus {µn }1≤n≤∞ is tight. Then, by (i), for > 0, choose δ > 0 so that if |h| < δ, then |ϕn (t + h) − ϕn (t)| < , for 1 ≤ n ≤ ∞. Without loss of generality, assume the compact set is [−K, K]. Choose integer m > 1 δ and let xj = j m · K for −m ≤ j ≤ m. By Continuity Theorem, we have ϕn (t) → ϕ∞ (t) for all t. Thus, for xj , there exists nj ∈ N such that if n > nj , then |ϕn (xj ) − ϕ∞ (xj )| < . Let n0 = max {n−m , · · · , nm }. Then, for all t ∈ [−K, K], there exists j such that t − jK m < δ. Now, for n > n0 , jK jK jK jK + ϕn + ϕ∞ |ϕn (t) − ϕ∞ (t)| ≤ ϕn (t) − ϕn − ϕ∞ − ϕ∞ (t) m m m m < 3. Therefore, ϕn → ϕ∞ uniformly on a compact set. (iii) Let Xn = it 1 n. Then, ϕn (t) = e n Sn = ϕ n (i) ch.f of a = eiat Exercise 3.3.14. Claim. ϕ t n t n n . → eiat , for all t. Proof. Since ϕ0 (0) = ia, ϕ( nt ) − ϕ(0) t = ia ⇒ lim n ϕ − 1 = iat. n→∞ n→∞ t/n n lim Thus, for all t, !n n n ϕ( nt ) − 1 t lim ϕ = lim 1 + n→∞ n→∞ n n = eiat . The second equality holds by Theorem 3.4.2. By Claim and Theorem 3.3.17, Sn a ⇒ a, and thus Sn P a → a. (ii) In probability convergence implies weak convergence. By Theorem 3.3.17, we get the desired result. (iii) (I think (ϕ(h) − 1)/h → +ia.) (a) Since log z is differentiable at z = 1, log ϕ( nt ) − log 1 log ϕ( nt ) = lim = (log z)0 n→∞ n→∞ ϕ( t ) − 1 ϕ( nt ) − 1 n t n iat (b) By (ii), ϕ n → e . By taking logarithm, we get t n log ϕ → iat. n lim 53 z=1 = 1. By (a), (b), we get ϕ( nt ) − 1 t t n ϕ −1 = · n log ϕ n n log ϕ( nt ) → iat. By Exercise 3.3.13, Sn n ⇒ a implies that ϕ( nt )n → eiat uniformly on a compact set. Fix small 0 < δ < 1. For any h > 0, we can express h = t n with large n and t ∈ [1 − δ, 1 + δ]. Then, ϕ( nt ) − 1 ϕ(h) − 1 − ia = − ia h t/n = 1 (n(ϕ(t/n) − 1) − iat) t ≤ 1 · |n(ϕ(t0 /n) − 1) − iat0 | t0 sup t0 ∈[1−δ,1+δ] ≤ 1 1−δ |n(ϕ(t0 /n) − 1) − iat0 |. sup t0 ∈[1−δ,1+δ] Since n(ϕ(t/n) − 1) → iat uniformly on compact set, " # (((( ( ( 1 ϕ(h) − 1 ( (( lim sup sup (( |n(ϕ(t/n) − 1) − iat| = 0. lim sup − ia ≤ ((( h 1 − δ n→∞ (t∈[1−δ,1+δ] h→0 ( ( ( Therefore, ϕ(h)−1 h → ia as h ↓ 0. Exercise 3.3.15. From Example 3.3.15, R∞ −∞ 1−cos x πx2 dx = 1. Let x = |y|t. Then, Z ∞ −∞ Thus, 2 R∞ 0 1−cos yt dt πt2 1 − cos |y|t 1 |y|dt = π|y|2 t2 |y| Z ∞ −∞ 1 − cos yt dt = 1. πt2 = |y|. Now, integrate with repect to F (y). Then, we get Z ∞Z ∞ Z ∞ 1 − cos yt 2 dtdF (y) = |y|dF (y). πt2 −∞ −∞ 0 Since 1−cos yt πt2 ≥ 0, by Fubini, Z ∞ Z ∞ ∞Z ∞ 1 − Re(ϕ(t)) 1 − cos yt 2 dF (y)dt = 2 dt = |y|dF (y). πt2 πt2 0 −∞ 0 −∞ Z = c > −∞, Exercise 3.3.16. Since limt↓0 ϕ(t)−1 t2 ϕ(t) − 1 ϕ(t) − 1 ϕ(−t) − 1 lim = lim = lim = c. 2 2 t↓0 t↓0 t↓0 t t t2 Thus, lim t↓0 ϕ(t) + ϕ(−t) − 2ϕ(0) = 2c > −∞. t2 54 By Theorem 3.3.21, EX 2 < ∞. By Theorem 3.3.20, ϕ(t) = 1 + itEX − t2 2 2 EX + o(t2 ). Then, 2 lim t↓0 itEX − t2 EX 2 + o(t2 ) ϕ(t) − 1 = lim t↓0 t2 t2 iEX 1 = lim − EX 2 t↓0 t 2 = c > −∞. Thus, EX = 0 and EX 2 = −2c < ∞. ϕ(t)−1 t2 Especially, if ϕ(t) = 1 + o(t2 ), we get limt↓0 = limt↓0 o(t2 ) t2 = 0. Then, EX = 0 and EX 2 = 0. Thus, Var(X) = 0, which means that X is constant a.s. Therefore, X = 0 a.s. Thus, ϕ(t) = EeitX = 1. Exercise 3.3.17. (⇒) By Continuity Theorem, ϕn (t) → ϕ0 (t), for all t, where ϕ0 is ch.f of 0. Since ϕ0 (t) = Eeit0 = 1, ϕn (t) → 1, for all t. (⇐) Step 1. {Yn } is tight. |f (t) ≤ 2 I n α [−α,α] (t) : integrable Let fn (t) = α1 (1−ϕn (t))I[−α,α] (t) for α < δ. Then, f (t) → 0 by assumption n By DCT, we have Z fn (t) = 1 α Z α (1 − ϕn (t))dt → 0. −α Note that Z Z Z ∞ 1 u 1 u ity (1 − ϕn (t))dt = 1− e dµn (y) dt u −u u −u −∞ Z Z 1 ∞ u = 1 − eity dtdµn (y) u −∞ −u u Z 1 ∞ 1 = t − eity dµn (y) u −∞ iy −u " Z 1 ∞ 1 = u − (cos uy + i sin uy) u −∞ iy # 1 − −u − (cos uy − i sin uy) dµn (y) iy Z ∞ 1 2 = 2u − sin uy dµn (y) u −∞ y Z ∞ sin uy =2 1− dµn (y) uy −∞ Z 1 ≥2 1− dµn (y) 2 u|y| |y|≥ u 55 . Z ≥ 1dµn (y) 2 |y|≥ u 2 = µn y : |y| ≥ u 2 . ≥ µn y : |y| > u The second equality holds by Fubini. The first inequality holds since sin uy uy ≤ sin uy uy ≤ 1 and 1 u|y| . Step 2. Yn ⇒ 0. Fix a subsequence {Ynk }. By Helly and Tightness Theorem, there exists sub-subsequence Ynm(k) such that Ynm(k) ⇒ X. Since ϕnm(k) → 1 for |t| < δ, ϕX (t) = 1 for |t| < δ. Then, for |t| < δ, ϕX (t)−1 t2 = 0, and thus lim t↓0 ϕ(t) − 1 = 0. t2 By Exercise 3.3.16, ϕX (t) ≡ 1. This means that X ≡ 0. Since this holds for all subsequence, by Theorem 3.2.15, Yn ⇒ 0. P Exercise 3.3.18. Step 1. Sm − Sn → 0. Let ϕ(t) be ch.f of X1 . Then, ch.f of Sn is given by ϕ(t)n and it converges to for some ch.f ψ(t). Since ψ is ch.f, there exists δ > 0 such that ψ(t) > 0 for |t| ≤ δ. Now, suppose m > n. Pm ch.f of Sm − Sn = l=n+1 Xl is given by ϕ(t)m−n = ψ(t)m . ψ(t)n For |t| ≤ δ, as m, n → ∞, ϕ(t)m−n → 1. By Exercise 3.3.17, Sm − Sn ⇒ 0. P By Exercise 3.2.12, Sm − Sn → 0. Step 2. Sn converges in prob. By Exercise 2.3.6, d(Sn − Sm , 0) → 0 ⇐⇒ d(Sn , Sm ) → 0. P By Exercise 2.3.7, there exists a r.v. S∞ such that Sn → S∞ . Exercise 19-25 (?) Exercise 3.3.19. Exercise 3.3.20. Exercise 3.3.21. Exercise 3.3.22. Exercise 3.3.23. 56 Exercise 3.3.24. Exercise 3.3.25. 57 3.4. Central Limit Theorems. Exercise 3.4.1. Φ(−1). Exercise 3.4.2. (a) Let σ 2 = Var(Xi ) ∈ (0, ∞). By CLT, we have √Sn nσ ⇒ Z ∼ N (0, 1). Then, for any M , Sn Sn M lim sup P √ > M = lim sup P √ > = P (Z > M/σ) > 0. σ n nσ n n Therefore, S S √n > M i.o. ≥ lim sup P √n > M > 0. n n n o n S Note that lim supn √nn > M ∈ T : tail σ-algebra. Sn By Kolmogorov’s 0-1 law, P lim supn √ > M = 1, for all M . n P Sn √ = ∞ a.s. n P Sn √ → U for some n Thus, lim supn (b) Assume that U. Then, as m → ∞, S(m+1)! S P √m! − p → 0. m! (m + 1)! Since Sm! and S(m+1)! − Sm! are independent, we get ! S(m+1)! − Sm! Sm! Sm! P 1< √ < 2, p < −3 = P 1 < √ <2 P m! m! (m + 1)! ! S(m+1)! − Sm! p < −3 (m + 1)! → P (1 < T < 2)P (T < −3) > 0, (2a) where T ∼ N (0, σ 2 ). For the second term, √ 1 Sm! (m+1)! =√ 1 (m+1)! · Sm! P √ → m! 0. Also, P ! S(m+1)! − Sm! Sm! 1< √ < −3 = P < 2, p m! (m + 1)! S(m+1)! Sm! Sm! 1< √ < −3 + p < 2, p m! (m + 1)! (m + 1)! ! S(m+1)! Sm! Sm! < −3 + √ 1< √ < 2, p m! m! (m + 1)! ! S(m+1)! Sm! 1< √ , p < −1 m! (m + 1)! ! S(m+1)! Sm! √ −p >2 . m! (m + 1)! ≤P ≤P ≤P (2b) By combining (2a) and (2b), we get lim inf P m ! S(m+1)! Sm! √ −p > 2 > 0, m! (m + 1)! which contradicts to the assumption that Sm! √ m! 58 S P − √(m+1)! → 0. (m+1)! ! Exercise 3.4.3. Xi0 : independent copy of Xi Y = X − X 0 i i i Ui = Yi 1(|Yi |≤A) V = Y 1 i i (|Yi |>A) We know that (Xi ), (Xi0 ), (Yi ), (Ui ), (Vi ) are i.i.d sequences. Suppose EXi2 = ∞. Pn Sn ⇒ W , Sn0 = i=1 Xi0 also weakly converges Since there exists some r.v. W such that √ n to W . That is, for all t, ϕW (t) = limn→∞ ϕSn /√n (t) = limn→∞ ϕSn0 /√n (t). Then, ch.f of Pn Ym m=1 √ n is given by Sn Sn0 √ √ ϕ Pnm=1 (t) = E exp it − Ym √ n n n Sn (t) · ϕ S 0 (−t) = ϕ√ √n n n → ϕW (t) · ϕW (−t), ∀t. SincePϕW (t) · ϕW (−t) is continuous at 0, by Continuity Theorem, there exists a r.v. T such that n m=1 √ Ym n ⇒ T . By hint, for any K, which is a continuity point of d.f of T , Pn 1 m=1 Ym √ ≥K ≥ . P (Y ≥ K) = lim P n→∞ 5 n This means that limK→∞ P (T ≥ K) ≥ 1 5 since discontinuity points of d.f of T is at most countable, which is a contradiction. One can derive hint from CLT and MCT. Exercise 3.4.4. Let Sn = X1 + · · · + Xn . CLT for i.i.d sequence implies that (4a) Note that S√ n −n nσ √ = By WLLN, we have Sn − ESn Sn − n √ = √ ⇒ χ. nσ nσ √ √ · ( Sn − n). √ S√ n+ n nσ Sn P n → 1. (EX1 = 1, Var(X1 ) < ∞). By Continuous Mapping Theorem, (4b) 1 σ √ S P 2 √ n +1 → . σ n By Converging Together Lemma with (4a) and (4b), √ −1 p √ 1 σ Sn − n S √ √ n +1 Sn − n = · ⇒ χ σ 2 nσ n Exercise 3.4.5. By CLT, (5a) By WLLN, Pn m=1 Xm √ ⇒ χ. nσ Pn m=1 2 Xm n 59 P → σ2 . By Continuous Mapping Theorem, r Pn m=1 (5b) 2 Xm n P → σ2 . By Converging Together Lemma with (5a) and (5b), s Pn Pn nσ 2 m=1 Xm m=1 Xm √ ⇒ χ. = · Pn Pn 2 1/2 2 ) nσ ( m=1 Xm m=1 Xm Exercise 3.4.6. By CLT for i.i.d sequence, we know that quence San √ σ an S√n σ n ⇒ N (0, 1). Thus, its subseP ⇒ N (0, 1). For Yn and Zn defined above, if one can show that Yn − Zn → 0, SNn Sa Sa − SNn Sa = √ n + (Yn − Zn ) ⇒ N (0, 1) √ = √n + n √ σ an σ an σ an σ an P by Converging Together Lemma. Claim. Yn − Zn → 0. Nn Nn P Fix > 0. Since an → 1, P an − 1 > = P |Nn − an | > an → 0. Then, h i lim sup P (|Yn − Zn | > δ) = lim sup P (|Yn − Zn | > δ, |Nn − an | > an ) + P (|Yn − Zn | > δ, |Nn − an | ≤ an ) n n ≤ lim sup P (|Yn − Zn | > δ, |Nn − an | > an ) + lim sup P (|Yn − Zn | > δ, |Nn − an | ≤ an ) n n = lim sup P (|Yn − Zn | > δ, |Nn − an | ≤ an ). n Note that √ P (|Yn − Zn | > δ, |Nn − an | ≤ an ) = P (|SNn − San | > δ · σ an , |Nn − an | ≤ an ). Since |Nn − an | ≤ an , h i √ P |SNn − San | > δσ an , |Nn − an | ≤ an ≤ P √ δσ an max |Sm − S[(1−)an ] | > 2 (1−)an ≤m≤(1+)an 4 ≤ 2 2 Var S[(1+)an ] − S[(1−)an ] δ σ an 4 = 2 2 σ 2 [(1 + )an ] − [(1 − )an ] δ σ an 8 ≤ 2 δ The second inequality holds by Kolmogorov’s Maximal Inequality. Thus, lim supn P (|Yn − Zn > δ) ≤ 8 δ2 . Since was arbitrarily chosen, P (|Yn − Zn | > δ) → 0, P for any δ > 0. Therefore, Yn → Zn . Exercise 3.4.7. 7 Exercise 3.4.8. 8 Exercise 3.4.9. (1) Var(Sn )/n → 2. 2 Var(Xn ) = EXm = 2 − m12 . Then, Var(Sn ) = 2n − we get the desired result. 60 Pn 1 m=1 m2 . Since Pn 1 m=1 m2 converges, √ (2) Sn / n ⇒ χ. Define Ym = 1 , Xm > 0 . Then, Ym ’s are i.i.d sequence with EYm = 0 and , Xm < 0 Var(Ym ) = 1. By CLT for i.i.d, we have −1 n X Ym .√ n ⇒ χ. m=1 P∞ Note that P (Xm 6= Ym ) = P (Xm = ±m) = m12 . Then, m=1 P (Xm 6= Ym ) = P∞ 1 < ∞. By Borel-Cantelli Lemma 1, P (X = 6 Y i.o.) = 0. Thus, P (Xm = m m m=1 m2 Ym eventually) = 1. This means that Pn Pn Ym Sn − m=1 Ym S √n = m=1 √ √ + ⇒χ n n n by converging together Lemma since the second term converges to 0. Pn Exercise 3.4.10. Let s2n = m=1 Var(Xm ) and Ynm = P n t2n := m=1 Var(Ynm ) = 1. For Ynm , Pn 2 2 (i) m=1 EYnm = tn = 1 Xm −EXm . sn Then, EYmn = 0 and (ii) For > 0, |Ynm | > ⇐⇒ |Xm − EXm | > sn . Note that since |Xi | ≤ M , |Xm − EXm | ≤ 2M , ∀m. However, since P∞ n=1 Var(Xn ) = ∞, Sn > 2M for sufficiently large n. This means that {|Ynm | > } does not occur eventually. Thus, lim n→∞ n Z X m=1 |Ynm |2 dP = 0. |Ynm |> Pn By (i) and (ii), Lindeberg Theorem gives us that m=1 Ynm ⇒ χ. Therefore, Pn Pn Sn − ESn m=1 Xm − E ( m=1 Xm ) pPn =p ⇒ χ. Var(Sn ) Var(X ) m m=1 Exercise 3.4.11. By Exercise 3.4.12, it suffices to show that {Xn } satisfies Lyapounov’s condition. lim n→∞ = lim n→∞ αn−(2+δ) 1 n1+δ/2 n X E(|Xm − EXm |2+δ ) m=1 n X E|Xm |2+δ m=1 C = lim δ/2 n→∞ n =0. Exercise 3.4.12. Let Xn,m = Xm −EXm . αn It suffices to show that {Xn,m } satisfies the Lin- deberg condition. Pn 2 (i) m=1 EXn,m = 1. 61 (ii) For given > 0, n X 2 E(Xn,m ; |Xn,m | > ) = m=1 n Z X |Xn,m |> m=1 = 1 αn2 1 ≤ 2 αn 2 Xn,m dP n X Z |Xm − EXm |2 dP |Xm −EXm |>αn m=1 n Z X m=1 |Xm −EXm |>αn " ≤ |Xm − EXm | αn # δ |Xm − EXm |2 dP n Z 1 X 1 · |Xm − EXm |2+δ dP → 0 δ αn2+δ m=1 by assumption. By (i) and (ii), we get the desired result. Exercise 3.4.13. (i) β > 1. P (Xj 6= 0) = P (Xj = ±j) = ∞ X 1 . jβ Thus, P (Xj 6= 0) = j=1 ∞ X 1 < ∞. β j j=1 By Borel-Cantelli Lemma 1, P (Xj = 0eventually) = 1. For ω ∈ Ω0 := {ω : Xj (ω) = 0 eventually}, P∞ j=1 Xj (ω) converges. Thus, Sn → S∞ a.s. , where S∞ is a r.v. such that S∞ (ω) = P∞ j=1 Xj (ω) for ω ∈ Ω0 (ii) β < 1. Let X nm Ynm T n = Xm for 1 ≤ m ≤ n, ∀n = Xnm n(3−β)/2 = Pn m=1 Ynm = Sn n(3−β)/2 1 2 Then, EYnm = 0 and EYnm = (3−β m2−β . Pn Pn n 2−β 1 1 2 ≤ n(3−β) (a) · n · n2−β = 1. m=1 Ynm = n(3−β) m=1 m Pn Pn 2 2 → c2 ≤ 1 Since m=1 Ynm is bounded and monotonically increasing, m=1 Ynm for some c > 0. Pn Pn 2 (3−β)/2 2 ). m=1 E(|Ynm | ; |Ynm | > ) = limn→∞ m=1 E(|Ynm | ; |Xnm | > ·n 3−β (3−β)/2 (3−β)/2 Since β < 1, 2 > 1. Then, for large n, ·n > n.Then, |Xnm | > · n = Pn 2 (3−β)/2 ∅. Thus, limn→∞ m=1 E(|Ynm | ; |Xnm | > · n ) = 0. Sn (a) and (b), Lindeberg Theorem implies that Tn = n(3−β)/2 ⇒ cχ. (b) limn→∞ By (iii) β = 1. ϕSn /n (t) = n Y ϕXj (t/n) j=1 62 n Y jt 1 = 1 − cos 1− j n j=1 = n Y j=1 1 1− · n −1 ! j jt · 1 − cos . n n Note that −1 Z 1 n X 1 jt 1 j 1 − cos → (1 − cos xt)dx. n n n 0 x j=1 jt 1 Also, 1j cos jt n − 1 ≤ 0 for all j and limn→∞ j cos n − 1 → 0. Thus, max1≤j≤n 1j cos jt n − 1 → 0. Exercise ?? with cj,n = − 1j 1 − cos jt n implies n Y jt 1 Sn (t) = 1 − cos 1+ − ϕ√ n j n j=1 Z 1 1 → exp − (1 − cos tx)dx . 0 x Since the last one is the ch.f of χ, it is continuous at t = 0. By Continuity theorem, we get Sn /n ⇒ χ. 63 3.5. Local Limit Theorems. No exercise 3.6. Poisson Convergence. No exercise 3.7. Poisson Processes. Exercise 3.7.1. 1 Exercise 3.7.2. 2 Exercise 3.7.3. 3 Exercise 3.7.4. 4 Exercise 3.7.5. 5 Exercise 3.7.6. 6 Exercise 3.7.7. 7 Exercise 3.7.8. 8 Exercise 3.7.9. 9 3.8. Stable Laws∗ . Exercise 3.8.1. 1 Exercise 3.8.2. 2 Exercise 3.8.3. 3 Exercise 3.8.4. 4 Exercise 3.8.5. 5 Exercise 3.8.6. 6 Exercise 3.8.7. 7 3.9. Infinitely Divisible Distributions∗ . Exercise 3.9.1. 1 Exercise 3.9.2. 2 Exercise 3.9.3. 3 Exercise 3.9.4. 4 64 3.10. Limit Theorems in Rd . Exercise 3.10.1. P (Xi ≤ x) = P (Xi ≤ x, Xj ∈ (−∞, ∞) for j 6= i) = lim P (X1 ≤ y, · · · , Xi−1 ≤ y, Xi ≤ x, Xi+1 ≤ y, · · · , Xd ≤ y) y→∞ = lim F (y, · · · , y, x, y, · · · , y), y→∞ where x lies at i-th coordinate. Exercise 3.10.2. Omit. Exercise 3.10.3. Omit. Exercise 3.10.4. Portmanteu Exercise 3.10.5. Let X be a r.v. whose ch.f is ϕ and Y = (Y1 , . . . , Yn ) be a random vector whose ch.f is ψ. Then, ϕ(t1 + · · · + td ) = E exp[i(t1 + · · · td )X] ψ(t1 , . . . , td ) = E exp[itT Y] = E exp[i(t1 Y1 + · · · + td Yd )]. Thus, Y = (X, . . . , X) Exercise 3.10.6. Let X = (X1 , . . . , Xk )T . T (⇒) ϕX1 ,··· ,Xk (t) = Eeit X = Eeit1 X1 +···+tk Xk = Eeit1 X1 · · · Eeitk Xk = k Y ϕXj (tj ) j=1 The third equality holds since Xi ’s are independent. (⇐) For A = [a1 , b1 ] × · · · × [ak , bk ] with µX (∂A) = 0, Z k Y µX (A) = lim (2π)−k ψj (tj ) ϕX (t)dt T →∞ [−T,T ]k = lim (2π)−k T →∞ = lim T →∞ k Y j=1 j=1 k Y Z [−T,T ]k j=1 (2π)−1 ψj (tj )ϕXj (tj ) dt1 · · · dtj ! Z ψj (tj )ϕXj (tj )dtj [−T,T ] 65 = k Y j=1 = k Y " −1 # Z lim (2π) T →∞ ψj (tj )ϕXj (tj )dtj [−T,T ] µXj [aj , bj ] . j=1 Since it holds for any A, X1 , · · · , Xk are independent. Exercise 3.10.7. Let X = (X1 , . . . , Xd )T , θ = (θ1 , . . . , θd )T and Γ = (Γij ) d X d X 1 Γij ti tj (⇒) ϕX (t) = exp itT θ − 2 i=1 j=1 = E exp(itT X) = E exp(it1 X1 + · · · itd Xd ) = E exp(it1 X1 ) · · · E exp(itd Xd ) # " d X 1 Γll t2l = exp itT θ − 2 l=1 The fourth equality holds since Xi ’s are independent. The fifth equality comes from Exercise 3.10.8 with c = ei . Thus, Γij = 0 for i 6= j. (⇐) Similarly, d X d X 1 ϕX (t) = exp itT θ − Γij ti tj 2 i=1 j=1 " # d X 1 = exp itT θ − Γll t2l 2 l=1 d Y 1 = exp iti θi − Γii t2i 2 i=1 = d Y ϕXi (ti ) i=1 The last equality comes from Exercise 3.10.8 with c = ei . Exercise 3.10.8. Let X = (X1 , . . . , Xd )T , θ = (θ1 , . . . , θd )T , c = (c1 , . . . , cd )T and Γ = (Γij ) : d × d matrix. We want to show that X ∼ Nd (θ, Γ) ⇐⇒ ∀c ∈ Rd , cT X ∼ N (cT θ, cT Γc). Note that ch.f of X = Nd (θ, Γ) is given by 1 ϕX (t) = exp itT θ − tT Γt . 2 66 T (⇒) ϕcT X (t) = Eeitc X T = Eei(tc) X 1 = exp i(tc)T θ − (tc)T Γ(tc) 2 1 = exp it(cT θ) − (cT Γc)t2 , 2 which is ch.f of N (cT θ, cT Γc). Thus, cT X = c1 X1 + · · · + cd Xd ∼ N (cT θ, cT Γc). T (⇐) ϕX (t) = Eeit X T = Eei·1·(t X) = ϕtT X (1) 1 T T 2 = exp i · 1 · (t θ) − t Γt · 1 2 1 T T = exp it θ − t Γt , 2 which is ch.f of Nd (θ, Γ). Thus, X ∼ Nd (θ, Γ). 4. Martingales 4.1. Conditional Expectation. Exercise 4.1.1. Since G, Ω ∈ G, Z Z P(A|G) dP = 1A dP = P(A ∩ G) ZG ZG P(A|G) dP = 1A dP = P(A) Ω Ω and the result follows. Exercise 4.1.2. Let F ∈ F. Z Z Z Z |X|2 P(|X| ≥ a|F) dP = 1|X|≥a dP ≤ dP = a−2 E |X|2 |F dP. 2 F F F a F Since F ∈ F was arbitrary, the result follows. Exercise 4.1.3. Almost surely, we have 0 ≤ EG [(|X| + θ|Y |)2 ] = EG [X 2 ] + 2θEG |XY | + θ2 EG [Y 2 ]. Thus, almost surely, D/4 = (EG |XY |)2 − EG [X 2 ]EG [Y 2 ] ≤ 0. Exercise 4.1.4. Let µ : Ω × F → [0, 1] be an r.c.p. given G. Z EG |XY | = |X(ω)Y (ω)|µ(ω̃, dω) 67 Z ≤ 1/p Z 1/q q |X(ω)| µ(ω̃, dω) |Y (ω)| µ(ω̃, dω) p = (EG |X|p )1/p (EG |Y |q )1/q . Exercise 4.1.5. Let P(a) = P(b) = P(c) = 1/3 and let F1 = σ({a}), F2 = σ({b}). Define X as a 7→ 1 X= b 7→ 2 . c 7→ 3 Then, EF1 X = a 7→ 1 c 7→ 2.5 b 7→ 2.5 , a 7→ 2 F2 E X= b 7→ 2 c 7→ 2 and F2 E F1 E X= a 7→ 1.75 b 7→ 2.5 , F1 E F2 E X= c 7→ 1.75 a 7→ 2 c 7→ 2 b 7→ 2 . Exercise 4.1.6. E[(X − EF X)2 ] + E[(EF X − EG X)2 ] = EX 2 + E[(EF X)2 ] − 2E[XEF X] + E[(EF X)2 ] + E[(EG X)2 ] − 2E[(EF X)(EG X)]. Note that E[(EF X)2 ] = E[(EF X)(EF X)] = E[EF [XEF X]] = E[XEF X], and E[(EF X)(EG X)] = E[EF [XEG X]] = E[XEG X] = E[(EG X)2 ]. Thus the above becomes = EX 2 − E[(EG X)2 ] = E[(X − EG X)2 ]. Exercise 4.1.7. Note that Var[X|F] = EF X 2 − (EF X)2 = EF [(X − EF X)2 ]. Thus, E[Var[X|F]] = E[(X − EF X)2 ]. Also, Var[EF X] = E[(EF X − EX)2 ]. Thus the result follows from Exercise 4.1.6. 68 Exercise 4.1.8. Let F = σ(N ). First, we need to check that E|X|2 < ∞. Indeed, X E|X|2 = E[|X|2 1{N =n} ] n = X = X = X ≤ X E[(Y1 + · · · + Yn )2 1{N =n} ] n E[(Y1 + · · · + Yn )2 ]E[1{N =n} ] n (nσ 2 + n2 µ2 )P[N = n] n Cn2 P[N = n] n = CEN 2 < ∞ where for the third equality we used the independence of Yi and N . Now we claim that EF X = N µ and Var[X|F] = N σ 2 . Note that these are F-measurable. For An = {N = n}, we have Z Z (Y1 + · · · + Yn ) dP X dP = An An = E[(Y1 + · · · + Yn )1An ] = E[Y1 + · · · + Yn ]E[1An ] = nµE[1An ] Z = N µ dP An and Z X 2 dP = An Z (Y1 + · · · + Yn )2 dP An = E[(Y1 + · · · + Yn )2 1An ] = E[(Y1 + · · · + Yn )2 ]E[1An ] = (nσ 2 + n2 µ2 )E[1An ] Z = (N σ 2 + N 2 µ2 ) dP. An Since F is generated by An , n ≥ 1 by complement and countable disjoint union (cf. Dynkin’s π-λ theorem), we have the above equalities for all A ∈ F. Thus EF X = N µ and Var[X|F] = EF X 2 − (EF X)2 = N σ 2 . Finally, from Exercise 4.1.7, Var[X] = Var[EF X] + E[Var[X|F]] = µ2 Var[N ] + σ 2 EN. Exercise 4.1.9. From Exercise 4.1.7, E[(Y − X)2 ] + E[(X − EX)2 ] = E[(Y − EY )2 ]. 69 Note that E[(X − EX)2 ] = EX 2 − (EX)2 = EY 2 − (EY )2 = E[(Y − EY )2 ], where the second equality holds since EX 2 = EY 2 and EX = EY . Thus E[(Y − X)2 ] = 0 and X = Y a.s. d Exercise 4.1.10. Fix c ∈ Q. Since E[Y |G] − c = Y − c, we have E|E[Y |G] − c| = E|Y − c|. Meanwhile, we also have E|E[Y |G] − c| = E|E[Y − c|G]| ≤ EE[|Y − c||G] = E|Y − c| and thus E|E[Y − c|G]| = EE[|Y − c||G] a.s. Since |E[Y − c|G]| ≤ E[|Y − c||G], it follows that |E[Y − c|G]| = E[|Y − c||G] a.s. Now let A = {E[Y − c|G] > 0} ∈ G. Multiplying 1A both sides, the last display becomes E[(Y − c)1A |G] = E[Y − c|G]1A = |E[Y − c|G]|1A = E[|Y − c||G]1A = E[|Y − c|1A |G]. Since (Y − c)1A ≤ |Y − c|1A , it again implies that (Y − c)1A = |Y − c|1A a.s. A similar argument with B = {E[Y − c|G] ≤ 0} in place of A will show that −(Y − c)1B = |Y − c|1B a.s. Thus sgn(Y − c) = sgn(E[Y − c|G]) = sgn(E[Y |G] − c). Now suppose that {Y 6= E[Y |G]} has positive probability. Without loss of generality we may assume that {Y < E[Y |G]} has a S positive probability. Since {Y < E[Y |G]} = c∈Q {Y < c < E[Y |G]}, there must exist some c ∈ Q such that {Y < c < E[Y |G]} has a positive probability. This is a contradiction because then with a positive probability, sgn(Y − c) 6= sgn(E[Y |G] − c). Therefore Y = E[Y |G] a.s. 4.2. Martingales, Almost Sure Convergence. Exercise 4.2.1. Since X1 , · · · , Xn are measurable with respect to Gn , Fn ⊂ Gn . Note that EFn [Xn+1 ] = EFn [EGn Xn+1 ] = EFn Xn = Xn . Thus Xn is a martingale with respect to Fn . Exercise 4.2.2. Let Xn = −1/n. Then EFn Xn+1 = −1/(n + 1) > −1/n = Xn and 2 = 1/(n + 1)2 < 1/n2 = Xn2 . EFn Xn+1 Exercise 4.2.3. Note that, almost surely, EFn [Xn+1 ∨ Yn+1 ] = EFn [Xn+1 1{Xn+1 ≥Yn+1 } ] + EFn [Yn+1 1{Xn+1 <Yn+1 } ] ≥ EFn [Yn+1 1{Xn+1 ≥Yn+1 } ] + EFn [Yn+1 1{Xn+1 <Yn+1 } ] = EFn [Yn+1 ] = Yn , and similarly EFn [Xn+1 ∨ Yn+1 ] ≥ Xn a.s. Thus EFn [Xn+1 ∨ Yn+1 ] ≥ Xn ∨ Yn . 70 Exercise 4.2.4. Let TM = inf{n ≤ 0 : Xn+ > M }. Consider the stopped process XTM ∧n . Note that XT+M ∧n ≤ M + supj ξj+ , which leads to EXT+M ∧n ≤ M + E supj ξj+ < ∞ for all n. Thus XTM ∧n converges a.s. That is, Xn converges on the event {TM = ∞}. Since we have supn Xn < ∞, if we let M → ∞, {TM = ∞} ↑ Ω up to a measure zero set. Therefore, Xn converges a.s. Exercise 4.2.5. Let ξj be defined as 1 with probability −1 j + 1 ξj = 1 j with probability j j + 1 P where {j } satisfies j j < ∞. By the Borel-Cantelli lemma, P(ξj 6= −1 i.o.) = 0 since P P(ξj 6= −1) < ∞. That is, ξj = −1 eventually a.s. and thus Xn → −∞ a.s. Exercise 4.2.6. (i) Note that Xn ≥ 0. Since X is a martingale, it is a supermartingale, and thus Xn converges a.s. to an L1 random variable X. To see X = 0 a.s., note that P(|Xn+1 − Xn | > δ) = P(Xn |Yn+1 − 1| > δ) ≥ P(Xn > δ, |Yn+1 − 1| > ) = P(Xn > δ)P(|Yn+1 − 1| > ). Here, P(|Xn+1 − Xn | > δ) converges to 0 since Xn converges a.s. Since P(|Yn+1 − 1| > p ) > 0 for small , this implies that P(Xn > δ) → 0. That is, Xn → 0. Therefore X = 0 a.s. (ii) By applying Jensen’s inequality to log(Y ∨ δ), we have E log(Y ∨ δ) ≤ log E[Y ∨ δ]. Letting δ → 0, we have E log Y ≤ log EY = 0. By the strong law of large numbers, it follows that 1 X 1 log Xn = log Ym → E log Y ≤ 0. n n m≤n (Note that this holds even when E log Y = −∞.) Exercise 4.2.7. Confer Proposition 3.1 in Chapter 5 of Complex Analysis by Stein and Shakarchi. Exercise 4.2.8. Observe that, from the given inequality, Xn+1 Xn E Qn Fn ≤ Qn−1 . (1 + Y ) m m=1 m=1 (1 + Ym ) Thus the random variable Wn = Qn−1 Xn m=1 (1 + Ym ) , which is also adapted and L1 , is a nonnegative supermartingale, and thus converges a.s. Since Qn−1 m=1 (1 + Ym ) converges a.s. by Exercise 4.2.7, it follows that Xn converges a.s. 71 Exercise 4.2.9. (i) On {N > n}, 1 2 E[Yn+1 1N >n |Fn ] = E[Xn+1 1N >n+1 + Xn+1 1n<N ≤n+1 |Fn ] 1 1 ≤ E[Xn+1 1N >n+1 |Fn ] + E[Xn+1 1n<N ≤n+1 |Fn ] 1 = E[Xn+1 1N >n |Fn ] 1 = E[Xn+1 |Fn ]1N >n = Xn1 1N >n . Meanwhile, on {N ≤ n}, 2 E[Yn+1 1N ≤n |Fn ] = E[Xn+1 1N ≤n |Fn ] 2 = E[Xn+1 |Fn ]1N ≤n ≤ Xn2 1N ≤n . Thus E[Yn+1 |Fn ] ≤ Xn1 1N >n + Xn2 1N ≤n = Yn , and Yn is a supermartingale. (ii) Use the similar argument as the above. Exercise 4.2.10. content 4.3. Examples. Exercise 4.3.1. Let X0 = 0. Define Xn+1 as follows: If Xn = 0, then Xn+1 = 1 or −1 with probability 1/2 each. If Xn 6= 0, then Xn+1 = 0 or n2 Xn with probability 1 − 1/n2 and 1/n2 respectively. Note that E[Xn+1 |Fn ] = Xn . i.e. Xn is indeed a martingale. Note also that, P since n 1/n2 < ∞, by Borel-Cantelli, P(Xn+1 = n2 Xn i.o.|Xn 6= 0) = 0. That is, given Xn 6= 0, Xn+1 = 0 except for finitely many times. Thus P(Xn = a i.o.) = 1 for a = −1, 0, 1. Exercise 4.3.2. Exercise 4.3.3. Observe that from the given inequality, E[Xn+1 − n X Yk |Fn ] ≤ Xn − k=1 n−1 X Yk . k=1 Thus the random variable Zn = Xn − n−1 X Yk , k=1 Pn which is also adapted and L1 , is a supermartingale. Let NM = inf{ k=1 Yk > M }. Then PNM ∧n ZNM ∧n is also a supermartingale. Since ZNM ∧n + M = XNM ∧n + M − k=1 Yk ≥ 0, it is a nonnegative supermartingale, and thus converges a.s. to an L1 limit. That is, Zn P converges to an L1 limit on the event {NM = ∞}. Since Yk < ∞ a.s., if we let M → ∞, {NM = ∞} ↑ Ω up to a measure zero set. Therefore, Zn converges a.s. to an L1 limit Z∞ Pn−1 P a.s. and thus Xn = Zn + k=1 Yk → Z∞ + Yk < ∞. 72 Exercise 4.3.4. Let (Am )m be a sequence of independent events with P(Am ) = pm . By the P Borel-Cantelli lemma, we know that pm = ∞ if and only if P(Am i.o.) = 1. Thus we show Q that P(Am i.o.) = 1 if and only if (1 − pm ) = 0. Suppose first that P(Am i.o.) = 1. Then 0 = P(Acm eventually) \ Acm ) = lim P( n→∞ m≥n Y = lim n→∞ (1 − pm ). m≥n Note that ∞ Y (1 − pm ) ≤ m=1 Y (1 − pm ) → 0 m≥n Q Q and thus (1 − pm ) = 0. Now for the other direction, suppose that (1 − pm ) = 0. Since Q pm < 1, 1≤m≤n (1 − pm ) 6= 0, and thus it follows that Y \ 0= (1 − pm ) = P( Acm ) m≥n+1 for all n. Letting n → ∞, we get P(Acm m≥n+1 eventually) = 0 and thus P(Am i.o.) = 1. c Exercise 4.3.5. Suppose first that P(An | ∩n−1 m=1 Am ) = 1 for some n. This means that n−1 c n c c An ⊃ ∩n−1 m=1 Am up to a measure zero set. That is, ∩m=1 Am = ∩m=1 Am \ An is measure zero. n−1 c c Then we immediately have P(∩∞ m=1 Am ) = 0. Thus we may assume that P(An | ∩m=1 Am ) < 1 for all n. By Exercise 4.3.4, it follows that 0= = = ∞ Y c P(Acn | ∩n−1 m=1 Am ) n=2 ∞ Y P(∩nm=1 Acm ) c P(∩n−1 m=1 Am ) n=2 c P(∩∞ m=1 Am ) c P(A1 ) c c and thus P(∩∞ m=1 Am ) = 0. (Note that P(A1 ) 6= 0 from the above assumption.) Exercise 4.3.6. (i) Note first that Xn is Fn -measurable and in L1 . To see it is a martin- gale, we show that Z Z Xn+1 dν = Ik,n Xn dν Ik,n for all k. If ν(Ik,n ) = 0, then Xn 1Ik,n = 0 and Xn+1 1Ik,n = 0, and thus the above holds. Now assume ν(Ik,n ) 6= 0 and suppose that Ik,n = Ip+1,n+1 ∪ · · · ∪ Iq,n+1 ∪ Iq+1,n+1 ∪ · · · ∪ Ir,n+1 where ν(Ip+1,n+1 ), · · · , ν(Iq,n+1 ) are nonzero and ν(Iq+1,n+1 ) = · · · = ν(Ir,n+1 ) = 73 0. Then we have Z Xn+1 dν = µ(Ip+1,n+1 ) + · · · + µ(Iq,n+1 ) Ik,n = µ(Ip+1,n+1 ) + · · · + µ(Iq,n+1 ) + µ(Iq+1,n+1 ) + · · · + µ(Ir,n+1 ) = µ(Ik,n ) Z = Xn dν. Ik,n Thus Xn is a martingale. (ii) Again, note that Xn is Fn -measurable and in L1 . Here, we show that Z Z Xn+1 dν ≤ Xn dν Ik,n Ik,n for all k. If ν(Ik,n ) = 0, then the inequality trivially holds as the above. Assume ν(Ik,n ) 6= 0. With the same notation as the above, we have Z Xn+1 dν = µ(Ip+1,n+1 ) + · · · + µ(Iq,n+1 ) Ik,n ≤ µ(Ip+1,n+1 ) + · · · + µ(Iq,n+1 ) + µ(Iq+1,n+1 ) + · · · + µ(Ir,n+1 ) = µ(Ik,n ) Z = Xn dν. Ik,n Exercise 4.3.7. We first assume that µ and ν are finite measures. Let Fn = σ(Am : 1 ≤ m ≤ n) and let µn and νn be restrictions of µ and ν to Fn respectively. Obviously µn νn . Since Fn is finitely generated for each n, there exists a family of partitions (Ik,n )k,n such that Fn = σ(Ik,n : 1 ≤ k ≤ Kn ) and {Ik,n+1 : 1 ≤ k ≤ Kn+1 } is a refinement of {Ik,n : 1 ≤ k ≤ Kn }. (Check this.) Define Xn as in Example 4.3.7. Then by definition, Xn satisfies Z (∗) µn (A) = Xn dνn A for A ∈ Fn . (We don’t define Xn as the Radon-Nikodym derivative because we are in the way of proving the Radon-Nikodym theorem.) Moreover, as we saw in Exercise 4.3.6, Xn is a martingale, and in particular it is a nonnegative supermartingale. By Theorem 4.2.12, a.s. Xn → X where X ∈ L1 (ν). Now by the proof of Theorem 4.3.5, Z µ(A) = X dν + µ(A ∩ {X = ∞}). A (Even though Xn is not a Radon-Nikodym derivative, it plays the same role as in the proof of Theorem 4.3.5 because of (∗) above. Thus the proof works.) Since X ∈ L1 (ν), ν({X = ∞}) = 0 and by the absolute continuity µ({X = ∞}) = 0. Therefore, Z µ(A) = X dν. A 74 Finally, note that X is measurable with respect to F = σ(An : n ∈ N). This proves the Radon-Nikodym theorem for finite measures. Now, let µ and ν be σ-finite. Let Ωk ↑ Ω be satisfying µ(Ωk ) < ∞ and ν(Ωk ) < ∞ for all k. (Check that we can always find such Ωk ’s.) Define µ(k) (·) = µ(· ∩ Ωk ) and ν (k) (·) = ν(· ∩ Ωk ). Then µ(k) and ν (k) are finite measures. By the above, we have Z (k) µ (A) = X (k) dν (k) A Where X (k) is F-measurable and supported on Ωk . Define X to be equal a.s. to Xk on Ωk . Note that for j < k, µ(j) (A) = µ(k) (A ∩ Ωj ) = Z X (k) dν (k) = A∩Ωj Z X (k) dν (j) . A Because of the uniqueness of (almost sure) limit, X (k) =X (j) on Ωj . Thus X is well-defined. Letting k → ∞ and using the continuity of measures and the definition of Lebesgue integral, Z Z (k) (k) µ(A) = lim µ (A) = lim X dν = X dν. k→∞ k→∞ A A Finally it is obvious that X is F-measurable. Exercise 4.3.8. (i) Note that Z 1 − αn = Fn (0) = qn (x) dGn (x) = qn (0)(1 − βn ), (−∞,0] Z qn (x) dGn (x) = qn (0)(1 − βn ) + qn (1)βn . 1 = Fn (1) = (−∞,1] Thus qn (0) = (1 − αn )/(1 − βn ) and qn (1) = αn /βn . By Theorem 4.3.8, µ ν if and only if ∞ Z Y √ qm dGm > 0. m=1 Note that ∞ Z Y √ qm dGm = m=1 = ∞ r Y 1 − αm m=1 ∞ p Y 1 − βm r (1 − βm ) + (1 − αm )(1 − βm ) + αm βm βm p αm βm . m=1 Thus the condition becomes ∞ p Y p (1 − αm )(1 − βm ) + αm βm > 0. m=1 (ii) 75 P Exercise 4.3.9. By the Borel-Cantelli lemma, αn < ∞ implies that ξn = 0 eventually µP a.s. and βn = ∞ implies that ξn = 1 infinitely often ν-a.s. Since {ξn = 0 eventually}∩{ξn = 1 i.o.} = ∅, we have µ ⊥ ν. Exercise 4.3.10. content Exercise 4.3.11. content Exercise 4.3.12. content Exercise 4.3.13. content 4.4. Doob’s Inequality, Convergence in Lp , p > 1. Exercise 4.4.1. Since X is a submartingale, E[Xk 1N =j |Fj ] = E[Xk |Fj ]1N =j ≥ Xj 1N =j . Now take the expectation of both sides. Exercise 4.4.2. On {M = N }, the result is trivial. On {M < N }, let Kn = 1M <n≤N . Then K is predictable. Note that (K · X)n = XN ∧n − XM ∧n . Since (K · X) is a submartingale, from E(K · X)0 ≤ E(K · X)k , we have EXM ≤ EXN . (Another proof ) Let Yn = XN ∧n . Then Y is a submartingale. By Theorem 4.4.1, EXM = EYM ≤ EYk = EXN . Exercise 4.4.3. We first observe that A ∈ FN . To see this, note that {N ≤ n} ∩ A = {N ≤ n} ∩ {M ≤ n} ∩ A, and that {M ≤ n} ∩ A ∈ Fn and {N ≤ n} ∈ Fn . Thus {N ≤ n} ∩ A ∈ Fn , meaning that A ∈ FN . Now consider {L ≤ n}. Observe that {L ≤ n} = ({L ≤ n} ∩ A) ∪ ({L ≤ n} ∩ Ac ) = ({M ≤ n} ∩ A) ∪ ({N ≤ n} ∩ Ac ) ∈ Fn since A ∈ FM and A ∈ FN . Thus L is a stopping time. Exercise 4.4.4. Let A ∈ FM and L be the stopping time defined as above. Since L ≤ N , by Exercise 4.4.2, EXL ≤ EXN . Note that EXL = EXL 1A + EXL 1Ac = EXM 1A + EXN 1Ac . Plugging this into the inequality, we have EXN ≥ EXM 1A + EXN 1Ac , which leads to EXN 1A ≥ EXM 1A . Since A ∈ FM was arbitrary, it follows that E[XN |FM ] ≥ XM . Exercise 4.4.5. We show that E[E[Y |G]E[Y |F]] = E[E[Y |F]2 ]. Since F ⊂ G, E[E[Y |G]E[Y |F]] = E[E[Y E[Y |F]|G]] = E[Y E[Y |F]] = E[E[Y E[Y |F]|F]] = E[E[Y |F]2 ]. This proves the result. 76 Exercise 4.4.6. Let A = {max1≤m≤n |Sm | > x}, and let N = inf{m ≥ 0 : |Sm | > x} and T = N ∧ n. Then N and T are stopping times. Since Sn2 − s2n is a martingale, by Exercise 4.4.1, 0 = E[S02 − s20 ] = E[ST2 − s2T ] 2 = E[(SN − s2N )1A ] + E[(Sn2 − s2n )1Ac ] ≤ (x + K)2 P(A) + (x2 − s2n )P(Ac ). Thus we have 0 ≤ (x + K)2 − (x + K)2 P(Ac ) + (x2 − s2n )P(Ac ) and it follows that P(Ac ) ≤ (x + K)2 (x + K)2 ≤ . s2n + (x + K)2 − x2 s2n Exercise 4.4.7. Let c ≥ −λ. Since (Xn + c)2 is a nonnegative submartingale, by Doob’s inequality, P( max Xm ≥ λ) = P( max (Xm + c)2 ≥ (λ + c)2 ) 1≤m≤n 1≤m≤n ≤ E(Xn + c)2 (λ + c)2 = EXn2 + c2 (λ + c)2 Optimizing over c ≥ −λ, the right-hand-side is minimized when c = EXn2 /λ. Plugging this c into the inequality, the result follows. Exercise 4.4.8. (i) Z ∞ E(X̄n ∧ M ) = P(X̄n ∧ M > x) dx 0 Z 1 ≤ 0 Z dx + Z ∞ ∞ P(X̄n ∧ M > x) dx 1 1 E(X̄n+ 1X̄n ∧M >x ) dx x 1 Z ∞Z 1 + =1+ X̄n 1X̄n ∧M >x dP dx x 1 Z Z X̄n ∧M 1 + =1+ X̄ dx dP x n Z 1 = 1 + X̄n+ log(X̄n ∧ M ) dP ≤1+ 77 (ii) Let f (a, b) = (a log b − a log a)/b. From ∇f = 0 we get log b = log a + 1, or simply b = ea. Thus the maximum of f is f (a, ea) = 1/e which proves the result. Exercise 4.4.9. From EXn < ∞ and EYn < ∞, all the expectations are finite. Note that EXm−1 Ym = E[E[Xm−1 Ym |Fm−1 ]] = E[Xm−1 E[Ym |Fm−1 ]] = EXm−1 Ym−1 . Thus Pn Pn m=1 E(Xm − Xm−1 )(Ym − Ym−1 ) = m=1 (EXm Ym − EXm−1 Ym−1 ) = EXn Yn − EX0 Y0 . Exercise 4.4.10. Note that for all n, EXn2 = E(X0 + ξ1 + · · · + ξn )2 = EX02 + n X 2 Eξm +2 m=1 = EX02 + ≤ EX02 + n X m=1 ∞ X n X EX0 ξm + 2 m=1 n X Eξm ξk m,k=1 2 Eξm 2 Eξm <∞ m=1 where the third equality is due to Theorem 4.4.7. By Theorem 4.4.6, the result follows. Exercise 4.4.11. For all n, n Xn2 1 1 X 2 2 = EX + Eξ . 0 b2n b2n b2n m=1 m Pn 2 By Kronecker’s lemma (Theorem 2.5.9), m=1 Eξm /b2m converges to 0. Also, since EX02 < ∞ E and bn ↑ ∞, EX02 /b2n converges to 0. Therefore EXn2 /b2n converges to 0, which implies that sup EXn2 /b2n < ∞. Since Xn /bn is a submartingale, by Theorem 4.4.6 Xn /bn a.s. and in L2 . As Xn /bn → 0 in L2 , Xn /bn → 0 a.s. 4.6. Uniform Integrability, Convergence in L1 . Exercise 4.6.1. Let Fn = σ(Y1 , · · · , Yn ) and define Xn = E[θ|Fn ]. Then X is a martingale with filtration (Fn ). By Theorem 4.6.8, Xn converges a.s. to E[θ|F∞ ]. Now it remains to show that θ = E[θ|F∞ ]. i.e. θ is F∞ -measurable. It follows by the strong law of large numbers that almost surely Y1 + · · · + Yn →θ n and thus θ is F∞ -measurable. Exercise 4.6.2. Obviously Xn is Fn -measurable and Xn ∈ L1 for all n since |Xn | ≤ K as f is Lipschitz. Note that EXn+1 1Ik,n = EXn+1 1I2k+1,n+1 + EXn+1 1I2k,n+1 78 = f 2k+2 2n+1 2k+1 2n+1 −f 1 1 2n+1 2n+1 2k + 2 2k =f −f 2n+1 2n+1 + f 2k+1 2n+1 −f 1 2n+1 2k 2n+1 1 2n+1 and that f k+1 2n k 2n −f 1 2n 2k + 2 2k =f − f . 2n+1 2n+1 EXn 1Ik,n = 1 2n Thus EXn+1 1Ik,n = EXn 1Ik,n for all k, and by the definition of conditional expectation we have Xn = E[Xn+1 |Fn ]. That is, (Xn ) is a martingale. Since |Xn | ≤ K for all n, the family (Xn ) is uniformly integrable. Therefore, by Theorem 4.6.7, Xn converges a.s. and in L1 to an L1 random variable X∞ . Let ak and bk be endpoints of some Ik,n . Observe that Z bk f (bk ) − f (ak ) = Xn (ω) dω. ak By the dominated convergence, it follows that Z bk f (bk ) − f (ak ) = X∞ (ω) dω ak for all such points ak and bk . Since f is continuous and |X∞ | ≤ K, letting ak → a and bk → b we have the last assertion. Exercise 4.6.3. By Theorem 4.6.8, E[f |Fn ] → E[f |F∞ ] a.s. and in L1 . Now it remains to show that f is F∞ -measurable, which is obvious. Exercise 4.6.4. Let ω ∈ {lim sup Xn < ∞}. We will show that ω ∈ D. Let M (ω) = lim sup Xn (ω) < ∞. Then there exists N (ω) such that Xn (ω) < 2M (ω) for all n ≥ N (ω). From the given assumption, it follows that P(D|X1 , · · · , Xn )(ω) ≥ δ(x) for all n ≥ N (ω). Since D ∈ σ(X1 , X2 , · · · ), by Theorem 4.6.9, the left-hand-side of the above display converges to 1D (ω) for all ω ∈ {lim sup Xn < ∞} except on a measure zero set. That is, we have 1D ≥ δ(x) > 0 for all ω ∈ {lim sup Xn < ∞} \ E where P(E) = 0. It follows that 1D (ω) = 1 and thus ω ∈ D. i.e. {lim sup Xn < ∞} \ E ⊂ D. The result now immediately follows. Exercise 4.6.5. content Exercise 4.6.6. Note that Xn is a martingale as E[Xn+1 |Fn ] = Xn (α+βXn )+(1−Xn )βXn = Xn . Since |Xn | ≤ 1 for all n, it is a uniformly integrable martingale. Thus Xn converges to an L1 random variable X almost surely (and in L1 ). 79 To see X ∈ {0, 1}, we first let Bn = {Xn+1 = α + βXn } and B = lim sup Bn . For ω ∈ B, infinitely often, Xn+1 (ω) − Xn (ω) = α + βXn (ω) − Xn (ω) = α(1 − Xn (ω)). Since Xn converges almost surely, |Xn+1 (ω) − Xn (ω)| < for large enough n. That is, infinitely often we have α|1 − Xn (ω)| < . Thus X(ω) = 1 for almost every ω ∈ B. Now consider C = B c = lim inf Bnc . For ω ∈ C, eventually, Xn+1 = βXn which then immediately implies that X(ω) = 0. Thus X(ω) = 0 for almost every ω ∈ C. This proves that X ∈ {0, 1} Finally, since Xn is a martingale, we have EX0 = EXn and by the dominated convergence EXn → EX. Thus θ = EX0 = EX = 1 · P(X = 1) + 0 · P(X = 0) = P(X = 1). Exercise 4.6.7. Note that E|E[Yn |Fn ] − E[Y |F∞ ]| ≤ E|E[Yn |Fn ] − E[Y |Fn ]| + E|E[Y |Fn ] − E[Y |F∞ ]| ≤ E|Yn − Y | + E|E[Y |Fn ] − E[Y |F∞ ]| By Theorem 4.6.8, E[Y |Fn ] → E[Y |F∞ ] in L1 and Yn → Y in L1 by the assumption. 4.7. Backwards Martingales. Exercise 4.7.1. By the Lp maximum inequality, for all n ≤ 0, p p p E sup |Xm | ≤ E|X0 |p . p−1 n≤m≤0 Letting n → −∞, it follows that supm≤0 |Xm | ∈ Lp . Let X−∞ be the limit of the backward martingale Xn (which exists by Theorem 4.7.1). Noting that |Xn −X−∞ |p ≤ (2 supm≤0 |Xm |)p , by the dominated convergence we have E|Xn − X−∞ |p → 0. Exercise 4.7.2. Note that |E[Yn |Fn ] − E[Y−∞ |F−∞ ]| ≤ |E[Yn |Fn ] − E[Y−∞ |Fn ]| + |E[Y−∞ |Fn ] − E[Y−∞ |F−∞ ]| ≤ E[|Yn − Y−∞ ||Fn ] + |E[Y−∞ |Fn ] − E[Y−∞ |F−∞ ]|. 80 By Theorem 4.7.3, |E[Y−∞ |Fn ] − E[Y−∞ |F−∞ ]| → 0. To see the convergence of the first term, let WN = supm,n≤N |Ym − Yn |. Then WN ≤ 2Z ∈ L1 and WN → 0. It follows that lim sup E[|Yn − Y−∞ ||Fn ] ≤ lim sup E[WN |Fn ] n→−∞ n→−∞ = E[WN |F−∞ ] →0 where the last convergence is due to the dominated convergence theorem. Exercise 4.7.3. Let p = P(X1 = 1, · · · , Xm = 1). Because (Xn ) is exchangeable, P(Xn1 = n 1, · · · , Xnm = 1) = p for any subsequence (nk ). Thus P(Sn = m) = m p. Similarly, P(X1 = 1, · · · , Xk = 1, Xn1 = 1, · · · , Xnm−k = 1) = p for any (nk ), and thus P(X1 = n−k 1, · · · , Xk = 1, Sn = m) = m−k p. Therefore, n−k n n−k n P(X1 = 1, · · · , Xk = 1|Sn = m) = = m−k m n−m m Exercise 4.7.4. Note that from the exchangeability, 0 ≤ E(X1 + · · · + Xn )2 = nEX12 + n(n − 1)EX1 X2 and thus EX1 X2 ≥ − EX12 → 0. (n − 1) Exercise 4.7.5. Let ϕ(x, y) = (x − y)2 and let −1 X n An (ϕ) = 2 ϕ(Xi , Xj ). 1≤i<j≤n Since An (ϕ) ∈ En , we have An (ϕ) = E[An (ϕ)|En ] −1 X n = 2 E[ϕ(Xi , Xj )|En ] 1≤i<j≤n = E[ϕ(X1 , X2 )|En ]. By Theorem 4.7.3, E[ϕ(X1 , X2 )|En ] → E[ϕ(X1 , X2 )|E]. (Note that this is possible because ϕ(X1 , X2 ) is integrable.) Now since E is trivial due to Example 4.7.6, it follows that E[ϕ(X1 , X2 )|E] = Eϕ(X1 , X2 ) = EX12 + EX22 − 2EX1 EX2 = 2σ 2 . 4.8. Optional Stopping Theorems. 81 Exercise 4.8.1. Let A ∈ FL and let N = L1A + M 1Ac . Note that N is a stopping time according to Exercise 4.4.3 and that N ≤ M . Since YM ∧n is a uniformly integrable submartingale, applying Theorem 4.8.3 to this leads to EY0 ≤ EYN ≤ EYM . Observe that EYN = EYL 1A + EYM 1Ac . It follows that EYL 1A ≤ EYM 1A . Since A ∈ FL was arbitrary, by the definition of conditional expectation, YL ≤ E[YM |FL ] and also EYL ≤ EYM . Exercise 4.8.2. Let N = inf{m : Xm > λ}. Since XN 1N <∞ ≥ λ, λP(N < ∞) ≤ EXN 1N <∞ ≤ EX0 by Theorem 4.8.4. Observe that {N < ∞} = {supm Xm > λ}. Exercise 4.8.3. Let Xn = Sn2 − nσ 2 . Note that E|XT | ≤ EST2 = EST2 1T <∞ + EST2 1T =∞ ≤ E(a + ξ)2 1T <∞ + a2 ≤ 2a2 + σ 2 < ∞. To see E[|Xn 1T >n |; |Xn 1T >n | > a2 + M ] = 0 for all n, we observe that |Xn 1T >n | ≤ a2 1T >n ≤ a2 for all n and thus {|Xn 1T >n | > a2 +M } = ∅. By Theorem 4.8.2, XT ∧n is a uniformly integrable martingale. It follows that 0 = EXT ∧n → EXT = E[ST2 − T σ 2 ]. If ET = ∞, then the result is trivial. For ET < ∞, we have T < ∞ a.s. and thus σ 2 ET = EST2 ≥ a2 . Exercise 4.8.4. Let Xn = Sn2 − nσ 2 . Then Xn is a martingale. Since T is a stopping time with ET < ∞, by the monotone convergence, EST2 ∧n = σ 2 E[T ∧ n] → σ 2 ET. In particular, EST2 ∧n ≤ σ 2 ET for all n. i.e. supn EST2 ∧n < ∞. Since ST ∧n is a martingale, this implies that ST ∧n converges a.s. and in L2 to ST . Therefore, EST2 ∧n → EST2 and the result follows. 82 Exercise 4.8.5. (a) Let A = p − q and B = 1 − (p − q)2 . Using the same argument as in Exercise 4.8.3, it follows that (SV0 ∧n − (V0 ∧ n)A)2 − (V0 ∧ n)B is a uniformly integrable martingale. Thus we have BEV0 = E(SV0 − V0 A)2 = ESV20 − 2AESV0 V0 + A2 EV02 . Now since p < 1/2, V0 < ∞ a.s. and SV0 = 0. Thus the above display becomes BEV0 = A2 EV02 . Therefore, the second moment of V0 is (not the variance; probably a typo) EV02 = B 1 − (p − q)2 EV = x. 0 A2 (q − p)3 Exercise 4.8.6. (a) Let φ(θ) = E exp(θξi ) = peθ + qe−θ and Xn = exp(θSn )/φ(θ)n . Then Xn is a martingale and thus eθx = EX0 = EXV0 ∧n =E exp(θSV0 ∧n ) . φ(θ)V0 ∧n Now note that, since θ ≤ 0 and SV0 ∧n ≥ 0, exp(θSV0 ∧n ) ≤ 1 and φ(θ) ≥ 1. By the bounded convergence theorem, it follows that exp(θSV0 ∧n ) φ(θ)V0 ∧n exp(θSV0 ) →E φ(θ)V0 eθx = E = E[φ(θ)−V0 ]. (b) From φ(θ) = 1/s, we get pe2θ − eθ /s + q = 0. Solving this with respect to eθ and noting that eθ ≤ 1, we have θ e = 1− p 1 − 4pqs2 . 2ps Thus V0 E[s ] = 1− p 1 − 4pqs2 2ps !x . 4 2 Exercise 4.8.7. Note that E[Sn+1 |Fn ] = Sn4 + 6Sn2 + 1 and E[Sn+1 |Fn ] = Sn2 + 1. From E[Yn+1 |Fn ] = Yn , we get (2b − 6)n + (b + c − 5) = 0 and thus b = 3 and c = 2. Now from EY0 = EYT ∧n , 0 = EST4 ∧n − 6E(T ∧ n)ST2 ∧n + 3E(T ∧ n)2 + 2E(T ∧ n). 83 Noting that |ST ∧n | ≤ a and that ET < ∞ (from the proof of Theorem 4.8.7), apply the DCT and MCT to get 0 = a4 − 6a2 ET + 3ET 2 + 2ET. Now ET = a2 leads to ET 2 = 5a4 − 2a2 . 3 Exercise 4.8.8. Note that Sτ ∧n ≥ a, and hence Xn ≤ eaθ0 for all n. Thus Xτ ∧n is a uniformly integrable martingale, and by Theorem 4.8.2, 1 = EX0 = EXτ . Now note that 1 = EXτ ≥ E[Xτ ; Sτ ≤ a] ≥ eaθ0 P(Sτ ≤ a) from which the result follows. Exercise 4.8.9. By the same argument as in Exercise 4.8.8, XT ∧n is a uniformly integrable martingale and we have 1 = EX0 = EXT . Now observe that 1 = EXT = E[XT ; T < ∞] + E[XT ; T = ∞] = eaθ0 P(T < ∞) + E[X∞ ; T = ∞]. Here, X∞ = exp(θ0 S∞ ) = 0 since S∞ = lim Sn = ∞ a.s. by the strong law of large numbers. Therefore, P(T < ∞) = e−aθ0 . Exercise 4.8.10. Note that φ(θ) = E exp(θξ) = (e−θ + eθ + e2θ )/3. Solving φ(θ) = 1 with √ √ θ ≤ 0, we get eθ = 2 − 1. By Exercise 4.8.9, P(T < ∞) = eθi = ( 2 − 1)i . Exercise 4.8.11. Let φ(θ) = E exp(θξ) = exp((c − µ)θ + σ 2 θ2 /2). Note that θ0 = −2(c − µ)/σ 2 ≤ 0 satisfies φ(θ0 ) = 1. For this θ0 , Xn = exp(θ0 Sn ) is a martingale. Now let T = inf{Sn ≤ 0}. By the same argument as in Exercise 4.8.8, XT ∧n is a uniformly integrable martingale. Thus E exp(θ0 S0 ) = E exp(θ0 ST ) ≥ E[exp(θ0 ST ); T < ∞] ≥ P(T < ∞). Now plugging θ0 = −2(c − µ)/σ 2 into the last display, the result follows. 4.9. Combinatorics of Simple Random Walk. Exercise 4.9.1. content 5. Markov Chains 5.1. Examples. 84 Exercise 5.1.1. P(Xn+1 = j|Xn = i, Xn−1 1 − i N = in−1 , · · · , X0 = i0 ) = i N =: p(i, j) for j = i + 1, for j = i. Exercise 5.1.2. Note that P(X4 = 2, X3 = 1, X2 = 0, X1 = 0) P(X3 = 1, X2 = 0, X1 = 0) P(S4 = 2, S3 = 1, S2 = 0, S1 = −1) 1 = = , P(S3 = 1, S2 = 0, S1 = −1) 2 P(X4 = 2|X3 = 1, X2 = 0, X1 = 0) = P(X4 = 2, X3 = 1, X2 = 1, X1 = 1) P(X3 = 1, X2 = 1, X1 = 1) P(S4 = 2, S3 = 1, S2 = 0, S1 = 1) 1 = = . P(S2 = 0, S1 = 1) 4 P(X4 = 2|X3 = 1, X2 = 1, X1 = 1) = Thus Xn is not a Markov chain. Exercise 5.1.3. Note that Xi ∈ σ(ξn , · · · , ξ0 ) for i ≤ n − 1. Since Xn+1 = (ξn+1 , ξn+2 ), Xn+1 is independent of {Xn−1 , · · · , X0 }. Thus, P(Xn+1 = j|Xn = i, Xn−1 = in−1 , · · · , X0 = i0 ) = P(Xn+1 = j|Xn = i). That is, Xn is a Markov chain. The transition probabilities are p HH HT TH TT HH HT TH 1/2 1/2 0 0 0 1/2 1/2 1/2 0 0 0 1/2 TT 0 1/2 0 1/2 p2 HH HT TH TT HH 1/4 1/4 1/4 1/4 TT 1/4 1/4 1/4 1/4 and HT 1/4 1/4 1/4 1/4 TH 1/4 1/4 1/4 1/4 Indeed, the distribution of Xn+2 = (ξn+2 , ξn+3 ) is independent of Xn = (ξn , ξn+1 ). Exercise 5.1.4. The transition probabilities are 85 AA,AA AA,Aa AA,aa Aa,Aa Aa,aa aa,aa AA,AA 1 1/4 0 1/16 0 0 AA,Aa 0 1/2 0 1/4 0 0 AA,aa 0 0 0 1/8 0 0 Aa,Aa 0 1/4 1 1/4 1/4 0 Aa,aa 0 0 0 1/4 1/2 0 aa,aa 0 0 0 1/16 1/4 1 Exercise 5.1.5. P(Xn+1 m−i × b−i m m m−i m−b+i = j|Xn = i) = + m × m i × m−b+i m for j = i + 1 i m × b−i m for j = i for j = i − 1 m Exercise 5.1.6. Let Kn = |{1 ≤ i ≤ n : Xi = 1}|. Note that Sn = Kn − (n − Kn ) = 2Kn − n, or Kn = (Sn + n)/2. From the elementary definition of conditional probability, R 1 K +1 θ n (1 − θ)n−Kn . P(Xn+1 = 1|Xn , · · · , X1 ) = 0R 1 θKn (1 − θ)n−Kn 0 Here, integrating by parts, Z 1 Z θm (1 − θ)k = (−1)2m 0 0 1 m!k! m! (1 − θ)k+m = . (k + 1) · · · (k + m) (k + m + 1)! Thus, P(Xn+1 = 1|Xn , · · · , X1 ) = Kn + 1 Sn + n + 2 = . n+2 2(n + 2) Now, note that P(Sn+1 = sn+1 |Sn , · · · , S1 ) = P(X n+1 = 1|Xn , · · · , X1 ) for sn+1 = Sn + 1 1 − P(X for sn+1 = Sn − 1 n+1 = 1|Xn , · · · , X1 ) Sn +n+2 for sn+1 = Sn + 1 2(n+2) = 1 − Sn +n+2 for s = S − 1. 2(n+2) n+1 n That is, the transition probability depends only on Sn . Therefore, Sn is a Markov chain. 5.2. Construction, Markov Properties. Exercise 5.2.1. Pµ (A ∩ B|Xn ) = Eµ [Eµ [1A 1B |Fn ]|Xn ] = Eµ [1A Eµ [1B |Fn ]|Xn ] (∗) = Eµ [1A Eµ [1B |Xn ]|Xn ] = Eµ [1A |Xn ] Eµ [1B |Xn ] Here, (∗) is from the Markov property. 86 Exercise 5.2.2. Let Fn = σ(X1 , · · · , Xn ), F∞ = σ(X1 , X2 , · · · ). Note that S m≥n+1 {Xm ∈ Bm } ∈ F∞ . By Lévy’s 0-1 law, on {Xn ∈ An }, [ [ (∗) δ ≤ P( {Xm ∈ Bm }|Xn ) = P( {Xm ∈ Bm }|Fn ) m≥n+1 m≥n+1 → 1S m≥n+1 {Xm ∈Bm } where (∗) is from the Markov property. That is, for ω ∈ {Xn ∈ An }, 1Sm≥n+1 {Xm ∈Bm } (ω) = 1, or in other words, [ {Xn ∈ An } ⊂ {Xm ∈ Bm }. m≥n+1 Thus, [ [ {Xn ∈ An } ⊂ n≥k {Xm ∈ Bm } m≥k+1 and letting k → ∞, {Xn ∈ An i.o.} ⊂ {Xn ∈ Bn i.o.}. Exercise 5.2.3. Let Aδ = {x : Px (D) ≥ δ}. Then obviously PXn (D) ≥ δ on {Xn ∈ Aδ }. Note that PXn (D) = P [ {Xm = a}|Xn ≥ δ m≥n+1 on {Xn ∈ Aδ }. By the previous exercise, P({Xn ∈ Aδ i.o.} − {Xn = a i.o.}) = 0. From {Xn = a i.o.} ⊂ D, we have {Xn ∈ Aδ i.o.} ⊂ {Xn = a i.o.} ⊂ D. This implies that P(Xn ∈ Acδ eventually) = 1 on Dc . That is, on Dc , PXn (D) ≤ δ eventually almost surely. Since δ is arbitrary, we have the desired result. Exercise 5.2.4. pn (x, y) = Px (Xn = y) = = (∗) = = n X m=1 n X Px (Ty = m)Px (Xn = y|Ty = m) Px (Ty = m)Px (Xn = y|X1 6= y, · · · , Xm−1 6= y, Xm = y) m=1 n X Px (Ty = m)Px (Xn = y|Xm = y) m=1 n X Px (Ty = m)pn−m (y, y) m=1 where (∗) is by the Markov property. 87 Exercise 5.2.5. Let T (k) = inf{i ≥ k : Xi = x} ≥ k. Then, similarly to the previous exercise, Px (Xm = x) = m X Px (T (k) = j)Px (Xm−j = x). j=k Summing over m and applying Fubini, n+k X Px (Xm = x) = m=k n+k m XX Px (T (k) = j)Px (Xm−j = x) m=k j=k = n+k X n+k X Px (T (k) = j)Px (Xm−j = x) j=k m=j = n+k X n−(j−k) Px (T (k) X = j) ≤ n+k X Px (Xm = x) m=0 j=k n X Px (T (k) = j) Px (Xm = x) m=0 j=k = Px (T (k) ≤ n + k) n X Px (Xm = x) m=0 ≤ n X Px (Xm = x). m=0 Exercise 5.2.6. Since Px (TC < ∞) > 0, there exist nx and x such that Px (TC ≤ nx ) = x > 0. Let N = maxs∈S−C nx and = mins∈S−C x . Then for all y ∈ S − C, Py (TC > N ) ≤ 1 − , and thus the desired result holds for k = 1. Now we proceed by induction. Note that Py (TC > kN ) = Py (X1 ∈ / C, · · · , XkN ∈ / C) = Py (XN +1 ∈ / C, · · · , XkN ∈ / C|X1 ∈ / C, · · · , XN ∈ / C)Py (X1 ∈ / C, · · · , XN ∈ / C) ≤ (1 − )Py (XN +1 ∈ / C, · · · , XkN ∈ / C|XN ∈ / C) Here, Py (XN ∈ / C, · · · , XkN ∈ / C) Py (XN ∈ / C) Z 1 = 1{XN +1 ∈C,··· dPy / ,XkN ∈C} / Py (XN ∈ / C) {XN ∈C} / Z 1 = Ey [1{XN +1 ∈C,··· |XN ] dPy / ,XkN ∈C} / Py (XN ∈ / C) {XN ∈C} / Z 1 = EXN [1{X1 ∈C,··· ] dPy . / ,X(k−1)N ∈C} / Py (XN ∈ / C) {XN ∈C} / Py (XN +1 ∈ / C, · · · , XkN ∈ / C|XN ∈ / C) = By the induction hypothesis, EXN [1{X1 ∈C,··· ] ≤ (1−)k−1 on {XN ∈ / C} Py -almost / ,X(k−1)N ∈C} / surely. Thus the desired result follows for general k. Exercise 5.2.7. 88 5.3. Recurrence and Transience. 5.4. Recurrence of Random Walks. 5.5. Stationary Measures. 5.6. Asymptotic Behavior. 5.7. Periodicity, Tail σ-Field. 5.8. General State Space. 6. Ergodic Theorems 6.1. Definitions and Examples. (i) Since ϕ−1 Ω ⊂ Ω and ϕ−1 Ω has probability 1, ϕ−1 Ω = Ω up to a null Exercise 6.1.1. set. Thus Ω ∈ I. (ii) Let A ∈ I. Then up to a null set, ϕ−1 (Ω \ A) = ϕ−1 Ω \ ϕ−1 A = Ω \ A, and thus Ω \ A ∈ I. (iii) Let Aj ∈ Ω for j ≥ 1. Then up to a null set, [ [ [ ϕ−1 ( Aj ) = ϕ−1 Aj = Aj , j and thus S j j j Aj ∈ I. Now to see that X ∈ I if and only if X ◦ ϕ = X a.s., suppose first that X ∈ I. Then up to a null set, {X ◦ ϕ > a} = ϕ−1 {X > a} = {X > a} since {X > a} ∈ I. Since this is true for arbitrary a, X ◦ ϕ = X a.s. (Confer the argument in Exercise 4.1.10.) For the other direction, suppose X ◦ ϕ = X a.s. Then up to a null set, {X > a} = {X ◦ ϕ > a} = ϕ−1 {X > a}, and thus {X > a} ∈ I. Exercise 6.1.2. (i) ϕ−1 (B) = ϕ−1 ( ∞ [ ϕ−n (A)) = n=0 ∞ [ ϕ−n (A) ⊂ B n=1 (ii) −1 ϕ −1 (C) = ϕ ( ∞ \ −n ϕ n=0 (B)) = ∞ \ −n ϕ n=1 since ϕ−n (B) ∩ B = ϕ−n (B). 89 (B) = ∞ \ (ϕ−n (B) ∩ B) = C n=1 (iii) Suppose that A is almost invariant. Define B = S∞ n=0 ϕ−n (A) and C = T∞ n=0 ϕ−n (B). Then by the above, C is invariant in the strict sense. Note that C= = = ∞ \ ϕ−n ( n=0 ∞ [ ∞ \ ∞ [ ϕ−k (A)) k=0 ϕ−(n+k) (A) n=0 k=0 ∞ [ \ ϕ−k (A) n=0 k≥n = lim sup ϕ−n (A). n→∞ −n Thus both C\A = lim sup(ϕ (A)\A) and A\C = lim inf(A\ϕ−n (A)) are measure-zero. Conversely, suppose that there exists C invariant in the strict sense with P(A∆C) = 0. Then both ϕ−1 (A \ C) = ϕ−1 (A) \ C and ϕ−1 (C \ A) = C \ ϕ−1 (A) are measurezero. It follows that both ϕ−1 (A) \ A ⊂ (ϕ−1 (A) \ C) ∪ (C \ A) and A \ ϕ−1 (A) ⊂ (A \ C) ∪ (C \ ϕ−1 (A)) are measure-zero. Exercise 6.1.3. (i) Suppose for the sake of contradiction that xn = xm for some n 6= m. Then (n − m)θ = k ∈ N and θ ∈ Q, which is a contradiction. Thus xn are distinct in [0, 1), and hence for any N , |xm − xn | < 1/N for some m < n ≤ N . Now let > 0 and x ∈ [0, 1). Fix N > 1/. Let k be satisfying x ∈ [k/N, (k + 1)/N ). (ii) (iii) Exercise 6.1.4. Let (Yk )k∈Z be a sequence of random variables with joint distribution defined by d (Y−k , · · · , Yk ) = (X0 , · · · , X2k ). To see (Yk ) is well-defined, note that, by the stationarity of X, d d (Y−k , · · · , Yk ) = (X1 , · · · , X2k+1 ) = (X0 , · · · , X2k ). Thus, distributions of (Y−k , · · · , Yk )k are consistent, and by the Kolmogorov extension theorem, the sequence (Yk ) is well-defined. To check (Yk ) is stationary, observe that d d d (Y0 , · · · , Yk ) = (Xk , · · · , X2k ) = (X2n+k , · · · , X2n+2k ) = (Yn , · · · , Yn+k ). d d Finally, note that (X0 , · · · , Xk ) = (Xk , · · · , X2k ) = (Y0 , · · · , Yk ), and therefore (Xk ) is indeed embedded in (Yk ). Exercise 6.1.5. To see that Yk is stationary, note that (Y0 , · · · , Yk ) = (g((Xj )j≥0 ), · · · , g((Xk+j )j≥0 )) d = (g((Xn+j )j≥0 ), · · · , g((Xn+k+j )j≥0 )) 90 = (Yn , · · · , Yn+k ). To see that Yk is ergodic, note that Yk+1 = g(Xk+1 , Xk+2 , · · · ) = g(Xk ◦ ϕ, Xk+1 ◦ ϕ, · · · )) = Yk ◦ ϕ. Thus, for B ∈ B(R{0,1,··· } ), {(Y1 , Y2 , · · · ) ∈ B} = {(Y0 , Y1 , · · · ) ◦ ϕ ∈ B} = ϕ−1 {(Y0 , Y1 , · · · ) ∈ B} Thus, if we let A = {(Y0 , Y1 , · · · ) ∈ B}, then since Xn is ergodic (or in other words, since ϕ is ergodic), A = ϕ−1 A implies that A is trivial. Therefore Yn is also ergodic. Exercise 6.1.6. Fix N > n. Let xν and yν be the smallest integer ≥ 0 that satisfy ν +xν ≡ 0 mod n and ν + N − yν ≡ 0 mod n. Then xν and yν are uniformly distributed on {1, · · · , n}. Note that (Z1 , · · · , ZN ) = (Yν+1 , · · · , Yν+N ) = (Yν+1 , · · · , Yν+xν , iid blocks of length n, Yν+N −yν +1 , · · · , Yν+N ) and all the blocks that appear in the above (including the first and the last incomplete blocks) are independent. Now fix k and let x̃ν and ỹν be the smallest integer ≥ 0 that satisfy k + ν + x̃ν ≡ 0 mod n and k + ν + N − ỹν ≡ 0 mod n. Then again x̃ν and ỹν are uniformly distributed on {1, · · · , n}. Note again that (Zk+1 , · · · , Zk+N ) = (Yk+ν+1 , · · · , Yk+ν+N ) = (Yk+ν+1 , · · · , Yk+ν+x̃ν , iid blocks of length n, Yk+ν+N −ỹν +1 , · · · , Yk+ν+N ). Since ν is independent of Y , so is xν , yν , x̃ν , and ỹν . Thus, the distribution of (Z1 , · · · , ZN ) and (Zk+1 , · · · , Zk+N ) are equal. To see that Z is ergodic, note that the tail σ-field T of Z is contained in the tail σ-field of blocks (Ykn+1 , · · · , Ykn+n ). Since the blocks (Ykn+1 , · · · , Ykn+n ) are iid, T is trivial. As the invariant σ-field is contained in T , it is also trivial. Exercise 6.1.7. Note first that µ([a, b]) = 1 log 2 Z a b dx 1 1+b = log . 1+x log 2 1+a From the definition of ϕ, note that x = 1/(ϕ(x) + n) for n ∈ N. It follows that 1 ∞ Z 1 X n+a dx µ(ϕ−1 [a, b]) = 1 log 2 n=1 n+b 1+x = ∞ 1 X 1+n+a n+b log log 2 n=1 n+a 1+n+b = 1 1+n+a 1+b lim log log 2 n→∞ 1+a 1+n+b 91 = 1 1+b log . log 2 1+a Thus µ(ϕ−1 [a, b]) = µ([a, b]). Now let L = {A ⊂ (0, 1) : µ(ϕ−1 A) = µ(A)}. Then L is a λ-system (check) and contains all closed intervals. By Dynkin’s theorem, µ(ϕ−1 A) = µ(A) for any Borel set A. 6.2. Birkhoff ’s Ergodic Theorem. 0 00 0 Exercise 6.2.1. As in the proof of Theorem 6.2.1, let XM = X1|X|≤M and XM = X − XM . Following the proof, we know that n−1 1 X 0 0 X (ϕm ω) → E[XM |I] n m=0 M a.s.. Since n−1 1 X 0 0 X (ϕm ω) − E[XM |I] n m=0 M p ≤ (2M )p , 00 by the bounded convergence theorem, the convergence occurs in Lp as well. To handle XM part, observe that, n−1 1 X 00 m 00 X (ϕ ω) − E[XM |I] n m=0 M ≤ p n−1 1 X 00 00 kX 00 (ϕm ω)kp + kE[XM |I]kp ≤ 2 kXM kp . n m=0 M Let A= n−1 1 X 0 0 X (ϕm ω) − E[XM |I] n m=0 M B= n−1 1 X 00 m 00 X (ϕ ω) − E[XM |I]. n m=0 M and It follows that lim sup n→∞ n−1 1 X X(ϕm ω) − E[X|I] n m=0 00 ≤ lim sup(kAkp + kBkp ) ≤ 2 kXM kp . n→∞ p 00 p Letting M → ∞ and using the dominated convergence theorem, E|XM | → 0 and the result follows. Exercise 6.2.2. (i) Let hN = supm,n≥N |gm − gn |. hN ≤ 2 supk |gk |, so EhN < ∞. Note that n−1 1 X gm (ϕm ω) − E[g|I] n m=0 ≤ N −1 n−1 n−1 1 X 1 X 1 X gm (ϕm ω) + |gm (ϕm ω) − g(ϕm ω)| + g(ϕm ω) − E[g|I] n m=0 n n m=N m=N 92 ≤ N −1 n−1 n−1 1 X 1 X 1 X gm (ϕm ω) + hN (ϕm ω) + g(ϕm ω) − E[g|I] . n m=0 n n m=N m=N Since E supk |gk | < ∞, gm < ∞ a.s. and thus the first term → 0 a.s. as n → ∞. Also, since hN and g are integrable, the second term → E[hN |I] and the third term → 0 a.s. Thus lim sup n→∞ n−1 1 X gm (ϕm ω) − E[g|I] ≤ E[hN |I]. n m=0 Now letting N → ∞ and using the dominated convergence, the result follows. (ii) Note that E n−1 1 X gm (ϕm ω) − E[g|I] n m=0 ≤E ≤ n−1 n−1 n−1 1 X 1 X 1 X gm (ϕm ω) − g(ϕm ω) + E g(ϕm ω) − E[g|I] n m=0 n m=0 n m=0 n−1 n−1 1 X 1 X E |gm (ϕm ω) − g(ϕm ω)| + E g(ϕm ω) − E[g|I] n m=0 n m=0 Since gm → g in L1 , the first term → 0 as n → ∞. The second term → 0 from Theorem 6.2.1. 0 , and Mk0 = Exercise 6.2.3. Let X 0 = X − α, Xj0 (ω) = X 0 (ϕj ω), Sk0 = X00 + · · · + Xk−1 max(0, S10 , · · · , Sk0 ). By Lemma 6.2.2, E[X 0 ; Mk0 > 0] ≥ 0. That is, E[X; Mk0 > 0] ≥ αP(Mk0 > 0). Note that Mk0 > 0 if and only if Dk > α. It follows that E|X| ≥ E[X; Mk0 > 0] ≥ αP(Mk0 > 0) = αP(Dk > α). 6.3. Recurrence. Exercise 6.3.1. Note that Rn (ω) = 1{S2 −S1 6=0,··· ,Sn −S1 6=0} (ω) + · · · + 1{Sn −Sn−1 6=0} (ω) + 1 = 1{S1 6=0,··· ,Sn−1 6=0} (ϕω) + · · · + 1{S1 6=0} (ϕn−1 ω) + 1. Taking the expectation, we get ERn = P(S1 6= 0, · · · , Sn−1 6= 0) + · · · + P(S1 6= 0) + 1 = n X gm−1 . m=1 Exercise 6.3.2. Note that, since P(Xi > 1) = 0, 1, · · · , max Sm ≤ {S1 , · · · , Sn } ≤ min Sm , · · · , max Sm , m≤n m≤n 93 m≤n and thus max Sm ≤ Rn ≤ max Sm − min Sm m≤n m≤n m≤n Since our sequence is ergodic, Sn /n → EXi > 0 by the ergodic theorem. This implies that Sn → ∞ and that minm Sm > −∞ a.s. Thus, lim sup max n→∞ m≤n Rn Sm Sm ≤ lim ≤ lim inf max , n→∞ n n→∞ m≤n n n and by Theorem 6.3.1, P(A) = lim max n→∞ m≤n Sm . n To evaluate the right-hand-side, note that, for some K, lim n→∞ Sn Sm ≤ lim inf max n→∞ m≤n n n Sm ≤ lim sup max n→∞ m≤n n Sm ≤ lim sup max n→∞ K≤m≤n n Sm ≤ max . K≤m m Since K is arbitrary, by letting K → ∞, limn→∞ maxm≤n Sm /n = EXi . This completes the proof. Exercise 6.3.3. Note that X X E 1{Xm ∈B} X0 ∈ A = E 1{Xm ∈B} 1{T1 ≥m} |X0 ∈ A 1≤m≤T1 1≤m = X P(Xm ∈ B, T1 ≥ m|X0 ∈ A) 1≤m = P(X0 ∈ A)−1 X P(Xm ∈ B, T1 ≥ m, X0 ∈ A). 1≤m Let Cm = {X−m ∈ A, X−m+1 ∈ / A, · · · , X−1 ∈ / A, X0 ∈ B}. Observe that ! c K [ Cm = {Xj ∈ A for some − K ≤ j ≤ −1, X0 ∈ B}c m=1 = {Xj ∈ / A for all − K ≤ j ≤ −1} ∪ {X0 ∈ / B}. As K → ∞, from P(Xn ∈ A at least once) = 1, it follows that P(Xj ∈ / A for all − K ≤ j ≤ S −1) = P(Xj ∈ / A for all 1 ≤ j ≤ K) → 0. Thus P( Cm ) = P(X0 ∈ B). Now the first display becomes = P(X0 ∈ A)−1 X P(Cm ) = 1≤m 94 P(X0 ∈ B) . P(X0 ∈ A) Exercise 6.3.4. From Theorem 6.3.3, P̄(T1 ≥ n) = P(T1 ≥ n|X0 = 1)P(X0 = 1) ĒT1 = P(T1 ≥ n, X0 = 1). Note that P(T1 = n) = P(X1 = · · · = Xn−1 = 0, Xn = 1) ∞ X = = = m=0 ∞ X m=0 ∞ X P(X−m = 1, X−m+1 = · · · = Xn−1 = 0, Xn = 1) P(X0 = 1, X1 = · · · = Xm+n−1 = 0, Xm+n = 1) P(X0 = 1, T1 = m + n) m=0 = P(X0 = 1, T1 ≥ n) which completes the proof. 6.5. Applications. Exercise 6.5.1. content Exercise 6.5.2. content Exercise 6.5.3. content Exercise 6.5.4. content Exercise 6.5.5. content 7. Brownian Motion 7.1. Definition and Construction. Exercise 7.1.1. content Exercise 7.1.2. Let X = B1 , Y = B2 − B1 , and Z = B3 − B2 . Then X, Y , Z are iid normal random variables with mean 0 and variance 1. Note that B12 B2 B3 = X 2 (X + Y )(X + Y + Z) = X 4 + 2X 3 Y + X 2 Y 2 + X 3 Z + X 2 Y Z. Taking the expectation, it follows that EB12 B2 B3 = EX 4 + EX 2 Y 2 = 4. 7.2. Markov Property, Blumenthal’s 0-1 Law. 95 7.3. Stopping Times, Strong Markov Property. 7.4. Path Properties. 7.5. Martingales. 7.6. Itô’s Formula. 8. Applications to Random Walk 8.1. Donsker’s Theorem. 8.4. Empirical Distributions, Brownian Bridge. 8.5. Laws of Iterated Logarithm. 9. Multidimensional Brownian Motion 9.1. Martingales. 9.5. Dirichlet Problem. end of document Email address: hoil.lee@kaist.ac.kr Email address: swj777888@snu.ac.kr 96