Math 554 Iowa State University Introduction to Stochastic Processes Department of Mathematics

advertisement
Math 554
Introduction to Stochastic Processes
Instructor: Alex Roitershtein
Iowa State University
Department of Mathematics
Fall 2015
Solutions to Homework #2
1.7
In what follows, the state j is fixed throughout. We want to show that the probability of
avoiding a subset of the state space (a certain state j in particular) for a long time decays
exponentially with the time. Informally speaking, the idea of the proof is as follows. Find
Ki ∈ N so large such that the probability of visiting to j during the first Ki steps is positive
when X0 = i. This is possible since Xn is an irreducible Markov chain. Since the state space
is finite, the probability of visiting to j during the first K := max1≤i≤N Ki steps is bounded
away from zero by a strictly positive constant ρ ∈ (0, 1), no matter where we started. Then,
using the Markov property, argue that avoiding j during the first nK steps can be compared
to the “worst-case scenario” of getting n “failures” in a row in a sequence of n i.i.d. cointossing trials with the probability of “success” in a trial equals 1 − ρ > 0. In order to make
this argument rigorous, one needs to carefully verify that the Markov property indeed makes
the very last step of the proof working. This is not obvious straight from the definitions
because an application of the Markov property per se requires conditioning on a particular
current state. The problem is that just by indicating that a state j has not been visited for
a while, one does not specify the current location of the Markov chain. Thus a strong form
of the Markov property is required here. A sufficient for our purpose version of the Markov
property is provided below in Lemma 1.
Let Ω := {1, . . . , N } denote the state space of Xn . We will use without a further notice
the Markov property in the following form. For any integer times i, m ∈ N, sets Ak (including
the possibility of being empty sets, singletons, or the whole finite state space), and a state i,
P Xk ∈ Ak : n + 1 ≤ k ≤ n + mXn = i, Xk ∈ Ak : 0 ≤ k ≤ n − 1
= P Xk ∈ Ak : n + 1 ≤ k ≤ n + mXn = i
= P Xk ∈ Ak : 1 ≤ k ≤ mX0 = i
(1)
We will also need the following consequence of the Markov property:
Lemma 1. Let E ⊂ [0, 1] be a subset of the closed interval [0, 1]. Consider an integer m ∈ N
and a sequence of events Ak ⊂ Ω, k ≥ 0. If
P Xk ∈ Ak : 1 ≤ k ≤ mX0 = i ∈ E
(2)
for all i ∈ Ω, then
P Xk ∈ Ak : n + 1 ≤ k ≤ n + mXk ∈ Ak : 0 ≤ k ≤ n ∈ E,
for any integer n ∈ N.
1
Proof of the lemma. By virtue of the Markov property, we have:
P Xk ∈ Ak : n + 1 ≤ k ≤ n + mXk ∈ Ak : 0 ≤ k ≤ n
X
=
P Xn = iXk ∈ Ak : 0 ≤ k ≤ n
i∈Ω
×P Xk ∈ Ak : n + 1 ≤ k ≤ n + mXn = i, Xk ∈ Ak : 0 ≤ k ≤ n
X
=
P Xn = iXk ∈ Ak : 0 ≤ k ≤ n · P Xk ∈ Ak : n + 1 ≤ k ≤ n + mXn = i .
i∈Ω
This completes
P the proof of the lemma in view of the fact that condition (2) holds for any
i ∈ Ω while i∈Ω P Xn = iXk ∈ Ak : 0 ≤ k ≤ n = 1.
The following remark elucidates the result in lemma but its content is beyond the scope.
Ignoring this remark completely or partially is of absolutely no harm :) If you want to
read more about conditional expectation, you can consult, for instance, the introductory
section in the Chapter about martingales (Chapter 5 in two recent editions) of Durrett’s
text Probability Theory and Examples. We will discuss conditional expectation later in this
course while studying martingales.
Remark for an “overqualified fellow”. Using the machinery of conditional expectation,
the lemma can be proved as follows. Let Fn = σ(X0 , . . . , Xn ) be a σ-algebra of the past events
and F n = σ(Xk : k > n) be a σ-algebra of the future events. Then for any A ∈ F n , with
probability one,
P(A|Fn ) = E P(A|Xn , Fn ) = E P(A|Xn ) ,
from which the lemma follows. The first identity in the display is a general property of
conditional probabilities, the second one is the Markov property.
Back to the solution of Exercise 1.7. Since Xn is irreducible, for each state i there exist
an integer Ki ∈ N and a positive real δi > 0 such that
P(XKi = j|X0 = i) = δi .
Let K = max1≤i≤N Ki and δ = min1≤i≤N δi Then, for any i ∈ Ω
P Xm 6= j, m = 0, . . . , K X0 = i ≤ P Xm 6= j, m = 0, . . . , Ki X0 = i
≤ P XKi 6= j X0 = i = 1 − P XKi = j X0 = i = 1 − δi ≤ 1 − δ < 1.
For simplicity of notations, for an integer t ≥ 0 and a state i ∈ Ω, introduce the following
events:
At := Xm 6= j, K(t − 1) ≤ m ≤ Kt
Bt (i) := X0 = i and Xm 6= j, 0 ≤ m ≤ Kt .
2
For a real number x ∈ R let bxc denote the integer part of x, that is
bxc := max{n ∈ N : n ≤ x}.
By Lemma 1, for any n ≥ K and any state i ∈ Ω we have
P Xm 6= j, m = 0, . . . , nX0 = i ≤ P Xm 6= j, m = 0, . . . , Kbn/KcX0 = i
=P
∩ts=1 As X0
= i0
bn/Kc
Y
P At X0 = i, ∩t−1
A
= P A1 X0 = i0 ×
s
s=1
t=2
bn/Kc
≤ (1 − δ)
where C0 := (1 − δ)−
set
C1 :=
K−1
K
n
≤ C0 ρ ,
and ρ := (1 − δ)1/K . To cover the case n < K, one can for instance
max max ρ−n · P Xm 6= j, m = 0, . . . , nX0 = i ,
0≤n≤K−1 i∈Ω
and C = max{C0 , C1 }.
1.9 (a)
Then chain is irreducible. For instance, the following loop path has a positive probability
and visits at every state:
1 → 2 → 4 → 1 → 3 → 5 → 1.
1.9 (b)
Period of the chain equals to the period of any of its state. Consider, for instance, the state
5. From 5 one can go only to 1. From 1 the chain moves to either 2 or 3. From either 2 or
3 the chain can only move into {4, 5}. If it moves to 5, then 5 has been reached in 3 steps.
If it moves to 4, then in three steps (overall, in six steps from the beginning), the chain will
be again in {4, 5}. Using the induction, one can conclude that period os the state 5 is three.
Thus d = 3.
1.9 (d)
From 1 the chain moves into {2, 3}. From {2, 3} the chain can only move into {4, 5}. Finally,
from {4, 5} the chain can only move to 1. Thus
P(T = 3|X0 = 1) = 1
and hence
E(T |X0 = 1) = 3.
It follows from the law of large numbers stated on p. 25 of the textbook that
n
1
1 X
P(Xm = 1|X0 = 1) = .
π(1) = lim
n→∞ n + 1
3
m=0
1.11
3
Let Zn ∈ {0, 1 . . . , 7} be the reminder of Sn after division by 8. Clearly (verify this though!),
Zn is a Markov chain with P(Z0 = 0) = 1 and transition matrix given by


0 1/6 1/6 1/6 1/6 1/6 1/6 0
 0
0 1/6 1/6 1/6 1/6 1/6 1/6 


 1/6 0

0
1/6
1/6
1/6
1/6
1/6


 1/6 1/6 0

0
1/6
1/6
1/6
1/6

P=
 1/6 1/6 1/6 0
0 1/6 1/6 1/6 


 1/6 1/6 1/6 1/6 0

0
1/6
1/6


 1/6 1/6 1/6 1/6 1/6 0
0 1/6 
1/6 1/6 1/6 1/6 1/6 1/6 0
0
For i ∈ S, let Ti = inf{k ≥ 1 : Xk = i and ui = E(Ti |Z0 = 0). Using the result of Exercise 1.15
(stated also as formula (1.1) on p. 25), we can conclude that
E(T0 |Z0 = 0) = u0 =
1
.
π(0)
Since
π(i) =
1
for any state i,
8
(3)
we obtain that
u0 = 8.
In order to obtain (3), one can first produce a guess based on a heuristic symmetry argument
(for instance, putting the states on a circle, observing that P looks then the same from
every state, and using the interpretation of π as the vector of asymptotic frequencies) and
then make the argument rigorous by a simple checking that the proposed π is the (unique)
probability vector solving πP = π,
For i ∈ S, let vi = E(T1 |Z0 = i). It is computationally considerably more complicated to
compute without help of a calculator/ computer/ friend
E(T1 |Z0 = 0) = v0 .
Let v = (v0 , v2 , v3 , v4 , v5 , v6 , v7 ) ∈ R7 and denote e = (1, 1, 1, 1, 1, 1, 1) ∈ R7 . Then, conditioning on the first step of the Markov chain,
v = e + Qv,
where Q is the matrix obtained from P by removing the second row and second column
(corresponding to the state 1) from matrix P. Thus v = (I − Q)−1 e (we actually have
recovered the result obtained on the bottom of p. 28 in the textbook). You can use MATLAB
to find the solution of this linear system and hence to determine v0 . I believe, v0 ≈ 6.8.
4
Download