CSE 736: Markov Chains SUNY Buffalo, Spring 2003 Lecturer: Hung Q. Ngo Scribe: Peng Lin Lecture 6: Convergence to equilibrium, Ergodic Theorem 1 Definitions Definition 1.2 (Invariant measure and distribution). Let X be an irreducible recurrent markov chain , then is called an invariant measure with transition matrix P, if there exists some vector s.t. with respect to for . Similarly, if it holds that #%$&'($ "! We first give out the following definitions: consisting of coordinates Definition 1.1 (Measures and distributions). A vector , , is called a measure on . If the sum of all coordinates is , then it is called a distribution. , it’s called an invariant distribution. With the above mentioned properties of invariant distribution, we can come up with the following propositions: 2143 ) )5*,+ ) )(*,+ .-/0 -/6 Proposition 1.3. If is a Markov chain with and is invariant with respect to , then is also a Markov chain with for any N. It’s also called stationary, equilibrium. 78, . By applying the property recurDistribution of (1) 1 3 ) 1 ',29 1;:< =', 1>:< @?A?A?B' Proof. From the definition of invariant distribution, we have sively, it is easy to see that 2%C$ )ED F$ as$ GIQNHKJ B -MLNOP , i.e., the probability of ending up at N is an invariant distribution. F F Proof. We want to show that F $R "! F #%$ T E) UWV %C$ )ED F $ , we have , Since by definition, S F 9#Y$ S T )EUWV [ C )"D M\%$ X! Z"! S T )EUWV X! [ C )"D M\%$ S T )EUWV [ C $) 3#< D F$ Proposition 1.4. Suppose tends to independent on the starting state, then 1 is ] _` ] ^ 1 _` ^ 2 Figure 1: A two-state Markov chain Example 1.5 (convergency). If a Markov chain is convergent, it must have an invariant distribution. Let’s consider a two-state Markov chain as illustrated in Figure 1. The transition matrix ba dcIf e e fhg dc f f i e , we have Ignoring the case when e ) kj m 3l l m 3mm l m 3l l m 3 l'n when GoHJ . f From 1.3, we can see that the row components form an invariant distribution. If is larger, the final f p state will most likely end up at state 1. However if e , then will end up either at 1 or 2, and the Markov chain will not converge. Next, we will answer the question “when does a Markov chain have an invariant distribution”. Theorem 1.6. If is irreducible, then the following claims are equivalent: 1. All states are positive recurrent. 2. Some state is positive recurrent. 3. F has an invariant distribution . q q r [ t s [.uX v : < Qy{z}| w ~ )Ewx+ Before showing the proof for the theorem, we first give out two lemmas. For a certain state , we can find the expected time spent in state before visiting : where [ q is the first passage time to . C[ D . 1. r [ C [ D r C [ D . 2. r r C [ D J -ML o . 3. Lemma 1.8. If is irreducible, r C[ D. recurrent, then Lemma 1.7. If is irreducible and recurrent, then is invariant measure with [ , then I r C [ D . Moreover, if With the lemmas given above, we can show the proofs for the theorem as follows: 2 is M M . This is obvious. Md M . Suppose is positive recurrent, then is recurrent. Since is positive recurrent, the 2. 8 $X! r $ C D J . Take F E Y , it’s expected total number of steps before visits to state : . Z"! F . From lemma1.7, F is an invariant distribution. easy to see M4K M . Take any state N , F $R X! F 9 YC$ )"D -ML G . And since F is a distribution, X! F 3. is irreducible, i.e., q -& F [ -/ [ $2 , and from the definition of F $ , we have F $ [ , L .N .Since o E -5 -AAA , then is an invariant measure with [ . Hence, we have r C D Let by Lemma1.8. It v follows v that: [ $X! r $ C [ D $A"! FF $[ F [ J . And since F $ -MLN , we are done. That is, q is positive recurrent as long as F [ Proof. 1. Next, we will look at the problem when a markov chain will converge. First, we show that periodic is not a good property: Consider the example Markov chain given in W¡2 a £¢2 a ¤ a g. figure 1, g . g . Hence the Markov chain has a period of 2 and will not converge. Theorem 1.9. If is irreducible, aperiodic and has an invariant distribution F , then ¥§¦§¨ 2© ªN"«x F $ )EUWV ) In other words, S ¬T )"UWV %C$ )"D F $ ­ , define ® period of i p¯(° ®±G -/ ²C )"D B³ . If ® =´ G , ( ´ means not divisible Given , ²C )ED i by, means divisible), then If ® , we say is aperiodic. N \ ® $ Proposition 1.10. 1. If , are in same class, then ® -RQ ²C )ED -ML G G.µ . 2. If is aperiodic, then BG#µ -¶T 7 YC$ )"D -/ $¹C¸ · D . ²C ) 3 · D º ® R G­» T . Proof. 1. Since is irreducible, BG ) 3 · 3x¼ if ${¼ $ £ ® $6X½R- ® $0 G¾» T= ® $0 ® ±G 5 ²C )ED B³ then ® .t¯(° ® ¿ 4 . Note: 2. Let ¿ -¶Á¾ ¿ , then Â(ÀûªÄ Á ¿ ,L  - Ä Å 3 . (a) If À - -AAAQ- À · ¿ , ¯(° ® À < - À ¡ -AAA- À · W . Take À < ¿ and À < ÆW< Ç AAÈ [ Ç , Let (b) BÀ < À ¡ v ¿ < ±G ¿ 5 < ´ G ³2pÉ ÊB-AAAQ- ¿ [ ±G ¿ 5 [ ´ G ³'É Ê -AAAQ- 'ÅË ?A?A? -AAAQ- ,  [ 3#< -AAAQ-  · Æ . Define 3.  < Ä Â < À <  » · ?A?A? »Ì [  À < [ À Í< » cR [ 3#»t<  À · [ À3#· < » ?A?A ? . » Say cR  · < À · »p [ . Obviously, - ÄcÎ ¿ , we can Ä ¿ - ÏEÄ>ct ¿ - ÏEÄdcºÏ ¿ . In general, we have qÄdcªq to qÄ ¿ . If q is larger infer that ÏEÄ Ä ), all consecutive numbers are covered in the range. enough (i.e., q 3 .-/ -/ i G\) Ñ#be±G Markov( ¾) Ð),) Ð ³ ). be Markov(F ), and they are indeÅ À) -¶Á ) , Å ) )(*,+ is Markov( F -Ò ) with state space ¾´o . We have 1. Define ) 2©%Å + Í /-¬N(¬«.i2© À+ i Ó«2© Á + ªNE«x' F $ Ò C Ô $ D C [ Ô Õ D t [ $¹Õ ×ÖEØ Ä © J «x . We want to show Ò is irreducible and positive recurrent. (proof omitted.) 2. Ö"Ø Ä © [[ J « -ML q o Ù ÖEØ Ä © J «x 3. From Lemma1.7, we know Ù -¶Á -¶Á 4. To show that À) ) have the “same” distribution after T. Note that À.) ) have the same transition Now we show the proof to theorem 1.9. Proof. We use a coupling argument. Let pendent. Let T be a reference point: matrix . We have 2© À) t ¹- G «k ) A$ "! 2© À) t ¹- À · ªNE- tT « ·w< ) 2© À · ªNÚ- tT «¾© À,) i d À · ª N"« · w < $A"! ) $A"! 2© Á · ªNE- tT «2© Á ) t XÁ · Î"N « ·¾w © Á < ) t /- G « 2© À ) i Ó«Û 2© À ) t ¹- G « » ¾© À ) i /- G « 2© Á ) t /- G « » 2© À) t /- G « 2© Á ) t Ó« » 2© À ) t /- G « (a) (b) ¾© Á ) Ü Ó« 2© À ) t Ó« » 2© Á ) t ¹- G « 2© À,) t Ó« c 2© Á ) i Ó«M 2© À) t ¹- G « » ¾ © Á ) t /- G « Based on the same arguments, we have Hence, Summing up on both sides, 2© À ) t « c 2 © Á ) i Ó«M Ï 2 © G « X! Taking limits as n tends to infinity, the second term on the left tends to right hand side tends to 0. Hence, we have ¥§¦§¨ 2 © t Ó«Ý F )EUWV ,À ) 4 F , the term on the where Þ ß G =' )· :wx< + Q y{áâ w ~ . O) 2©Þ ß G H ² as GHàJ «x G Theorem 1.11 (Ergodic Theorem). Let be irreducible and let 5 be Markov( .-/ ), then