cs70 jean walrand lecture 33 review
play

CS70: Jean Walrand: Lecture 33. Review Distribution of X n n 0 . 3 - PowerPoint PPT Presentation

n m n m m X n m m m n n X n m m X n X CS70: Jean Walrand: Lecture 33. Review Distribution of X n n 0 . 3 3 0 . 7 2 0 . 2 2 0 . 4 Markov Chain: 1 1 1 3 0 . 6 0 . 8 Markov Chains 2 Finite set X ; 0 ; P = { P ( i


  1. n m n m m X n m m m n n X n m m X n X CS70: Jean Walrand: Lecture 33. Review Distribution of X n n 0 . 3 3 0 . 7 2 0 . 2 2 0 . 4 ◮ Markov Chain: 1 1 1 3 0 . 6 0 . 8 Markov Chains 2 ◮ Finite set X ; π 0 ; P = { P ( i , j ) , i , j ∈ X } ; ◮ Pr [ X 0 = i ] = π 0 ( i ) , i ∈ X + 1 Let π m ( i ) = Pr [ X m = i ] , i ∈ X . Note that ◮ Pr [ X n + 1 = j | X 0 ,..., X n = i ] = P ( i , j ) , i , j ∈ X , n ≥ 0. = ∑ 1. Review ◮ Note: Pr [ X m + 1 = j ] Pr [ X m + 1 = j , X m = i ] i Pr [ X 0 = i 0 , X 1 = i 1 ,..., X n = i n ] = π 0 ( i 0 ) P ( i 0 , i 1 ) ··· P ( i n − 1 , i n ) . 2. Distribution = ∑ Pr [ X m = i ] Pr [ X m + 1 = j | X m = i ] ◮ First Passage Time: 3. Irreducibility i = ∑ π m ( i ) P ( i , j ) . ◮ A ∩ B = / 4. Convergence 0 ; β ( i ) = E [ T A | X 0 = i ]; α ( i ) = P [ T A < T B | X 0 = i ] Hence, i ◮ β ( i ) = 1 + ∑ j P ( i , j ) β ( j ); α ( i ) = ∑ j P ( i , j ) α ( j ) . π m + 1 ( j ) = ∑ π m ( i ) P ( i , j ) , ∀ j ∈ X . i With π m , π m + 1 as a row vectors, these identities are written as π m + 1 = π m P . Thus, π 1 = π 0 P , π 2 = π 1 P = π 0 PP = π 0 P 2 ,.... Hence, π n = π 0 P n , n ≥ 0 . Distribution of X n Distribution of X n Distribution of X n n n n 0 . 3 0 . 3 3 0 . 3 3 3 0 . 7 0 . 7 2 2 0 . 2 0 . 2 2 0 . 7 2 2 2 1 0 . 4 1 1 3 1 3 1 1 1 1 1 1 0 . 6 1 3 0 . 8 0 . 8 + 1 π 0 = [0 . 5 , 0 . 1 , 0 . 4] π 0 = [0 . 2 , 0 . 3 , 0 . 5] π 0 = [0 , 1 , 0] π 0 = [1 , 0 , 0] π m (2) π m (2) π m (1) π m (2) π m (1) π m (1) π m (3) π m (3) π m (3) π m (2) π 0 = [0 . 5 , 0 . 3 , 0 . 2] π 0 = [1 , 0 , 0] π m (2) π m (2) π m (1) π m (1) π m (1) π m (3) π m (3) π m (3) As m increases, π m converges to a vector that depends on π 0 (obviously, As m increases, π m converges to a vector that does not depend on π 0 . As m increases, π m converges to a vector that does not depend on π 0 . since π m ( 1 ) = π 0 ( 1 ) , ∀ m ) .

  2. Balance Equations Balance Equations Balance Equations Question: Is there some π 0 such that π m = π 0 , ∀ m ? Theorem A distribution π 0 is invariant iff π 0 P = π 0 . These equations Definition A distribution π 0 such that π m = π 0 , ∀ m is said to be an are called the balance equations. invariant distribution. Theorem A distribution π 0 is invariant iff π 0 P = π 0 . These equations Example 1: are called the balance equations. Theorem A distribution π 0 is invariant iff π 0 P = π 0 . These equations Example 2: are called the balance equations. Proof: π n = π 0 P n , so that π n = π 0 , ∀ n iff π 0 P = π 0 . Thus, if π 0 is invariant, the distribution of X n is always the same as � 1 − a a � that of X 0 . π P = π ⇔ [ π ( 1 ) , π ( 2 )] = [ π ( 1 ) , π ( 2 )] b 1 − b Of course, this does not mean that X n does not move. It means that ⇔ π ( 1 )( 1 − a )+ π ( 2 ) b = π ( 1 ) and π ( 1 ) a + π ( 2 )( 1 − b ) = π ( 2 ) the probability that it leaves a state i is equal to the probability that it � � 1 0 ⇔ π ( 1 ) a = π ( 2 ) b . π P = π ⇔ [ π ( 1 ) , π ( 2 )] = [ π ( 1 ) , π ( 2 )] ⇔ π ( 1 ) = π ( 1 ) and π ( 2 ) = π ( 2 ) . enters state i . 0 1 These equations are redundant! We have to add an equation: The balance equations say that ∑ j π ( j ) P ( j , i ) = π ( i ) . Every distribution is invariant for this Markov chain. This is obvious, π ( 1 )+ π ( 2 ) = 1. Then we find That is, since X n = X 0 for all n . Hence, Pr [ X n = i ] = Pr [ X 0 = i ] , ∀ ( i , n ) . ∑ π ( j ) P ( j , i ) = π ( i )( 1 − P ( i , i )) = π ( i ) ∑ b a P ( i , j ) . π = [ a + b , a + b ] . j � = i j � = i Thus, Pr [ enter i ] = Pr [ leave i ] . Irreducibility Existence and uniqueness of Invariant Distribution Long Term Fraction of Time in States Definition A Markov chain is irreducible if it can go from every state i to every state j (possibly in multiple steps). Theorem A finite irreducible Markov chain has one and only Theorem Let X n be an irreducible Markov chain with invariant Examples: one invariant distribution. distribution π . That is, there is a unique positive vector π = [ π ( 1 ) ,..., π ( K )] 0 . 3 0 . 3 0 . 3 Then, for all i , such that π P = π and ∑ k π ( k ) = 1. 0 . 7 0 . 7 0 . 7 2 2 2 0 . 2 0 . 2 1 Proof: See EE126, or lecture note 24. (We will not expect you n − 1 1 ∑ 1 { X m = i } → π ( i ) , as n → ∞ . 1 1 0 . 4 1 to understand this proof.) 1 3 1 3 1 1 3 n 0 . 6 m = 0 0 . 8 0 . 8 Note: We know already that some irreducible Markov chains [B] [C] [A] have multiple invariant distributions. The left-hand side is the fraction of time that X m = i during Fact: If a Markov chain has two different invariant distributions steps 0 , 1 ,..., n − 1. Thus, this fraction of time approaches π ( i ) . π and ν , then it has infinitely many invariant distributions. [A] is not irreducible. It cannot go from (2) to (1). Proof: See EE126. Lecture note 24 gives a plausibility Indeed, p π +( 1 − p ) ν is then invariant since [B] is not irreducible. It cannot go from (2) to (1). argument. [ p π +( 1 − p ) ν ] P = p π P +( 1 − p ) ν P = p π +( 1 − p ) ν . [C] is irreducible. It can go from every i to every j . If you consider the graph with arrows when P ( i , j ) > 0, irreducible means that there is a single connected component.

  3. Long Term Fraction of Time in States Long Term Fraction of Time in States Convergence to Invariant Distribution Question: Assume that the MC is irreducible. Does π n Theorem Let X n be an irreducible Markov chain with invariant 1 n ∑ n − 1 approach the unique invariant distribution π ? distribution π . Then, for all i , m = 0 1 { X m = i } → π ( i ) , as n → ∞ . Theorem Let X n be an irreducible Markov chain with invariant n ∑ n − 1 1 distribution π . Then, for all i , m = 0 1 { X m = i } → π ( i ) , as n → ∞ . Example 2: Answer: Not necessarily. Here is an example: Example 1: Assume X 0 = 1. Then X 1 = 2 , X 2 = 1 , X 3 = 2 ,... . The fraction of time in state 1 converges to 1 / 2, which is π ( 1 ) . Thus, if π 0 = [ 1 , 0 ] , π 1 = [ 0 , 1 ] , π 2 = [ 1 , 0 ] , π 3 = [ 0 , 1 ] , etc. Hence, π n does not converge to π = [ 1 / 2 , 1 / 2 ] . Periodicity Convergence of π n Convergence of π n Theorem Assume that the MC is irreducible. Then Theorem Let X n be an irreducible and aperiodic Markov chain with Theorem Let X n be an irreducible and aperiodic Markov chain with invariant distribution π . Then, for all i ∈ X , invariant distribution π . Then, for all i ∈ X , d ( i ) := g.c.d. { n > 0 | Pr [ X n = i | X 0 = i ] > 0 } π n ( i ) → π ( i ) , as n → ∞ . π n ( i ) → π ( i ) , as n → ∞ . has the same value for all states i . Proof: See Lecture notes 24. Proof See EE126, or Lecture notes 24. Definition If d ( i ) = 1, the Markov chain is said to be aperiodic. Example Proof See EE126, or Lecture notes 24. Otherwise, it is periodic with period d ( i ) . Example Example [A]: { n > 0 | Pr [ X n = 1 | X 0 = 1 ] > 0 } = { 3 , 6 , 7 , 9 , 11 ,... } ⇒ d ( 1 ) = 1. { n > 0 | Pr [ X n = 2 | X 0 = 2 ] > 0 } = { 3 , 4 ,... } ⇒ d ( 2 ) = 1. [B]: { n > 0 | Pr [ X n = 1 | X 0 = 1 ] > 0 } = { 3 , 6 , 9 ,... } ⇒ d ( i ) = 3. { n > 0 | Pr [ X n = 5 | X 0 = 5 ] > 0 } = { 6 , 9 ,... } ⇒ d ( 5 ) = 3.

  4. Convergence of π n Calculating π Summary Theorem Let X n be an irreducible and aperiodic Markov chain with Let P be irreducible. How do we find π ? invariant distribution π . Then, for all i ∈ X ,  0 . 8 0 . 2 0  Example: P = 0 0 . 3 0 . 7  .  π n ( i ) → π ( i ) , as n → ∞ . 0 . 6 0 . 4 0 Markov Chains One has π P = π , i.e., π [ P − I ] = 0 where I is the identity matrix: Proof See EE126, or Lecture notes 24. Example  0 . 8 − 1 0 . 2 0  ◮ Markov Chain: Pr [ X n + 1 = j | X 0 ,..., X n = i ] = P ( i , j )  = [ 0 , 0 , 0 ] . π 0 0 . 3 − 1 0 . 7  ◮ FSE: β ( i ) = 1 + ∑ j P ( i , j ) β ( j ); α ( i ) = ∑ j P ( i , j ) α ( j ) . 0 . 6 0 . 4 0 − 1 ◮ π n = π 0 P n However, the sum of the columns of P − I is 0 . This shows that these equations are redundant: If all but the last one hold, so does the last one. Let ◮ π is invariant iff π P = π us replace the last equation by π 1 = 1, i.e., ∑ j π ( j ) = 1: ◮ Irreducible ⇒ one and only one invariant distribution π  0 . 8 − 1 0 . 2 1  ◮ Irreducible ⇒ fraction of time in state i approaches π ( i )  = [ 0 , 0 , 1 ] . π 0 0 . 3 − 1 1  0 . 6 0 . 4 1 ◮ Irreducible + Aperiodic ⇒ π n → π . ◮ Calculating π : One finds π = [ 0 , 0 ...., 1 ] Q − 1 where Q = ··· . Hence, − 1  0 . 8 − 1 0 . 2 1  π = [ 0 , 0 , 1 ] 0 0 . 3 − 1 1 ≈ [ 0 . 55 , 0 . 26 , 0 . 19 ]   0 . 6 0 . 4 1

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend