Randomized Algorithms Randomized Algorithms Markov Chains and - - PowerPoint PPT Presentation

randomized algorithms randomized algorithms
SMART_READER_LITE
LIVE PREVIEW

Randomized Algorithms Randomized Algorithms Markov Chains and - - PowerPoint PPT Presentation

Randomized Algorithms Randomized Algorithms Markov Chains and Random Walks Markov Chains and Random Walks Speaker: Chuang-Chieh Lin Advisor: Professor Maw-Shang Chang National Chung Cheng University 2007/4/25 2007/4/25 References


slide-1
SLIDE 1

2007/4/25 2007/4/25

Randomized Algorithms Randomized Algorithms

Markov Chains and Random Walks Markov Chains and Random Walks

Speaker: Chuang-Chieh Lin Advisor: Professor Maw-Shang Chang National Chung Cheng University

slide-2
SLIDE 2

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 2 2

References References

Professor S. C. Tsai’s slides. Randomized Algorithms, Rajeev Motwani and

Prabhakar Raghavan.

Probability and Computing - Randomized

Algorithms and Probabilistic Analysis, Michael Mitzenmacher and Eli Upfal.

Wikipedia: Markov Chains

slide-3
SLIDE 3

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 3 3

Outline Outline

Introduction to Markov chains Classification of states Stationary distribution Random walks on undirected graphs Connectivity problem

slide-4
SLIDE 4

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 4 4

I ntroduction to Markov I ntroduction to Markov Chains Chains

Markov chains provide a simple but powerful

framework for modeling random processes.

Markov chains can be used to analyze simple

randomized algorithms applying random walks.

slide-5
SLIDE 5

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 5 5

I ntroduction to Markov I ntroduction to Markov Chains (cont Chains (cont’ ’d) d)

Definition: F A stochastic process X = {X(t), t ∈ T} is a collection of random variables. F If T is a countable set, say T = {0, 1, 2, . . .}, we say that X is a discrete time stochastic process. F Otherwise it is called continuous time sto- chastic process. F Here we consider a discrete time stochastic process Xn, for n = 0, 1, 2, . . ..

slide-6
SLIDE 6

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 6 6

I ntroduction to Markov I ntroduction to Markov Chains (cont Chains (cont’ ’d) d)

F If Xn = i, then the process is said to be in state i at time n. F Denote Pr[Xn+1 = j | Xn = i, Xn−1 = in−1, . . . , X0 = i0] = Pi,j for all states i0, i1, . . . , in−1, i, j and all n ≥ 0. F Xn+1 depends only on Xn.

slide-7
SLIDE 7

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 7 7

I ntroduction to Markov I ntroduction to Markov Chains (cont Chains (cont’ ’d) d)

That is, Pi,j = Pr[Xn+1 = j | Xn = i ], for all states i, j and all n ≥ 0 Such a stochastic process is known as a Markov chain.

slide-8
SLIDE 8

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 8 8

Formal definitions.

slide-9
SLIDE 9

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 9 9

Markov property Markov property

In probability theory, a stochastic process has the Markov

property if the conditional probability distribution of future states of the process, given the present state and all past states, depends only upon the current state and not on any past states.

Mathematically, if X(t), t > 0, is a stochastic process, the

Markov property states that

Pr[X(t + h) = y | X(s) = x(s), ∀s ≤ t] = Pr[X(t + h) = y | X(s) = x(t)], ∀h > 0

slide-10
SLIDE 10

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 10 10

Markov chain Markov chain

In mathematics, a

Markov chain, named after Andrey Markov, is a discrete-time stochastic process with the Markov property.

June 14, 1856 – July 20, 1922

slide-11
SLIDE 11

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 11 11

Homogeneous Homogeneous

Markov processes are typically termed

(time-) homogeneous if

Pr[X(t + h) = y | X(t) = x(t)] = Pr[X(h) = y | X(0) = x(t)], ∀t, h > 0

slide-12
SLIDE 12

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 12 12

Transition matrix Transition matrix

i 1 2 j Pi,1 Pi,2 Pi,j

. . .

⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎦ ⎤ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎣ ⎡ = O M O M M L L O M O M M L L L L

j i i i j j

P P P P P P P P P

, 1 , , , 1 1 , 1 , 1 , 1 , ,

P transition matrix F Pi,j ≥ 0. F P

j

Pi,j = 1.

slide-13
SLIDE 13

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 13 13

Transition probability Transition probability

The m-step transition probability P m

i,j of the Markov

chain is defined as the conditional probability, given that the chain is currently in state i, that will be in state j after m additional transitions. That is, P m

i,j = Pr[Xn+m = j | Xn = i], for m ≥ 0, i, j ≥ 0.

Conditioning on the first transition from i, we have the following equation: P m

i,j =

X

k≥0

Pi,kP m−1

k,j

.

slide-14
SLIDE 14

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 14 14

Chapman Chapman-

  • Kolmogorov

Kolmogorov equation equation

Generalize the previous result, we have

Chapman-Kolmogorov equation as follows.

P n+m

i,j

= X

k≥0

P n

i,kP m k,j.

slide-15
SLIDE 15

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 15 15

Chapman Chapman-

  • Kolmogorov

Kolmogorov equation equation (cont (cont’ ’d) d)

¥

Proof: By the definition of the n-step transition probability, P n+m

i,j

= X

k≥0

Pr[Xn+m = j, Xn = k | X0 = i] = X

k≥0

Pr[Xn = k | X0 = i] · Pr[Xn+m = j | Xn = k, X0 = i] By the Markov property, Pr[Xn+m = j | Xn = k, X0 = i] = Pr[Xn+m = j | Xn = k] = P m

k,j. With the additional

  • bservation that Pr[Xn = k | X0 = i] = P n

i,k, the theorem

immediately follows.

slide-16
SLIDE 16

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 16 16

Recall: Transition matrix Recall: Transition matrix

i 1 2 j Pi,1 Pi,2 Pi,j

. . .

⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎦ ⎤ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎣ ⎡ = O M O M M L L O M O M M L L L L

j i i i j j

P P P P P P P P P

, 1 , , , 1 1 , 1 , 1 , 1 , ,

P transition matrix F Pi,j ≥ 0. F P

j

Pi,j = 1.

slide-17
SLIDE 17

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 17 17

Recall: Homogeneous Recall: Homogeneous

Markov processes are typically termed

(time-) homogeneous if

Pr[X(t + h) = y | X(t) = x(t)] = Pr[X(h) = y | X(0) = x(t)], ∀t, h > 0

slide-18
SLIDE 18

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 18 18

F Let P(n) denote the matrix of n-step tran- sition probabilities P n

i,j, then the Chapman-

Kolmogorov equations implies that P(n) = Pn.

For example,

slide-19
SLIDE 19

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 19 19

2 3 1

0.3 0.4 0.5 0.6 0.8 0.2 0.5 0.7

⎥ ⎥ ⎥ ⎥ ⎦ ⎤ ⎢ ⎢ ⎢ ⎢ ⎣ ⎡ = 8 . 2 . 6 . 4 . 5 . 5 . 3 . 7 . P

0 1 2 3 1 2 3

⎥ ⎥ ⎥ ⎥ ⎦ ⎤ ⎢ ⎢ ⎢ ⎢ ⎣ ⎡ = 64 . 1 . 16 . 1 . 48 . 2 . 12 . 2 . 3 . 15 . 2 . 35 . 18 . 21 . 12 . 49 .

2

P

0 1 2 3 1 2 3

slide-20
SLIDE 20

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 20 20

Classification of states Classification of states

A first step in analyzing the long-term

behavior of a Markov chain is to classify its states.

In the case of a finite Markov chain, this is

equivalent to analyzing the connected connectivity structure of the directed graph representing the Markov chain.

slide-21
SLIDE 21

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 21 21

Basic definitions Basic definitions

F State j is said to be accessible from state i if P n

i,j > 0 for some n ≥ 0.

F We say states i and j communicate if they are both accessible from each other. (i ↔ j) F The Markov chain is said to be irreducible if all states communicate with each other.

slide-22
SLIDE 22

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 22 22

Basic definitions (cont Basic definitions (cont’ ’d) d)

F Let rt

i,j denote the probability that starting at

state i, the first transition to state j occurs at time t. That is, rt

i,j = Pr[Xt = j and, for 1 ≤ s ≤ t − 1, Xs 6=

j | X0 = i]. F State i is said to be recurrent if P

t≥1 rt i,i = 1,

and transient if P

t≥1 rt i,i < 1.

slide-23
SLIDE 23

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 23 23

Basic definitions (cont Basic definitions (cont’ ’d) d)

F A state j in a discrete time Markov chain is periodic if there exists an integer ∆ > 1 such that Pr[Xt+s = j | Xt = j] = 0 for some integer t ≥ 0 unless s is divisible by ∆. F A state i has period d if d = gcd{n | P n

i,i > 0},

where gcd means the greatest common divisor. F A discrete time Markov chain is periodic if there exists at least one periodic state in the chain.

slide-24
SLIDE 24

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 24 24

Basic definitions (cont Basic definitions (cont’ ’d) d)

F A state with period 1 is said to be aperiodic. F We denote by hi,j the expected time from state i to state j. So we have hi,j = P

t≥1 t · rt i,j.

F A recurrent state i is said to be positive recur- rent, if hi,i < ∞. Otherwise it is null recurrent. F Positive recurrent, aperiodic states are called ergodic.

slide-25
SLIDE 25

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 25 25

Null recurrent? Null recurrent?

For example, consider a Markov chain whose

states are the positive integers.

From state i, the probability of going to state i+1

is i/(i+1).

With probability 1/(i+1), the chain returns to

state 1.

slide-26
SLIDE 26

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 26 26

Null recurrent? (cont Null recurrent? (cont’ ’d) d)

Starting at state 1, the probability of not having

returned to state 1 within the first t steps is thus

Hence the probability of never returning to state 1

from state 1 is 0, then we have state 1 is recurrent. It follows that

rt

1,1 = 1 t · 1 t+1 = 1 t(t+1).

t

Q

j=1 j j+1 = 1 t+1.

slide-27
SLIDE 27

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 27 27

Null recurrent? (cont Null recurrent? (cont’ ’d) d)

However, the expected number of steps until the

first return to state 1 from state 1 is which is unbounded.

Thus this Markov chain has null recurrent states.

h1,1 =

P

t=1

t · rt

1,1 = ∞

P

t=1 1 t+1,

slide-28
SLIDE 28

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 28 28

In the foregoing example, the Markov

chain had an infinite number of states.

This is necessary for null recurrent states

to exist.

Yet for a finite Markov chain, we have the

following lemma.

slide-29
SLIDE 29

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 29 29

Lemma 1 Lemma 1

In a finite Markov chain:

– At least one state is recurrent; and – All recurrent states are positive recurrent.

We omit the proof here, though it is not

hard.

slide-30
SLIDE 30

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 30 30

Recall that Recall that… …

F State i is said to be recurrent if P

t≥1 rt i,i = 1,

and transient if P

t≥1 rt i,i < 1.

slide-31
SLIDE 31

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 31 31

Proposition 1 Proposition 1

Proof of this proposition is a little bit complicated, so we

  • mit it here.

F State i is recurrent if P

n≥0 1 · P n i,i = ∞.

I That is, the expected number of visits to state i over all time is infinite.

F State i is transient if P

n≥0 1 · P n i,i < ∞.

I That is, the expected number of visits to state i over all time is finite.

slide-32
SLIDE 32

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 32 32

Proof of the second statement Proof of the second statement

F Let Ni be the number of visits to state i over all time, then E[Ni] = P

n≥0 Pi,i = 0.

F Given an initial state distribution, let Vi denote the event that the system eventually goes to state i.

I Obviously, Pr[Vi] ≤ 1.

F If Vi does not occur, then Ni = 0.

I This implies that E[Ni | V c

i ] = 0.

slide-33
SLIDE 33

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 33 33

Proof of the second statement Proof of the second statement (cont (cont’ ’d) d)

F Otherwise (i.e., Vi occurs), there exists a time t when the system first enters state i. F In this case, given that the state is i, let Vii denote the event that the system eventually returns to state i.

I Thus V c

ii is the event that the system never re-

turns to state i.

slide-34
SLIDE 34

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 34 34

Proof of the second statement Proof of the second statement (cont (cont’ ’d) d)

F Since i is transient, there exists a state, say j, such that for some t0, P t0

i,j > 0 but i is not

accessible from j.

I Thus if we enter state j at time t0, the event V c

ii

will occur.

F Since this is one possible way that V c

ii can occur,

Pr[V c

ii] ≥ P t0 i,j > 0.

slide-35
SLIDE 35

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 35 35

Proof of the second statement Proof of the second statement (cont (cont’ ’d) d)

F After each return to i, there is a probability Pr[V c

ii] > 0 that state i will never be reentered.

F Hence, given Vi, the expected number of visits to i is geometric with conditional expected value E[Ni | Vi] = 1/Pr[V c

ii] ≤ 1/P t0 i,j.

F Finally we have E[Ni] = E[Ni | V c

i ] · Pr[V c i ] + E[Ni | Vi] · Pr[Vi]

= E[Ni | Vi] · Pr[Vi] < ∞.

slide-36
SLIDE 36

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 36 36

Corollary 1 Corollary 1

If state i is recurrent, and state i communicates

with state j, then state j is recurrent.

Proof:

– Exercise.

slide-37
SLIDE 37

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 37 37

Stationary Distribution Stationary Distribution

Definition: A stationary distribution (also called an equilibrium distribution) of a Markov chain is a probability distribution ¯ π such that ¯ π = ¯ πP. Recall: P is the one-step transition probability matrix of a Markov chain.

slide-38
SLIDE 38

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 38 38

Computing the stationary Computing the stationary distribution of a finite Markov chain distribution of a finite Markov chain

One way to compute the stationary distribution

  • f a finite Markov chain is to solve the system of

linear equations

This is particularly useful if one is given a

specific chain.

¯ π = ¯ πP.

slide-39
SLIDE 39

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 39 39

For example, given the transition matrix

we have five equations for the four unknowns π0, π1, π2, and π3 given by and

, 4 / 1 4 / 1 2 / 1 2 / 1 4 / 1 4 / 1 6 / 1 3 / 1 2 / 1 4 / 3 4 / 1 ⎥ ⎥ ⎥ ⎥ ⎦ ⎤ ⎢ ⎢ ⎢ ⎢ ⎣ ⎡ = P

¯ π = ¯ πP

P3

i=0 πi = 1.

slide-40
SLIDE 40

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 40 40

Another technique Another technique

Another useful technique is to study the cut-sets

  • f the Markov chain.

For any state i of the chain,

  • r

n

P

j=0

πjPj,i = πi = πi

n

P

j=0

Pi,j

n

P

j6=i

πjPj,i =

n

P

j=0

πiPi,j.

slide-41
SLIDE 41

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 41 41

That is, in the stationary distribution the

probability that a chain leaves a state equals the probability that it enters the state.

This observation can be generalized to sets of

states as follows.

slide-42
SLIDE 42

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 42 42

Theorem Theorem

Let S be a set of states of a finite, irreducible,

aperiodic Markov chain. In the stationary distribution, the probability that the chain leaves the set S equals the probability that it enters S.

slide-43
SLIDE 43

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 43 43

In other words, if C is a cut-set in the graph

representation of the Markov chain, then in the stationary distribution the probability of crossing the cut-set in one direction is equal to the probability of crossing the cut-set in the other direction.

slide-44
SLIDE 44

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 44 44

That is, in the stationary distribution the

probability that a chain leaves a state equals the probability that it enters the state.

slide-45
SLIDE 45

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 45 45

The transition matrix is

. 1 1 ⎥ ⎦ ⎤ ⎢ ⎣ ⎡ − − = q q p p P

1

1 − q 1 − p p q

For example, For example,

slide-46
SLIDE 46

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 46 46

This Markov chain is often used to represent

bursty behavior.

For example, when bits are corrupted in

transmissions they are often corrupted in large blocks, since errors are often caused by an external phenomenon of some duration.

1

1 − q 1 − p p q

slide-47
SLIDE 47

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 47 47

In this setting, begin in state 0 after t steps

represents that the tth bit was sent successfully, while being in state 1 represents that the bit was corrupted.

1

1 − q 1 − p p q

slide-48
SLIDE 48

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 48 48

Blocks of successfully sent bits and corrupted

bits both have lengths that follow a geometric distribution.

When p and q are small, state changes are rare,

and the bursty behavior is modeled.

1

1 − q 1 − p p q

slide-49
SLIDE 49

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 49 49

Solving corresponds to solving the

following system of three equations:

¯ π = ¯ πP

π0(1 − p) + π1q = π0; π0p + π1(1 − q) = π1; π0 + π1 = 1.

1

1 − q 1 − p p q

slide-50
SLIDE 50

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 50 50

Using the cut-set formulation,

we have that in the stationary distribution the probability of leaving state 0 must equal the probability of entering state 0.

π0p = π1q.

slide-51
SLIDE 51

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 51 51

Again, now using π0+π1 = 1 yields π0 =

q/(p+q) and π1 = p/(p+q).

For example, with the natural parameters p =

0.005 and q = 0.1, in the stationary distribution more than 95% of the bits are received successfully.

slide-52
SLIDE 52

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 52 52

Theorem 1 Theorem 1

Any finite, irreducible, and ergodic Markov

chain has the following properties:

F The chain has a unique stationary distribution ¯ π = (π0, π1, . . . , πn); F For all j and i, lim

t→∞ P t j,i exists and it is indepen-

dent of j; F πi = lim

t→∞ P t j,i = 1 hi,i .

positive recurrent or aperiodic

slide-53
SLIDE 53

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 53 53

Proof of Theorem 1 Proof of Theorem 1

Please refer to the textbook for the details.

– page 168−170 in [MU05]

We omit the proof here. Yet we can still eye on the following

lemma.

slide-54
SLIDE 54

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 54 54

Lemma 2 Lemma 2

For any irreducible, ergodic Markov chain

and for any state i,

We explain instead of proving the lemma as

follows.

lim

t→∞P t i,i exists and lim t→∞P t i,i = 1 hi,i .

slide-55
SLIDE 55

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 55 55

I nformal justifications I nformal justifications… …

The expected time between visits to i is hi,i and

therefore state i is visited 1/hi,i of the time.

Thus the limit of , which represents the

probability that, a state chosen far in the future is at state i when the chain starts at state i, must be 1/hi,i.

Since the limit exists, we can show that

P t

i,i

l li im m

t t→ →∞ ∞ P

P t

t j j, ,i i =

= l li im m

t t→ →∞ ∞ P

P t

t i i, ,i i =

=

1 1 h hi

i, ,i i .

.

We omit the detail here for simplicity.

slide-56
SLIDE 56

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 56 56

Markov Chains and Random Walks on Undirected Graphs…

slide-57
SLIDE 57

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 57 57

Random walks on undirected Random walks on undirected graphs graphs

A random walk

random walk on G is a Markov chain defined by the sequence of moves of a particle between vertices of G.

In this process, the place of the particle at a given

time step is the state of the system.

If the particle is at vertex i and if i has d(i)

  • utgoing edges, then the probability that the

particle follows edge (i, j) and moves to neighbor j is 1/d(i).

slide-58
SLIDE 58

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 58 58

A random walk on an undirected graph G

is aperiodic iff G is not bipartite.

Lemma 3 Lemma 3

slide-59
SLIDE 59

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 59 59

A graph is bipartite iff it does not have any odd

cycle.

In an undirected graph, there is always a path of

length 2 from a vertex to itself.

Thus, if the graph is bipartite then the random

walk is periodic (d = 2).

If not bipartite, then it has an odd cycle and

gcd(2,odd-number) = 1. Thus, the Markov chain is aperiodic.

Proof of Lemma 3 Proof of Lemma 3

⇒ ⇐

slide-60
SLIDE 60

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 60 60

Some technical restrictions Some technical restrictions

Due to some technical reasons, for the

remainder of our discussion, we assume that the graph G that we will discuss is not bipartite.

slide-61
SLIDE 61

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 61 61

Something needs to be Something needs to be clarified clarified… …

A random walk on a finite, undirected, connected,

and non-bipartite graph G satisfies the conditions

  • f Theorem 1.

– A Markov chain that is finite, irreducible, and ergodic.

Hence the random walk converges to a stationary

  • distribution. (Due to Theorem 1.)

The following Theorem shows that this distribution

depends only on the degree sequence of the graph.

aperiodic ergodic

slide-62
SLIDE 62

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 62 62

Theorem 2 Theorem 2

A random walk on G converges to a stationary distribution ¯ π, where πv = d(v) 2|E|.

slide-63
SLIDE 63

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 63 63

Proof of Theorem 2 Proof of Theorem 2

Since P

v∈V d(v) = 2|E|, P v∈V πv = P v∈V d(v) 2|E| = 1.

That is, ¯ π is indeed a distribution. Let P be the transition probability matrix and N(v) be the neighbors of v.

(¯ πP)v = P

u∈N(v) d(u) 2|E| · 1 d(u) = d(v) 2|E|.

thus the theorem follows. ¥

slide-64
SLIDE 64

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 64 64

Corollary 2 Corollary 2

For any vertex u in G, hu,u = 2|E|

d(u).

Recall that hv,u is the expected number of steps to reach u from v. So the corollary immediately follows.

slide-65
SLIDE 65

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 65 65

Lemma 4 Lemma 4

If (u, v) ∈ E, then hv,u ≤ 2|E|.

slide-66
SLIDE 66

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 66 66

Proof of Lemma 4 Proof of Lemma 4

u

. . .

w’s

Recall that N(u) denotes the neighbors of u in the given graph G. Since 2|E|

d(u) = hu,u = 1 d(u) ·

P

w∈N(u)

(1 + hw,u), we have 2|E| = P

w∈N(u)

(1 + hw,u). Therefore hv,u < 2|E| (since (u, v) ∈ E.) ¥

slide-67
SLIDE 67

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 67 67

The Covering Time The Covering Time

Definition:

– The cover time of a graph G = (V, E) is the maximum over all v ∈ V of the expected time to visit all of the nodes in the graph by a random walk starting from v.

slide-68
SLIDE 68

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 68 68

Lemma 5 Lemma 5

The cover time of G = (V, E) is bounded by 4|V| ⋅ |E|.

slide-69
SLIDE 69

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 69 69

Proof of Lemma 5 Proof of Lemma 5

Choose a spanning tree of G. Then there exists a

cyclic (Eulerian) tour on this tree, where each edge is traversed once in each direction, which can be found by doing a DFS.

Let v0,v1,…,v2|V|−2= v0 be the sequence of vertices

in the tour, starting from v0.

slide-70
SLIDE 70

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 70 70

Proof of Lemma 5 Proof of Lemma 5 (cont (cont’ ’d) d)

Clearly the expected time to go through the

vertices in the tour is an upper bound on the cover time.

Hence the cover time is bounded above by ¥

2|V |−3

X

i=0

hvi,vi+1 < (2|V | − 2)(2|E|) < 4|V | · |E|.

from Lemma 4

slide-71
SLIDE 71

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 71 71

Suppose we are given an undirected graph

G (V, E) and two vertices s s and t t in G.

– Let n = |V| and m = |E|.

We want to determine if there is a path

connecting s s and t t.

Application: Application: s s -

  • t

t connectivity connectivity

slide-72
SLIDE 72

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 72 72

Application: Application: s s -

  • t

t connectivity connectivity (cont (cont’ ’d) d)

This is easily done in linear time using a

standard breadth-first search or depth-first search.

However, such algorithms require Ω(n)

space.

The following randomized algorithm

works only O(log n) bits of memory.

slide-73
SLIDE 73

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 73 73

Application: Application: s s -

  • t

t connectivity connectivity (cont (cont’ ’d) d)

s-t Connectivity Algorithm: Input: G and two vertices s and t Output: Yes or No F Start a random walk from s. F If the walk reaches t within 4n3 steps, return YES. Otherwise, return No.

slide-74
SLIDE 74

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 74 74

Application: Application: s s -

  • t

t connectivity connectivity (cont (cont’ ’d) d)

We have assumed that G has no bipartite

connected component.

– The results can be made to apply to bipartite graphs with some additional technical work.

Let us consider the following theorem.

slide-75
SLIDE 75

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 75 75

Theorem 3 Theorem 3

The s-t Connectivity Algorithm returns the cor- rect answer with probability 1

2 and it only errs by

returning that there is no path from s to t when there is such a path.

slide-76
SLIDE 76

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 76 76

Proof of Theorem 3 Proof of Theorem 3

The algorithm gives correct answer, when G has

no s-t path.

If G has an s-t path, the algorithm errs if it does

not find the path in 4n3 steps.

Now we consider the case that G has an s-t path.

slide-77
SLIDE 77

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 77 77

Proof of Theorem 3 (cont Proof of Theorem 3 (cont’ ’d) d)

The expected time to reach t from s is bounded by

the cover time, which is at most 4|V| ⋅ |E| < 2n3.

– |E| ≤ n(n − 1)/2

Let a random variable X denote the number of

steps needed for the s-t Connectivity Algorithm.

By Markov’s inequality, ¥

Pr[X > 4n3] ≤ Pr[X ≥ 4n3] ≤ E[X] 4n3 < 2n3 4n3 = 1 2.

slide-78
SLIDE 78

Computation Theory Lab, CSIE, CCU, Taiwan Computation Theory Lab, CSIE, CCU, Taiwan 2007/4/25 2007/4/25 78 78

The algorithm must keep track of its current

position, which takes O(log n) bits, as well as the number of steps taken in the random walk, which also takes only O(log n) bits.

– Since we count up only 4n3, which requires log(4n3) = O(log n) bits

Proof of Theorem 3 (cont Proof of Theorem 3 (cont’ ’d) d)

slide-79
SLIDE 79

Thank you.