SLIDE 1
Martingales
Definition A sequence of random variables Z0, Z1, . . . is a martingale with respect to the sequence X0, X1, . . . if for all n ≥ 0 the following hold:
1 Zn is a function of X0, X1, . . . , Xn; 2 E[|Zn|] < ∞; 3 E[Zn+1|X0, X1, . . . , Xn] = Zn;
Definition A sequence of random variables Z0, Z1, . . . is a martingale when it is a martingale with respect to itself, that is
1 E[|Zn|] < ∞; 2 E[Zn+1|Z0, Z1, . . . , Zn] = Zn;
SLIDE 2 Conditional Expectation
Definition E[Y | Z = z] =
y Pr(Y = y | Z = z) , where the summation is over all y in the range of Y . Lemma For any random variables X and Y , E[X] = EY [EX[X | Y ]] =
Pr(Y = y)E[X | Y = y] , where the sum is over all values in the range of Y .
SLIDE 3 Lemma For any random variables X and Y , E[X] = EY [EX[X | Y ]] =
Pr(Y = y)E[X | Y = y] , where the sum is over all values in the range of Y . Proof.
Pr(Y = y)E[X | Y = y] =
Pr(Y = y)
x Pr(X = x | Y = y) =
x Pr(X = x | Y = y) Pr(Y = y) =
x Pr(X = x ∩ Y = y) =
x Pr(X = x) = E[X].
SLIDE 4 Example
Consider a two phase game:
- Phase I: roll one die. Let X be the outcome.
- Phase II: Flip X fair coins, let Y be the number of HEADs.
- You receive a dollar for each HEAD.
Y is distributed B(X, 1
2),
E[Y | X = a] = a 2 E[Y ] =
6
E[Y | X = i] Pr(X = i) =
6
i 2 Pr(X = i) = 7 4
SLIDE 5 Conditional Expectation as a Random variable
Definition The expression E[Y | Z] is a random variable f (Z) that takes on the value E[Y | Z = z] when Z = z. Consider the outcome of rolling two dice X1, X2, X = X1 + X2. E[X | X1] =
x Pr(X = x | X1) =
X1+6
x · 1 6 = X1 + 7 2 . Consider the two phase game E[Y | X] = X 2
SLIDE 6 If E[Y | Z] is a random variable, it has an expectation. Theorem E[Y ] = E[E[Y | Z]] . E[X | X1] = X1 + 7 2 . Thus E[E[X | X1]] = E
2
2 + 7 2 = 7 .
SLIDE 7
Martingales
Definition A sequence of random variables Z0, Z1, . . . is a martingale with respect to the sequence X0, X1, . . . if for all n ≥ 0 the following hold:
1 Zn is a function of X0, X1, . . . , Xn; 2 E[|Zn|] < ∞; 3 E[Zn+1|X0, X1, . . . , Xn] = Zn;
Definition A sequence of random variables Z0, Z1, . . . is a martingale when it is a martingale with respect to itself, that is
1 E[|Zn|] < ∞; 2 E[Zn+1|Z0, Z1, . . . , Zn] = Zn;
SLIDE 8
Martingale Example
A series of fair games (E[gain] = 0), not necessarily independent.. Game 1: bet $1. Game i > 1: bet 2i if won in round i − 1; bet i otherwise. Xi = amount won in ith game. (Xi < 0 if ith game lost). Zi = total winnings at end of ith game.
SLIDE 9
Example
Xi = amount won in ith game. (Xi < 0 if ith game lost). Zi = total winnings at end of ith game. Z1, Z2, . . . is martingale with respect to X1, X2, . . . E[Xi] = 0. E[Zi] = E[Xi] = 0 < ∞. E[Zi+1|X1, X2, . . . , Xi] = Zi + E[Xi+1] = Zi.
SLIDE 10
Gambling Strategies
I play series of fair games (win with probability 1/2). Game 1: bet $1. Game i > 1: bet 2i if I won in round i − 1; bet i otherwise. Xi = amount won in ith game. (Xi < 0 if ith game lost). Zi = total winnings at end of ith game. Assume that (before starting to play) I decide to quit after k games: what are my expected winnings?
SLIDE 11
Lemma If Z0, Z1, . . . , Zn is a martingale with respect to X0, X1, . . . , Xn, then E[Zn] = E[Z0] . Proof. Since Zi defines a martingale Zi = E[Zi+1|X0, X1, . . . , Xi] . Then E[Zi] = E[E[Zi+1|X0, X1, . . . , Xi]] = E[Zi+1] . Thus, E[Zn] = E[Zn−1] = . . . , = E[Z0] .
SLIDE 12
Gambling Strategies
I play series of fair games (win with probability 1/2). Game 1: bet $1. Game i > 1: bet 2i if I won in round i − 1; bet i otherwise. Xi = amount won in ith game. (Xi < 0 if ith game lost). Zi = total winnings at end of ith game. Assume that (before starting to gamble) we decide to quit after k games: what are my expected winnings? E[Zk] = E[Z1] = 0.
SLIDE 13 A Different Strategy
Same gambling game. What happens if I:
- play a random number of games?
- decide to stop only when I have won (or lost) $1000?
SLIDE 14 Stopping Time
Definition A non-negative, integer random variable T is a stopping time for the sequence Z0, Z1, . . . if the event “T = n” depends only on the value of random variables Z0, Z1, . . . , Zn. Intuition: corresponds to a strategy for determining when to stop a sequence based only on values seen so far. In the gambling game:
- first time I win 10 games in a row: is a stopping time;
- the last time when I win: is not a stopping time.
SLIDE 15
Consider again the gambling game: let T be a stopping time. Zi = total winnings at end of ith game. What are my winnings at the stopping time, i.e. E[ZT]? Fair game: E[ZT] = E[Z0] = 0? “T =first time my total winnings are at least $1000” is a stopping time, and E[ZT] > 1000...
SLIDE 16 Martingale Stopping Theorem
Theorem If Z0, Z1, . . . is a martingale with respect to X1, X2, . . . and if T is a stopping time for X1, X2, . . . then E[ZT] = E[Z0] whenever one of the following holds:
- there is a constant c such that, for all i, |Zi| ≤ c;
- T is bounded;
- E[T] < ∞, and there is a constant c such that
E
- |Zi+1 − Zi|
- X1, . . . , Xi
- < c.
SLIDE 17 Martingale Stopping Theorem Applications
We play a sequence of fair game with the following stopping rules:
1 T is chosen from distribution with finite expectation:
E[ZT] = E[Z0].
2 T is the first time we made $1000: E[T] is unbounded. 3 We double until the first win. E[T] = 2 but
E
- |Zi+1 − Zi|
- X1, . . . , Xi
- is unbounded.
SLIDE 18 Example: The Gambler’s Ruin
- Consider a sequence of independent, fair 2-player gambling
games.
- In each round, each player wins or loses $1 with probability 1
2.
- Xi = amount won by player 1 on ith round.
- If player 1 has lost in round i: Xi < 0.
- Zi = total amount won by player 1 after ith rounds.
- Z0 = 0.
- Game ends when one player runs out of money
- Player 1 must stop when she loses net ℓ1 dollars (Zt = −ℓ1)
- Player 2 terminates when she loses net ℓ2 dollars (Zt = ℓ2).
- q = probability game ends with player 1 winning ℓ2 dollars.
SLIDE 19 Example: The Gambler’s Ruin
- T = first time player 1 wins ℓ2 dollars or loses ℓ1 dollars.
- T is a stopping time for X1, X2, . . . .
- Z0, Z1, . . . is a martingale.
- Zi’s are bounded.
- Martingale Stopping Theorem: E[ZT] = E[Z0] = 0.
E[ZT] = qℓ2 − (1 − q)ℓ1 = 0 q = ℓ1 ℓ1 + ℓ2
SLIDE 20 Example: A Ballot Theorem
- Candidate A and candidate B run for an election.
- Candidate A gets a votes.
- Candidate B gets b votes.
- a > b.
- Votes are counted in random order:
- chosen from all permutations on all n = a + b votes.
- What is the probability that A is always ahead in the count?
SLIDE 21 Example: A Ballot Theorem
- Si = number of votes A is leading by after i votes counted
- If A is trailing: Si < 0).
- Sn = a − b.
- For 0 ≤ k ≤ n − 1: Xk = Sn−k
n−k .
- Consider X0, X1, . . . , Xn.
- This sequence goes backward in time!
E[Xk|X0, X1, . . . , Xk−1] = ?
SLIDE 22 Example: A Ballot Theorem
E[Xk|X0, X1, . . . , Xk−1] = ?
- Conditioning on X0, X1, . . . , Xk−1: equivalent to conditioning
- n Sn, Sn−1, . . . , Sn−k+1,
- ai = number of votes for A after first i votes are counted.
- (n − k + 1)th vote: random vote among these first n − k + 1
votes. Sn−k = Sn−k+1 + 1 if (n − k + 1)th vote is for B Sn−k+1 − 1 if (n − k + 1)th vote is for A Sn−k =
with prob. n−k+1−an−k+1
n−k+1
Sn−k+1 − 1 with prob.
an−k+1 n−k+1
SLIDE 23 E[Sn−k|Sn−k+1] = (Sn−k+1 + 1)n − k + 1 − an−k+1 (n − k + 1) + (Sn−k+1 − 1) an−k+1 (n − k + 1) = Sn−k+1 n − k n − k + 1 (Since 2an−k+1 − n − k + 1 = Sn−k+1) E[Xk|X0, X1, . . . , Xk−1] = E Sn−k n − k
= Sn−k+1 n − k + 1 = Xk−1 = ⇒ X0, X1, . . . , Xn is a martingale.
SLIDE 24 Example: A Ballot Theorem
T = min{k : Xk = 0} if such k exists n − 1
- therwise
- T is a stopping time.
- T is bounded.
- Martingale Stopping Theorem:
E[XT] = E[X0] = E[Sn] n = a − b a + b . Two cases:
1 A leads throughout the count. 2 A does not lead throughout the count.
SLIDE 25
1 A leads throughout the count.
For 0 ≤ k ≤ n − 1: Sn−k > 0, then Xk > 0. T = n − 1. XT = Xn−1 = S1. A gets the first vote in the count: S1 = 1, then XT = 1.
2 A does not lead throughout the count.
For some k: Sk = 0. Then Xk = 0. T = k < n − 1. XT = 0.
SLIDE 26
Example: A Ballot Theorem
Putting all together:
1 A leads throughout the count: XT = 1. 2 A does not lead throughout the count: XT = 0
E[XT] = a − b a + b = 1 ∗ Pr(Case 1) + 0 ∗ Pr(Case 2) . That is Pr(A leads throughout the count) = a − b a + b .
SLIDE 27 A Different Gambling Game
Two stages:
1 roll one die; let X be the outcome; 2 roll X standard dice; your gain Z is the sum of the outcomes
What is your expected gain?
SLIDE 28 Wald’s Equation
Theorem Let X1, X2, . . . be nonnegative, independent, identically distributed random variables with distribution X. Let T be a stopping time for this sequence. If T and X have bounded expectation, then E T
Xi
Corollary of the martingale stopping theorem.
SLIDE 29 Proof
For i ≥ 1, let Zi = i
j=1(Xj − E[X]).
The sequence Z1, Z2, . . . is a martingale with respect to X1, X2, . . .. E[Z1] = 0, E[T] < ∞, and since Xi are nonnegative E
- |Zi+1 − Zi|
- X1, . . . , Xi
- = E[|Xi+1 − E[X]|] ≤ 2E[X] .
Hence we can apply the martingale stopping theorem to compute E[ZT] = E[Z1] = 0 . We now find = E[ZT] = E
T
(Xj − E[X]) = E
T
Xj − TE[X] = E
T
Xj − E[T] · E[X] = 0,
SLIDE 30 A Different Gambling Game
Two stages:
1 roll one die; let X be the outcome; 2 roll X standard dice; your gain Z is the sum of the outcomes
What is your expected gain? Yi = outcome of ith die in second stage. E[Z] = E X
Yi
X is a stopping time for Y1, Y2, . . . . By Wald’s equation: E[Z] = E[X]E[Yi] = 7 2 2 .
SLIDE 31 Example: a k-run
- We flip a fair coin until we get a consecutive sequence of k
HEADs.
- What’s the expected number of times we flip the coin.
- A SWITCH is a HEAD followed by a TAIL.
- Let X1 be the number of flips till k HEADs or the first
SWITCH
- Let Xi be the number of flips following the i − 1 SWITCH till
k HEADs or the next SWITCH
- Let T be the first i with k HEADs
E[Xi] =
i2−i +
k
i2−i + k2k−1 E[T] = 2k
- The expected number of coin flips is E[Xi]E[T]
SLIDE 32 Hoeffding’s Bound
Theorem Let X1, . . . , Xn be independent random variables with E[Xi] = µi and Pr(Bi ≤ Xi ≤ Bi + ci) = 1, then Pr
Xi −
n
µi
−
2ǫ2 n i=1 c2 i
Do we need independence?
SLIDE 33 Tail Inequalities
Theorem (Azuma-Hoeffding Inequality) Let Z0, Z1, . . . , Zn be a martingale (with respect to X1, X2, . . . ) such that |Zk − Zk−1| ≤ ck. Then, for all t ≥ 0 and any λ > 0, Pr(|Zt − Z0| ≥ λ) ≤ 2e−λ2/(2 t
k=1 c2 k ) .
The following corollary is often easier to apply. Corollary Let X0, X1, . . . be a martingale such that for all k ≥ 1, |Xk − Xk−1| ≤ c . Then for all t ≥ 1 and λ > 0, Pr
√ t
SLIDE 34 Tail Inequalities: A More General Form
Theorem (Azuma-Hoeffding Inequality) Let Z0, Z1, . . . , be a martingale with respect to X0, X1, X2, . . . , such that Bk ≤ Zk − Zk−1 ≤ Bk + ck , for some constants ck and for some random variables Bk that may be functions of X0, X1, . . . , Xk−1. Then, for all t ≥ 0 and any λ > 0, Pr(|Zt − Z0| ≥ λ) ≤ 2e−2λ2/(t
k=1 c2 k ) .
SLIDE 35 Proof
Let X k = X0, . . . , Xk and Yi = Zi − Zi−1. Since E[Zi | X i−1] = Zi−1, E[Yi | X i−1] = E[Zi − Zi−1 | X i−1] = 0 . Since Pr(Bi ≤ Yi ≤ Bi + ci | X i−1) = 1, by Hoeffding’s Lemma: E[eβYi | X i−1] ≤ eβ2c2
i /8 .
Lemma (Hoeffding’s Lemma) Let X be a random variable such that Pr(X ∈ [a, b]) = 1 and E[X] = 0. Then for every λ > 0, E[eλX] ≤ eλ2(a−b)2/8.
SLIDE 36 Proof of the Lemma
Since f (x) = eλx is a convex function, for any α ∈ (0, 1) and x ∈ [a, b], f (X) ≤ αf (a) + (1 − α)f (b) . Thus, for α = b−x
b−a ∈ (0, 1),
eλx ≤ b − x b − a eλa + x − a b − aeλb . Taking expectation, and using E[X] = 0, we have E
≤ b b − aeλa + a b − aeλb ≤ eλ2(b−a)2/8 .
SLIDE 37 E
≤ eβ2c2
i /8 .
EX n
i=1 Yi
EX n−1
i=1 Yi
X n−1 = EX n
i=1 YiEXn−1
≤ eβ2c2
n/8EX n−1
i=1 Yi
eβ2 n
i=1 c2 i /8
SLIDE 38 E[eβ n
i=1 Yi] ≤ eβ2 n i=1 c2 i /8
Pr(Zt − Z0 ≥ λ) = Pr t
Yi ≥ λ
i=1 Yi]
eβλ ≤ e−λβeβ2 t
i=1 c2 i /8
≤ 2e−2λ2/(t
k=1 c2 k ),
For β =
4λ t
i=1 c2 i .
Pr(|Zt − Z0| ≥ λ) ≤ 2e−2λ2/(t
k=1 c2 k )
SLIDE 39
Example
Assume that you play a sequence of n fair games, where the bet bi in game i depends on the outcome of previous games. Let B = maxi bi. The probability of winning or losing more than λ is bounded by Pr(|Zn| ≥ λ) ≤ 2e−2λ2/nB2 Pr(|Zn| ≥ λB√n) ≤ 2e−2λ2
SLIDE 40
Doob Martingale
Let X1, X2, . . . , Xn be sequence of random variables. Let Y = f (X1, . . . , Xn) be a random variable with E[|Y |] < ∞. Let Z0 = E[Y ] Let Zi = E[Y |X1, X2, . . . , Xi], i = 0, 1, . . . , n Z0, Z1, . . . , Zn is martingale with respect to X1, X2, . . . , Xn.
SLIDE 41
Proof
Fact E[E[V |U, W ]|W ] = E[V |W ]. Y = f (X1, . . . , Xn) Z0 = E[Y ] Zi = E[Y |X1, . . . , Xi], for all i ∈ 1, . . . , n. E[Zi+1|X1, X2, . . . , Xi] = E[E[Y |X1, X2, . . . , Xi+1]|X1, X2, . . . , Xi] = E[Y |X1, X2, . . . , Xi] = Zi .
SLIDE 42 Example: Polya’s Urn
- Start with M balls, R red, M − R blue.
- Repeat n times: We pick a ball uniformly at random. If Red
we add a red ball, else we add a blue ball.
- Xi = 1 if we add a red ball in step i, else Xi = 0
- We want to estimate Sn = n
i=1 Xi
E[St+1 | St] = St + R + St M + t E[St+1] = E[E[St+1 | St]] = E
M + t
M + R + tR/M M + t = (t + 1) R M
SLIDE 43 Example: Polya’s Urn
- Start with M balls, R red, M − R blue. Repeat n times: We
pick a ball uniformly at random. If Red we add a red ball, else we add a blue ball.
- Xi = 1 if added a red ball in step i, else Xi = 0, Sn = n
i=1 Xi
Zi = E[Sn|X1, X2, . . . , Xi] = i
j=1 Xj + (n − i) R+i
j=1 Xj
M+i
E[Zi+1 | X1, . . . , Xi] = E[E[Sn|X1, X2, . . . , Xi+1] | X1 = x1, . . . , Xi = xi] = E
i
xj + Xi+1 + (n − i − 1) R + i
j=1 xj + Xi+1
M + i + 1 =E
i
xj + R + i
j=1 xj
M + i + (n − i − 1) R + i
j=1 xj + R+i
j=1 xj
M+i
M + i + 1 = Zi
SLIDE 44 Example: Edge Exposure Martingale
Let G random graph from Gn,p. Consider m = n
2
in arbitrary order. Xi = 1 if ith edge is present
F(G) = size of maximum clique in G. Z0 = E[F(G)] Zi = E[F(G)|X1, X2, . . . , Xi], for i = 1, . . . , m. Z0, Z1, . . . , Zm is a Doob martingale. (F(G) could be any finite-valued function on graphs.)
SLIDE 45 Tail Inequalities: Doob Martingales
Let X1, . . . , Xn be sequence of random variables. Random variable Y :
- Y is a function of X1, X2, . . . , Xn;
- E[|Y |] < ∞.
Let Zi = E[Y |X1, . . . , Xi], i = 0, 1, . . . , n. Z0, Z1, . . . , Zn is martingale with respect to X1, . . . , Xn. If we can use Azuma-Hoeffding inequality: Pr(|Zn − Z0| ≥ λ) ≤ ..... then we have, Pr(|Y − E[Y ]| ≥ λ) ≤ ...... We need a bound on |Zi − Zi−1|.
SLIDE 46 McDiarmid Bound
Theorem Assume that f (X1, X2, . . . , Xn) satisfies, |f (x1, . . . , , xi, . . . , xn) − f (x1, . . . , , yi, . . . , xn)| ≤ ci . and X1, . . . , Xn are independent, then Pr(|f (X1, . . . , Xn) − E[f (X1, . . . , Xn)]| ≥ λ) ≤ 2e−2λ2/(n
k=1 c2 k ) .
[Changing the value of Xi changes the value of the function by at most ci.]
SLIDE 47 Proof
Define a Doob martingale Z0, Z1, . . . , Zn:
- Z0 = E[f (X1, . . . , Xn)] = E[f ( ¯
X)]
- Zi = E[f (X0, . . . , Xn) | X1, . . . , Xi] = E[f (Xi, . . . , Xn) | X i]
- Zn = f (X1, . . . , Xn) = f ( ¯
X) We want to prove that this martingale satisfies the conditions of Theorem (Azuma-Hoeffding Inequality) Let Z0, Z1, . . . , be a martingale with respect to X0, X1, X2, . . . , such that Bk ≤ Zk − Zk−1 ≤ Bk + ck , for some constants ck and for some random variables Bk that may be functions of X0, X1, . . . , Xk−1. Then, for all t ≥ 0 and any λ > 0, Pr(|Zt − Z0| ≥ λ) ≤ 2e−2λ2/(t
k=1 c2 k ) .
SLIDE 48
Lemma If X1, . . . , Xn are independent then for some random variable Bk, Bk ≤ Zk − Zk−1 ≤ Bk + ck , Zk − Zk−1 = E[f ( ¯ X) | X k] − E[f ( ¯ X) | X k−1] . Hence Zk − Zk−1 is bounded above by sup
x E[f ( ¯
X) | X k−1, Xk = x] − E[f ( ¯ X) | X k−1] and bounded below by inf
y E[f ( ¯
X) | X k−1, Xk = y] − E[f ( ¯ X) | X k−1] . Thus, we need to show sup
x E[f ( ¯
X) | X k−1, Xk = x] − inf
y E[f ( ¯
X) | X k−1, Xk = y] ≤ c ,
SLIDE 49 Zk − Zk−1 = sup
x,y E[f ( ¯
X, x) − f ( ¯ X, y) | X k−1]. Because the Xi are independent, the values for Xk+1, . . . , Xn do not depend on the values of X1, . . . , Xk. Hence, for any values ¯ z, x, y we have Pr(¯ z, x) = Pr(¯ z, y), and therefore sup
x,y E[f ( ¯
X, x) − f ( ¯ X, y) | X1 = z1, . . . , Xk−1 = zk−1] = sup
x,y
Pr
- (Xk+1 = zk+1) ∩ . . . ∩ (Xn = zn)
- ∗
- f (¯
z, x) − f (¯ z, y)
But f (¯ z, x) − f (¯ z, y) ≤ ci and therefore E[f ( ¯ X, x) − f ( ¯ X, y) | X k−1] ≤ ci
SLIDE 50
Example: Pattern Matching
Given a string and a pattern: is the pattern interesting? Does it appear more often than is expected in a random string? Is the number of occurrences of the pattern concentrated around the expectation?
SLIDE 51
A = (a1, a2, . . . , an) string of characters, each chosen independently and uniformly at random from Σ, with s = |Σ|. pattern: B = (b1, . . . , bk) fixed string, bi ∈ Σ. F = number of occurrences of B in random string A. E[F] = ?
SLIDE 52
A = (a1, a2, . . . , an) string of characters, each chosen independently and uniformly at random from Σ, with m = |Σ|. pattern: B = (b1, . . . , bk) fixed string, bi ∈ Σ. F= number occurrences of B in random string S. E[F] = (n − k + 1) 1 m k . Can we bound the deviation of F from its expectation?
SLIDE 53
F= number occurrences of B in random string A. Z0 = E[F] Zi = E[F|a1, . . . , ai], for i = 1, . . . , n. Z0, Z1, . . . , Zn is a Doob martingale. Zn = F.
SLIDE 54 F= number occurrences of B in random string A. Z0 = E[F] Zi = E[F|a1, . . . , ai], for i = 1, . . . , n. Z0, Z1, . . . , Zn is a Doob martingale. Zn = F. Each character in A can participate in no more than k occurrences
|Zi − Zi+1| ≤ k . Azuma-Hoeffding inequality (version 1): Pr(|F − E[F]| ≥ λ) ≤ 2e−λ2/(2nk2) .
SLIDE 55 Application: Balls and Bins
We are throwing m balls independently and uniformly at random into n bins. Let Xi = the bin that the ith ball falls into. Let F be the number of empty bins after the m balls are thrown. Then the sequence Zi = E[F | X1, . . . , Xi] is a Doob martingale. F = f (X1, X2, . . . , Xm) satisfies the Lipschitz condition with bound 1, thus |Zi+1 − Zi| ≤ 1 We therefore obtain Pr(|F − E[F]| ≥ ǫ) ≤ 2e−2ǫ2/m Here E[F] = n
n m , but we could obtain the concentration result without knowing E[F].
SLIDE 56
Application: Chromatic Number
Given a random graph G in Gn,p, the chromatic number χ(G) is the minimum number of colors required to color all vertices of the graph so that no adjacent vertices have the same color. We use the vertex exposure martingale defined below: Let Gi be the random subgraph of G induced by the set of vertices 1, . . . , i, let Z0 = E[χ(G)], and let Zi = E[χ(G) | G1, . . . , Gi] . Since a vertex uses no more than one new color, again we have that the gap between Zi and Zi−1 is at most 1. We conclude Pr(|χ(G) − E[χ(G)]| ≥ λ√n) ≤ 2e−2λ2 . This result holds even without knowing E[χ(G)].
SLIDE 57 Example: Edge Exposure Martingale
Let G random graph from Gn,p. Consider m = n
2
in arbitrary order. Xi = 1 if ith edge is present
F(G) = size maximum clique in G. Z0 = E[F(G)] Zi = E[F(G)|X1, X2, . . . , Xi], for i = 1, . . . , m. Z0, Z1, . . . , Zm is a Doob martingale. (F(G) could be any finite-valued function on graphs.)