Persisting randomness in randomly growing discrete structures: - - PowerPoint PPT Presentation

persisting randomness in randomly growing discrete
SMART_READER_LITE
LIVE PREVIEW

Persisting randomness in randomly growing discrete structures: - - PowerPoint PPT Presentation

Persisting randomness in randomly growing discrete structures: graphs and search trees R. Gr ubel Leibniz Universit at Hannover Paris, AofA 2014 Examples: Coin tossing vs. P olya urn Examples: Coin tossing vs. P olya urn


slide-1
SLIDE 1

Persisting randomness in randomly growing discrete structures: graphs and search trees

  • R. Gr¨

ubel Leibniz Universit¨ at Hannover

Paris, AofA 2014

slide-2
SLIDE 2

Examples: Coin tossing vs. P´

  • lya urn
slide-3
SLIDE 3

Examples: Coin tossing vs. P´

  • lya urn
slide-4
SLIDE 4

Examples: Coin tossing vs. P´

  • lya urn
slide-5
SLIDE 5

Examples: Coin tossing vs. P´

  • lya urn
slide-6
SLIDE 6

Examples: Coin tossing vs. P´

  • lya urn
slide-7
SLIDE 7

Examples: Coin tossing vs. P´

  • lya urn
slide-8
SLIDE 8

Examples: Coin tossing vs. P´

  • lya urn
slide-9
SLIDE 9

Persistence of randomness: What is it?

slide-10
SLIDE 10

Persistence of randomness: What is it?

In words: The influence of early values may or may not go away in the long run.

slide-11
SLIDE 11

Persistence of randomness: What is it?

In words: The influence of early values may or may not go away in the long run. Formally, we have a sequence X = (Xn)n∈N of random variables on some probability space (Ω, A, P), and the tail σ-field T (X) :=

  • n=1

σ

  • {Xm : m ≥ n}
  • may or may not be P-trivial in the sense of

P(A) = 0 or P(A) = 1 for all A ∈ T (X).

slide-12
SLIDE 12

Persistence of randomness: What is it?

In words: The influence of early values may or may not go away in the long run. Formally, we have a sequence X = (Xn)n∈N of random variables on some probability space (Ω, A, P), and the tail σ-field T (X) :=

  • n=1

σ

  • {Xm : m ≥ n}
  • may or may not be P-trivial in the sense of

P(A) = 0 or P(A) = 1 for all A ∈ T (X). Kolmogorov’s zero-one law: No persisting randomness in i.i.d. sequences.

slide-13
SLIDE 13

Tail σ-fields (and topologies) via boundary theory

slide-14
SLIDE 14

Tail σ-fields (and topologies) via boundary theory

  • Let F be a combinatorial family, Fn: objects of size n.
slide-15
SLIDE 15

Tail σ-fields (and topologies) via boundary theory

  • Let F be a combinatorial family, Fn: objects of size n.
  • Let X be a Markov chain with P(Xn ∈ Fn) = 1, n ∈ N.
slide-16
SLIDE 16

Tail σ-fields (and topologies) via boundary theory

  • Let F be a combinatorial family, Fn: objects of size n.
  • Let X be a Markov chain with P(Xn ∈ Fn) = 1, n ∈ N.
  • Say that (yn)n∈N with yn ∈ Fn, n ∈ N, converges iff
  • P(X1 = x1, . . . , Xl = xl|Xn = yn)
  • n∈N

converges for all fixed l ∈ N, x1 ∈ F1, . . . , xl ∈ Fl.

slide-17
SLIDE 17

Tail σ-fields (and topologies) via boundary theory

  • Let F be a combinatorial family, Fn: objects of size n.
  • Let X be a Markov chain with P(Xn ∈ Fn) = 1, n ∈ N.
  • Say that (yn)n∈N with yn ∈ Fn, n ∈ N, converges iff
  • P(X1 = x1, . . . , Xl = xl|Xn = yn)
  • n∈N

converges for all fixed l ∈ N, x1 ∈ F1, . . . , xl ∈ Fl.

  • This leads to a compactification of F with boundary ∂F.
slide-18
SLIDE 18

Tail σ-fields (and topologies) via boundary theory

  • Let F be a combinatorial family, Fn: objects of size n.
  • Let X be a Markov chain with P(Xn ∈ Fn) = 1, n ∈ N.
  • Say that (yn)n∈N with yn ∈ Fn, n ∈ N, converges iff
  • P(X1 = x1, . . . , Xl = xl|Xn = yn)
  • n∈N

converges for all fixed l ∈ N, x1 ∈ F1, . . . , xl ∈ Fl.

  • This leads to a compactification of F with boundary ∂F.

Theorem (Doob, Dynkin,. . .)

(a) Xn → X∞ almost surely with P(X∞ ∈ ∂F) = 1. (b) X∞ generates T (X).

slide-19
SLIDE 19

Graph limits

slide-20
SLIDE 20

Graph limits

  • Let F = G be the family of finite simple graphs.
  • Let V (G) and E(G) be the vertices and edges of G ∈ G.
slide-21
SLIDE 21

Graph limits

  • Let F = G be the family of finite simple graphs.
  • Let V (G) and E(G) be the vertices and edges of G ∈ G.
  • For G, H ∈ G let Γ(H, G) be the set of injective functions

φ : V (H) → V (G).

  • Let T(H, G) be the set of all φ ∈ Γ(H, G) that satisfy

{i, j} ∈ E(H) ⇐ ⇒ {φ(i), φ(j)} ∈ E(G).

  • Let

t(H, G) := #T(H, G)/#ΓH

G.

slide-22
SLIDE 22

Graph limits

  • Let F = G be the family of finite simple graphs.
  • Let V (G) and E(G) be the vertices and edges of G ∈ G.
  • For G, H ∈ G let Γ(H, G) be the set of injective functions

φ : V (H) → V (G).

  • Let T(H, G) be the set of all φ ∈ Γ(H, G) that satisfy

{i, j} ∈ E(H) ⇐ ⇒ {φ(i), φ(j)} ∈ E(G).

  • Let

t(H, G) := #T(H, G)/#ΓH

G.

In words: Sample #V (H) vertices from V (G) without replacement. Then t(H, G) is the probability that the induced subgraph is isomorphic to H.

slide-23
SLIDE 23

Graph limits

  • Let F = G be the family of finite simple graphs.
  • Let V (G) and E(G) be the vertices and edges of G ∈ G.
  • For G, H ∈ G let Γ(H, G) be the set of injective functions

φ : V (H) → V (G).

  • Let T(H, G) be the set of all φ ∈ Γ(H, G) that satisfy

{i, j} ∈ E(H) ⇐ ⇒ {φ(i), φ(j)} ∈ E(G).

  • Let

t(H, G) := #T(H, G)/#ΓH

G.

In words: Sample #V (H) vertices from V (G) without replacement. Then t(H, G) is the probability that the induced subgraph is isomorphic to H.

The subgraph sampling topology: (Gn)n∈N ⊂ G of converges iff (t(H, Gn))n∈N converges for all H ∈ G. (Aldous, Lovasz, . . .).

slide-24
SLIDE 24

A structural view

slide-25
SLIDE 25

A structural view

  • Let F be a family of bounded functions f : F → R.
  • Embed F into RF via evaluation, y →
  • f → f (y)
  • .
slide-26
SLIDE 26

A structural view

  • Let F be a family of bounded functions f : F → R.
  • Embed F into RF via evaluation, y →
  • f → f (y)
  • .
  • Doob-Martin: F = {K(x, ·) : x ∈ F} with

K(x, y) = P(Xn = y|Xm = x) P(Xn = y) . This is the Martin kernel.

slide-27
SLIDE 27

A structural view

  • Let F be a family of bounded functions f : F → R.
  • Embed F into RF via evaluation, y →
  • f → f (y)
  • .
  • Doob-Martin: F = {K(x, ·) : x ∈ F} with

K(x, y) = P(Xn = y|Xm = x) P(Xn = y) . This is the Martin kernel.

  • Subgraph sampling: F = {t(H, ·) : H ∈ G}.
slide-28
SLIDE 28

A structural view

  • Let F be a family of bounded functions f : F → R.
  • Embed F into RF via evaluation, y →
  • f → f (y)
  • .
  • Doob-Martin: F = {K(x, ·) : x ∈ F} with

K(x, y) = P(Xn = y|Xm = x) P(Xn = y) . This is the Martin kernel.

  • Subgraph sampling: F = {t(H, ·) : H ∈ G}.

General questions for a given graph model of the Markovian type, growing one node at a time: (a) Are these topologies the same? (b) What is the respective tail σ-field?

slide-29
SLIDE 29

Graphs: The Erd˝

  • s-R´

enyi model

slide-30
SLIDE 30

Graphs: The Erd˝

  • s-R´

enyi model

Model description, as a Markov chain that grows by one node at a time, with parameter p, 0 < p < 1:

  • X ER

1

is the single element of G1.

slide-31
SLIDE 31

Graphs: The Erd˝

  • s-R´

enyi model

Model description, as a Markov chain that grows by one node at a time, with parameter p, 0 < p < 1:

  • X ER

1

is the single element of G1.

  • To move from X ER

n

to X ER

n+1,

– add the node n + 1, – add the edges {i, n + 1}, i ∈ {1, . . . , n}, independently and with probability p. – randomly relabel V (Xn+1),

slide-32
SLIDE 32

Graphs: The Erd˝

  • s-R´

enyi model

Model description, as a Markov chain that grows by one node at a time, with parameter p, 0 < p < 1:

  • X ER

1

is the single element of G1.

  • To move from X ER

n

to X ER

n+1,

– add the node n + 1, – add the edges {i, n + 1}, i ∈ {1, . . . , n}, independently and with probability p. – randomly relabel V (Xn+1),

Theorem

For X ER = (X ER

n )n∈N the Doob-Martin and the graph testing

topology coincide. In particular, T (X ER) is trivial.

slide-33
SLIDE 33

Graphs: The Erd˝

  • s-R´

enyi model

Model description, as a Markov chain that grows by one node at a time, with parameter p, 0 < p < 1:

  • X ER

1

is the single element of G1.

  • To move from X ER

n

to X ER

n+1,

– add the node n + 1, – add the edges {i, n + 1}, i ∈ {1, . . . , n}, independently and with probability p. – randomly relabel V (Xn+1),

Theorem

For X ER = (X ER

n )n∈N the Doob-Martin and the graph testing

topology coincide. In particular, T (X ER) is trivial.

– If we omit the relabelling then the Markov chain is of the complete memory type, and the Doob-Martin boundary is the projective limit (‘the sequence is the limit’).

slide-34
SLIDE 34

Graphs: The uniform attachment model

slide-35
SLIDE 35

Graphs: The uniform attachment model

  • Again, X ua

1 is the single element of G1.

  • Construct X ua

n+1 from X ua n by adding all edges {i, j} ⊂ [n + 1]

not yet in Xn, independently and with probability 1/(n + 1).

slide-36
SLIDE 36

Graphs: The uniform attachment model

  • Again, X ua

1 is the single element of G1.

  • Construct X ua

n+1 from X ua n by adding all edges {i, j} ⊂ [n + 1]

not yet in Xn, independently and with probability 1/(n + 1). Let 1{i,j}, 1 ≤ i < j, be the edge indicator functions.

slide-37
SLIDE 37

Graphs: The uniform attachment model

  • Again, X ua

1 is the single element of G1.

  • Construct X ua

n+1 from X ua n by adding all edges {i, j} ⊂ [n + 1]

not yet in Xn, independently and with probability 1/(n + 1). Let 1{i,j}, 1 ≤ i < j, be the edge indicator functions.

Theorem

In the Doob-Martin topology associated with X ua convergence of a sequence of graphs is equivalent to the pointwise convergence of all edge indicator functions. Further, T (X ua) is trivial.

slide-38
SLIDE 38

Graphs: The uniform attachment model

  • Again, X ua

1 is the single element of G1.

  • Construct X ua

n+1 from X ua n by adding all edges {i, j} ⊂ [n + 1]

not yet in Xn, independently and with probability 1/(n + 1). Let 1{i,j}, 1 ≤ i < j, be the edge indicator functions.

Theorem

In the Doob-Martin topology associated with X ua convergence of a sequence of graphs is equivalent to the pointwise convergence of all edge indicator functions. Further, T (X ua) is trivial.

– The result fits the structural view, with F the set of edge indicators. In essence, a graph is identified with its adjacency matrix.

slide-39
SLIDE 39

Graphs: The uniform attachment model

  • Again, X ua

1 is the single element of G1.

  • Construct X ua

n+1 from X ua n by adding all edges {i, j} ⊂ [n + 1]

not yet in Xn, independently and with probability 1/(n + 1). Let 1{i,j}, 1 ≤ i < j, be the edge indicator functions.

Theorem

In the Doob-Martin topology associated with X ua convergence of a sequence of graphs is equivalent to the pointwise convergence of all edge indicator functions. Further, T (X ua) is trivial.

– The result fits the structural view, with F the set of edge indicators. In essence, a graph is identified with its adjacency matrix. – This differs from the graph testing topology: The Doob-Martin topology depends on the transition mechanism.

slide-40
SLIDE 40

Graphs: The uniform attachment model

  • Again, X ua

1 is the single element of G1.

  • Construct X ua

n+1 from X ua n by adding all edges {i, j} ⊂ [n + 1]

not yet in Xn, independently and with probability 1/(n + 1). Let 1{i,j}, 1 ≤ i < j, be the edge indicator functions.

Theorem

In the Doob-Martin topology associated with X ua convergence of a sequence of graphs is equivalent to the pointwise convergence of all edge indicator functions. Further, T (X ua) is trivial.

– The result fits the structural view, with F the set of edge indicators. In essence, a graph is identified with its adjacency matrix. – This differs from the graph testing topology: The Doob-Martin topology depends on the transition mechanism.

Work in progress: Other graph models, especially cases with non-trivial tail σ-fields. P´

  • lya urns play a prominent role.
slide-41
SLIDE 41

Binary search trees

  • V = {0, 1}⋆ (0-1 words) is the set of nodes,
slide-42
SLIDE 42

Binary search trees

  • V = {0, 1}⋆ (0-1 words) is the set of nodes,
  • a prefix-stable x ⊂ V is a binary tree,

Bn is the family of binary trees with n nodes.

slide-43
SLIDE 43

Binary search trees

  • V = {0, 1}⋆ (0-1 words) is the set of nodes,
  • a prefix-stable x ⊂ V is a binary tree,

Bn is the family of binary trees with n nodes. ⋆ BST chain: L(Xn+1|Xn = x) = unif

  • {y ∈ Bn+1 : x ⊂ y}
  • .
slide-44
SLIDE 44

BST: The boundary

– Let F = B be the family of binary trees. – For x ∈ B and u ∈ V let #x(u) be the number of nodes in x with prefix u (fringe subtree size).

slide-45
SLIDE 45

BST: The boundary

– Let F = B be the family of binary trees. – For x ∈ B and u ∈ V let #x(u) be the number of nodes in x with prefix u (fringe subtree size).

Theorem (Evans, Gr., Wakolbinger 2012)

Let X = (Xn)n∈N be the BST chain. (a) Doob-Martin convergence of (xn)n∈N ⊂ B is equivalent to convergence of #xn(u)/#xn for all u ∈ V.

slide-46
SLIDE 46

BST: The boundary

– Let F = B be the family of binary trees. – For x ∈ B and u ∈ V let #x(u) be the number of nodes in x with prefix u (fringe subtree size).

Theorem (Evans, Gr., Wakolbinger 2012)

Let X = (Xn)n∈N be the BST chain. (a) Doob-Martin convergence of (xn)n∈N ⊂ B is equivalent to convergence of #xn(u)/#xn for all u ∈ V. (b) The Martin boundary ∂B may be represented by the set of all probability measures on {0, 1}∞, endowed with weak convergence.

slide-47
SLIDE 47

BST: The boundary

– Let F = B be the family of binary trees. – For x ∈ B and u ∈ V let #x(u) be the number of nodes in x with prefix u (fringe subtree size).

Theorem (Evans, Gr., Wakolbinger 2012)

Let X = (Xn)n∈N be the BST chain. (a) Doob-Martin convergence of (xn)n∈N ⊂ B is equivalent to convergence of #xn(u)/#xn for all u ∈ V. (b) The Martin boundary ∂B may be represented by the set of all probability measures on {0, 1}∞, endowed with weak convergence. (c) The support of X∞ is the whole of ∂B. In particular, T (X) is not trivial.

slide-48
SLIDE 48

BST: The boundary

– Let F = B be the family of binary trees. – For x ∈ B and u ∈ V let #x(u) be the number of nodes in x with prefix u (fringe subtree size).

Theorem (Evans, Gr., Wakolbinger 2012)

Let X = (Xn)n∈N be the BST chain. (a) Doob-Martin convergence of (xn)n∈N ⊂ B is equivalent to convergence of #xn(u)/#xn for all u ∈ V. (b) The Martin boundary ∂B may be represented by the set of all probability measures on {0, 1}∞, endowed with weak convergence. (c) The support of X∞ is the whole of ∂B. In particular, T (X) is not trivial. Plan: Use this to improve results on convergence in distribution to convergence almost surely.

slide-49
SLIDE 49

BST: the IDLA view

(Internal diffusion limited aggregation, Diaconis and Fulton 1991)

slide-50
SLIDE 50

BST: the IDLA view

(Internal diffusion limited aggregation, Diaconis and Fulton 1991)

  • An infinite graph (here: V) is gradually filled up by randomly

growing contiguous subsets.

slide-51
SLIDE 51

BST: the IDLA view

(Internal diffusion limited aggregation, Diaconis and Fulton 1991)

  • An infinite graph (here: V) is gradually filled up by randomly

growing contiguous subsets.

  • Describe these by their boundary or frontier function, here

Bx : ∂V → N, Bx(v) := min

  • k ∈ N : (v1, . . . , vk) /

∈ x

  • ,

for all v = (vi)i∈N ∈ ∂V.

slide-52
SLIDE 52

BST: the IDLA view

(Internal diffusion limited aggregation, Diaconis and Fulton 1991)

  • An infinite graph (here: V) is gradually filled up by randomly

growing contiguous subsets.

  • Describe these by their boundary or frontier function, here

Bx : ∂V → N, Bx(v) := min

  • k ∈ N : (v1, . . . , vk) /

∈ x

  • ,

for all v = (vi)i∈N ∈ ∂V.

  • ∂V = {0, 1}∞ is the boundary of the infinite binary tree V.
slide-53
SLIDE 53

BST: the IDLA view

(Internal diffusion limited aggregation, Diaconis and Fulton 1991)

  • An infinite graph (here: V) is gradually filled up by randomly

growing contiguous subsets.

  • Describe these by their boundary or frontier function, here

Bx : ∂V → N, Bx(v) := min

  • k ∈ N : (v1, . . . , vk) /

∈ x

  • ,

for all v = (vi)i∈N ∈ ∂V.

  • ∂V = {0, 1}∞ is the boundary of the infinite binary tree V.
  • The tree boundary ∂V can be mapped to the unit interval via

v = (vi)i∈N → 1 2 +

  • i=1

2vi − 1 2i+1 .

(useful for visualization)

slide-54
SLIDE 54

The frontier function

1 {0, 1}∞

slide-55
SLIDE 55

The frontier function

1 {0, 1}∞

slide-56
SLIDE 56

The frontier function

1 {0, 1}∞

slide-57
SLIDE 57

The frontier function

1 {0, 1}∞

slide-58
SLIDE 58

Dynamics: An example with n = 50, 100, 200

slide-59
SLIDE 59

Dynamics: An example with n = 50, 100, 200

slide-60
SLIDE 60

Dynamics: An example with n = 50, 100, 200

slide-61
SLIDE 61

Picture suggests that we need to shift and to smooth.

slide-62
SLIDE 62

Picture suggests that we need to shift and to smooth.

  • Let µ be the uniform distribution on ∂V (coin tossing),
  • let ‘’ be the lexicographic order on ∂V,
  • let Hn = n

k=1 1/k be the harmonic numbers.

slide-63
SLIDE 63

Picture suggests that we need to shift and to smooth.

  • Let µ be the uniform distribution on ∂V (coin tossing),
  • let ‘’ be the lexicographic order on ∂V,
  • let Hn = n

k=1 1/k be the harmonic numbers.

With (Xn)n∈N be the BST chain define Yn = (Yn(v))v∈∂V by Yn(v) :=

  • wv
  • BXn(w) − Hn
  • µ(dw), v ∈ ∂V.
slide-64
SLIDE 64

Picture suggests that we need to shift and to smooth.

  • Let µ be the uniform distribution on ∂V (coin tossing),
  • let ‘’ be the lexicographic order on ∂V,
  • let Hn = n

k=1 1/k be the harmonic numbers.

With (Xn)n∈N be the BST chain define Yn = (Yn(v))v∈∂V by Yn(v) :=

  • wv
  • BXn(w) − Hn
  • µ(dw), v ∈ ∂V.
  • For u, v ∈ ∂V, u = v, let d(u, v) := 2−|u∧v|, with u ∧ v the

common prefix of u and v.

  • (∂V, d) is a compact, totally disconnected metric space.
  • With the supremum norm · ∞, the space C(∂V) of

continuous functions is a separable Banach space.

  • The Yn’s are random element of this Banach space.
slide-65
SLIDE 65

Theorem (2009, 2014)

(a) Yn → Y∞ almost surely as n → ∞. (b) Almost all paths of the limit process Y∞ are not Lipschitz continuous. (c) The paths of Y∞ are H¨

  • lder continuous with exponent α for

all α < 1, again with probability 1.

slide-66
SLIDE 66

Theorem (2009, 2014)

(a) Yn → Y∞ almost surely as n → ∞. (b) Almost all paths of the limit process Y∞ are not Lipschitz continuous. (c) The paths of Y∞ are H¨

  • lder continuous with exponent α for

all α < 1, again with probability 1.

Yn(ω) for two different ω’s, with n = 500 (blue) and n = 1000 (red).

1 1 1 1