INFORMATION THEORY: SOURCES, DIRICHLET SERIES, REALISTIC ANALYSIS - - PowerPoint PPT Presentation

information theory sources dirichlet series realistic
SMART_READER_LITE
LIVE PREVIEW

INFORMATION THEORY: SOURCES, DIRICHLET SERIES, REALISTIC ANALYSIS - - PowerPoint PPT Presentation

INFORMATION THEORY: SOURCES, DIRICHLET SERIES, REALISTIC ANALYSIS OF DATA STRUCTURES Mathieu Roux and Brigitte Vall ee GREYC Laboratory (CNRS and University of Caen, France) Talk also based on joint works with Viviane Baladi , Eda Cesaratto ,


slide-1
SLIDE 1

INFORMATION THEORY: SOURCES, DIRICHLET SERIES, REALISTIC ANALYSIS OF DATA STRUCTURES

Mathieu Roux and Brigitte Vall´ ee GREYC Laboratory (CNRS and University of Caen, France) Talk also based on joint works with Viviane Baladi, Eda Cesaratto, Julien Cl´ ement, Jim Fill, Philippe Flajolet WORDS 2011, Prague, September 2011

slide-2
SLIDE 2

Description of a framework which – unifies the analyses for text algorithms and searching/sorting algorithms

slide-3
SLIDE 3

Description of a framework which – unifies the analyses for text algorithms and searching/sorting algorithms – provides a general model for sources – shows the importance of the Dirichlet generating functions – explains the importance of tameness for sources

slide-4
SLIDE 4

Description of a framework which – unifies the analyses for text algorithms and searching/sorting algorithms – provides a general model for sources – shows the importance of the Dirichlet generating functions – explains the importance of tameness for sources – defines a natural subclass of sources, the dynamical sources – provides sufficient conditions for tameness of dynamical sources

slide-5
SLIDE 5

Description of a framework which – unifies the analyses for text algorithms and searching/sorting algorithms – provides a general model for sources – shows the importance of the Dirichlet generating functions – explains the importance of tameness for sources – defines a natural subclass of sources, the dynamical sources – provides sufficient conditions for tameness of dynamical sources – provides probabilistic analyses for data structures built on tame sources.

slide-6
SLIDE 6

Plan of the talk. – General motivations: Dirichlet generating functions and tameness – An important class of sources: dynamical sources. – Tameness in the case of dynamical sources – Conclusion and possible extensions.

slide-7
SLIDE 7

Plan of the talk. – General motivations: Dirichlet generating functions and tameness. – An important class of sources: dynamical sources. – Tameness in the case of dynamical sources – Conclusion and possible extensions.

slide-8
SLIDE 8

The classical framework for analysis of algorithms in two main algorithmic domains: Text algorithms – Sorting or Searching algorithms.

slide-9
SLIDE 9

The classical framework for analysis of algorithms in two main algorithmic domains: Text algorithms – Sorting or Searching algorithms. – In text algorithms, algorithms deal with words – In sorting or searching algorithms, algorithms deal with keys. A word or a key are both a sequence of symbols ... but

slide-10
SLIDE 10

The classical framework for analysis of algorithms in two main algorithmic domains: Text algorithms – Sorting or Searching algorithms. – In text algorithms, algorithms deal with words – In sorting or searching algorithms, algorithms deal with keys. A word or a key are both a sequence of symbols ... but – for comparing two words, importance of the structure of words – for comparing two keys, transparence of the structure of keys

  • nly their relative order plays a role.
slide-11
SLIDE 11

Text algorithms and dictionaries : The trie structure Probabilistic study

a a a a a a b b b b b c c c c c abc b c b b b cba bbc cab

slide-12
SLIDE 12

Text algorithms and dictionaries : The trie structure Probabilistic study

a a a a a a b b b b b c c c c c abc b c b b b cba bbc cab

Main parameter on a node nw labelled with prefix w: Nw := the number of words which begin with prefix w. Nw := the number of words which go through the node nw

slide-13
SLIDE 13

Text algorithms and dictionaries : The trie structure Probabilistic study

a a a a a a b b b b b c c c c c abc b c b b b cba bbc cab

Main parameter on a node nw labelled with prefix w: Nw := the number of words which begin with prefix w. Nw := the number of words which go through the node nw The size, and the path length of a trie equal R =

  • w∈Σ⋆

1[Nw≥2] T =

  • w∈Σ⋆

1[Nw≥2] · Nw, Central role of pw :=the probability that a word begins with prefix w.

slide-14
SLIDE 14

A realistic framework for sorting or searching. Keys are viewed as words and are compared [wrt the lexicographic order]. The realistic unit cost is now the symbol–comparison.

slide-15
SLIDE 15

A realistic framework for sorting or searching. Keys are viewed as words and are compared [wrt the lexicographic order]. The realistic unit cost is now the symbol–comparison. The realistic cost of the comparison between two words A and B, A = a1 a2 a3 . . . ai . . . and B = b1 b2 b3 . . . bi . . . equals k + 1, where k is the length of their largest common prefix k := max{i; ∀j ≤ i, aj = bj}= the coincidence c(A, B)

slide-16
SLIDE 16

A realistic framework for sorting or searching. Keys are viewed as words and are compared [wrt the lexicographic order]. The realistic unit cost is now the symbol–comparison. The realistic cost of the comparison between two words A and B, A = a1 a2 a3 . . . ai . . . and B = b1 b2 b3 . . . bi . . . equals k + 1, where k is the length of their largest common prefix k := max{i; ∀j ≤ i, aj = bj}= the coincidence c(A, B)

slide-17
SLIDE 17

A realistic framework for sorting or searching. Keys are viewed as words and are compared [wrt the lexicographic order]. The realistic unit cost is now the symbol–comparison. The realistic cost of the comparison between two words A and B, A = a1 a2 a3 . . . ai . . . and B = b1 b2 b3 . . . bi . . . equals k + 1, where k is the length of their largest common prefix k := max{i; ∀j ≤ i, aj = bj}= the coincidence c(A, B) The probabilistic study of the coincidence deals with pw:= the probability that a word begins with prefix w. Pr[c(A, B) ≥ k] = Pr[A and B begin with the same w of length k]

slide-18
SLIDE 18

A realistic framework for sorting or searching. Keys are viewed as words and are compared [wrt the lexicographic order]. The realistic unit cost is now the symbol–comparison. The realistic cost of the comparison between two words A and B, A = a1 a2 a3 . . . ai . . . and B = b1 b2 b3 . . . bi . . . equals k + 1, where k is the length of their largest common prefix k := max{i; ∀j ≤ i, aj = bj}= the coincidence c(A, B) The probabilistic study of the coincidence deals with pw:= the probability that a word begins with prefix w. Pr[c(A, B) ≥ k] = Pr[A and B begin with the same w of length k] =

  • |w|=k

p2

w

slide-19
SLIDE 19

The example of the binary search tree (BST)

slide-20
SLIDE 20

The example of the binary search tree (BST)

Number of symbol comparisons needed for inserting F = abbbbbbb.

slide-21
SLIDE 21

The example of the binary search tree (BST)

Number of symbol comparisons needed for inserting F = abbbbbbb. = 7 for comparing to A c(F, A) = 6

slide-22
SLIDE 22

The example of the binary search tree (BST)

Number of symbol comparisons needed for inserting F = abbbbbbb. = 7 for comparing to A c(F, A) = 6 + 8 for comparing to B c(F, B) = 7

slide-23
SLIDE 23

The example of the binary search tree (BST)

Number of symbol comparisons needed for inserting F = abbbbbbb. = 7 for comparing to A c(F, A) = 6 + 8 for comparing to B c(F, B) = 7 + 1 for comparing to C c(F, C) = 0

slide-24
SLIDE 24

The example of the binary search tree (BST)

Number of symbol comparisons needed for inserting F = abbbbbbb. = 7 for comparing to A c(F, A) = 6 + 8 for comparing to B c(F, B) = 7 + 1 for comparing to C c(F, C) = 0 Total = 16 To be compared to the number of key comparisons [= 3]

slide-25
SLIDE 25

The example of the binary search tree (BST)

Number of symbol comparisons needed for inserting F = abbbbbbb. = 7 for comparing to A c(F, A) = 6 + 8 for comparing to B c(F, B) = 7 + 1 for comparing to C c(F, C) = 0 Total = 16 To be compared to the number of key comparisons [= 3]

This defines the symbol-path-length of a BST based on the coincidence We perform a probabilistic study of this symbol path-length

slide-26
SLIDE 26

Now, we work inside an unifying framework where searching and sorting algorithms are viewed as text algorithms.

slide-27
SLIDE 27

Now, we work inside an unifying framework where searching and sorting algorithms are viewed as text algorithms. In this context, the probabilistic behaviour of algorithms heavily depends

  • n the mechanism which produces words.
slide-28
SLIDE 28

Now, we work inside an unifying framework where searching and sorting algorithms are viewed as text algorithms. In this context, the probabilistic behaviour of algorithms heavily depends

  • n the mechanism which produces words.

A source:= a mechanism which produces symbols from alphabet Σ,

  • ne for each time unit.

When (discrete) time evolves, a source produces (infinite) words of ΣN.

slide-29
SLIDE 29

Now, we work inside an unifying framework where searching and sorting algorithms are viewed as text algorithms. In this context, the probabilistic behaviour of algorithms heavily depends

  • n the mechanism which produces words.

A source:= a mechanism which produces symbols from alphabet Σ,

  • ne for each time unit.

When (discrete) time evolves, a source produces (infinite) words of ΣN. For w ∈ Σ⋆, pw := probability that a word begins with the prefix w. The set {pw, w ∈ Σ⋆} defines the source S.

slide-30
SLIDE 30

Fundamental role of the Dirichlet generating functions of the source Λ(s) :=

  • w∈Σ⋆

ps

w,

Λk(s) =

  • w∈Σk

ps

w,

 Λ =

  • k≥0

Λk   Remark: Λk(1) = 1 for any k, Λ(1) = ∞.

slide-31
SLIDE 31

Fundamental role of the Dirichlet generating functions of the source Λ(s) :=

  • w∈Σ⋆

ps

w,

Λk(s) =

  • w∈Σk

ps

w,

 Λ =

  • k≥0

Λk   Remark: Λk(1) = 1 for any k, Λ(1) = ∞. – they encapsulate the main probabilistic properties of the source – they translate them into analytic properties

slide-32
SLIDE 32

Fundamental role of the Dirichlet generating functions of the source Λ(s) :=

  • w∈Σ⋆

ps

w,

Λk(s) =

  • w∈Σk

ps

w,

 Λ =

  • k≥0

Λk   Remark: Λk(1) = 1 for any k, Λ(1) = ∞. – they encapsulate the main probabilistic properties of the source – they translate them into analytic properties For instance, the entropy hS, the coincidence cS h(S) := lim

k→∞

−1 k

  • w∈Σk

pw log pw = −1 k lim

k→∞ Λ′ k(1)

Pr[cS ≥ k] =

  • w∈Σk

p2

w = Λk(2)

slide-33
SLIDE 33

Fundamental role of the Dirichlet generating functions of the source Λ(s) :=

  • w∈Σ⋆

ps

w,

Λk(s) =

  • w∈Σk

ps

w,

 Λ =

  • k≥0

Λk   Remark: Λk(1) = 1 for any k, Λ(1) = ∞. – they encapsulate the main probabilistic properties of the source – they translate them into analytic properties For instance, the entropy hS, the coincidence cS h(S) := lim

k→∞

−1 k

  • w∈Σk

pw log pw = −1 k lim

k→∞ Λ′ k(1)

Pr[cS ≥ k] =

  • w∈Σk

p2

w = Λk(2)

– they intervene in probabilistic analysis of algorithms and data structures.

slide-34
SLIDE 34

Exact average-case analysis for Tries or BST’s S(X)

n

:= the mean path-length for the Trie [X = T]

  • r the mean symbol path-length of the BST [X = B]

when built on n words independently drawn from the same source.

slide-35
SLIDE 35

Exact average-case analysis for Tries or BST’s S(X)

n

:= the mean path-length for the Trie [X = T]

  • r the mean symbol path-length of the BST [X = B]

when built on n words independently drawn from the same source. For each case [X = T or X = B] an exact formula for S(X)

n

S(X)

n

=

n

  • k=2

(−1)k n k

  • ̟X(k)

which involves a series ̟X at integer values k.

Cl´ ement, Flajolet, V. (2001) for X = T, Cl´ ement, Fill, Flajolet, V. (2009) for X = B

slide-36
SLIDE 36

Exact average-case analysis for Tries or BST’s S(X)

n

:= the mean path-length for the Trie [X = T]

  • r the mean symbol path-length of the BST [X = B]

when built on n words independently drawn from the same source. For each case [X = T or X = B] an exact formula for S(X)

n

S(X)

n

=

n

  • k=2

(−1)k n k

  • ̟X(k)

which involves a series ̟X at integer values k.

Cl´ ement, Flajolet, V. (2001) for X = T, Cl´ ement, Fill, Flajolet, V. (2009) for X = B

This series ̟X(s) is closely related to the Dirichlet series of the source ̟T (s) = sΛ(s) ̟B(s) = 2 Λ(s) s(s − 1) where Λ(s) :=

  • w∈Σ⋆

ps

w

slide-37
SLIDE 37

Exact average-case analysis for Tries or BST’s S(X)

n

:= the mean path-length for the Trie [X = T]

  • r the mean symbol path-length of the BST [X = B]

when built on n words independently drawn from the same source. For each case [X = T or X = B] an exact formula for S(X)

n

S(X)

n

=

n

  • k=2

(−1)k n k

  • ̟X(k)

which involves a series ̟X at integer values k.

Cl´ ement, Flajolet, V. (2001) for X = T, Cl´ ement, Fill, Flajolet, V. (2009) for X = B

This series ̟X(s) is closely related to the Dirichlet series of the source ̟T (s) = sΛ(s) ̟B(s) = 2 Λ(s) s(s − 1) where Λ(s) :=

  • w∈Σ⋆

ps

w

Nice exact formulae, not easy to deal with, due to the alternating signs

slide-38
SLIDE 38

Asymptotic analysis. The residue formula transforms the sum into an integral with 1 < d < 2. Sn =

n

  • k=2

(−1)k n k

  • ̟(k) =

1 2iπ d+i∞

d−i∞

̟(s) n! (−1)n+1 s(s − 1) . . . (s − n)ds,

slide-39
SLIDE 39

Asymptotic analysis. The residue formula transforms the sum into an integral with 1 < d < 2. Sn =

n

  • k=2

(−1)k n k

  • ̟(k) =

1 2iπ d+i∞

d−i∞

̟(s) n! (−1)n+1 s(s − 1) . . . (s − n)ds,

We shift the integral on the left, Usually, the first singularities occur at ℜs = 1.

slide-40
SLIDE 40

Asymptotic analysis. The residue formula transforms the sum into an integral with 1 < d < 2. Sn =

n

  • k=2

(−1)k n k

  • ̟(k) =

1 2iπ d+i∞

d−i∞

̟(s) n! (−1)n+1 s(s − 1) . . . (s − n)ds,

We shift the integral on the left, Usually, the first singularities occur at ℜs = 1. Behaviour of ̟(s) [or Λ(s)] near ℜs = 1?

slide-41
SLIDE 41

Asymptotic analysis. The residue formula transforms the sum into an integral with 1 < d < 2. Sn =

n

  • k=2

(−1)k n k

  • ̟(k) =

1 2iπ d+i∞

d−i∞

̟(s) n! (−1)n+1 s(s − 1) . . . (s − n)ds,

We shift the integral on the left, Usually, the first singularities occur at ℜs = 1. Behaviour of ̟(s) [or Λ(s)] near ℜs = 1? Where are the red singularities closest to ℜs = 1? Is Λ(s) of polynomial growth on the green contour?

slide-42
SLIDE 42

Asymptotic analysis. The residue formula transforms the sum into an integral with 1 < d < 2. Sn =

n

  • k=2

(−1)k n k

  • ̟(k) =

1 2iπ d+i∞

d−i∞

̟(s) n! (−1)n+1 s(s − 1) . . . (s − n)ds,

We shift the integral on the left, Usually, the first singularities occur at ℜs = 1. Behaviour of ̟(s) [or Λ(s)] near ℜs = 1? Where are the red singularities closest to ℜs = 1? Is Λ(s) of polynomial growth on the green contour?

Importance of the existence of a region R – which contains only s = 1 as a pole – where Λ(s) is of polynomial growth. Tameness of the source

slide-43
SLIDE 43

Main results

[Cl´ ement, Flajolet, V. (2001), Cl´ ement, Flajolet, Fill, V. (2009)]

Consider n words independently drawn from the same tame source. Then: The mean path-length Tn

  • f the Trie satisfies

Tn ∼ 1 hS n log n. The mean symbol path-length Bn

  • f the BST satisfies

Bn ∼ 1 hS n log2 n. Here, hS is the entropy hS of the source S, defined as hS := lim

k→∞

 −1 k

  • w∈Σk

pw log pw   , where pw is the probability that a word begins with prefix w.

slide-44
SLIDE 44

Plan of the talk. – General motivations: Dirichlet generating functions and tameness – An important class of “natural” sources: dynamical sources = sources associated to dynamical systems – Tameness in the case of dynamical sources – Conclusion and possible extensions.

slide-45
SLIDE 45

A dynamical source = a source built with a dynamical system [V. 1998]

slide-46
SLIDE 46

A dynamical source = a source built with a dynamical system [V. 1998]

A dynamical system (I, T) is defined by – an alphabet Σ denumerable (possibly infinite), – a topological partition of I :=]0, 1[ with open intervals Im,m∈Σ, – an encoding mapping σ equal to m on each Im, – a shift mapping T – each T|Im is a bijection of class C2 on Im – The local inverse of T|Im is denoted by hm.

slide-47
SLIDE 47

A dynamical source = a source built with a dynamical system [V. 1998]

A dynamical system (I, T) is defined by – an alphabet Σ denumerable (possibly infinite), – a topological partition of I :=]0, 1[ with open intervals Im,m∈Σ, – an encoding mapping σ equal to m on each Im, – a shift mapping T – each T|Im is a bijection of class C2 on Im – The local inverse of T|Im is denoted by hm.

x T x T x

2

T x

3

M(x) = (c, b, a, c . . .) This gives rise to a source: On an input x of I, it outputs the word M(x) := (σx, σTx, σT 2x, . . . ). When an initial density is chosen on I, this induces (via M) a probabilistic model on Σ∞ = a dynamical source

slide-48
SLIDE 48

Strong relations between the geometry of the system,

the correlations between symbols and the probabilistic properties of the source. Two geometric characteristics of the system: – The position of the branches T(Ik) w.r.t Im – The shape of the branches defined by the derivative of hm

Particular cases: simple sources and affine branches

x 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 x 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0

slide-49
SLIDE 49

Strong relations between the geometry of the system,

the correlations between symbols and the probabilistic properties of the source. Two geometric characteristics of the system: – The position of the branches T(Ik) w.r.t Im – The shape of the branches defined by the derivative of hm

Particular cases: simple sources and affine branches A memoryless source

:= a complete system with affine branches and uniform initial density

A Markov chain

:= a Markovian system with affine branches, with an initial density which is constant on each Im.

x 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 x 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0

slide-50
SLIDE 50

Strong relations between the geometry of the system,

the correlations between symbols and the probabilistic properties of the source. Two geometric characteristics of the system: – The position of the branches T(Ik) w.r.t Im – The shape of the branches defined by the derivative of hm

Particular cases: simple sources and affine branches A memoryless source

:= a complete system with affine branches and uniform initial density

A Markov chain

:= a Markovian system with affine branches, with an initial density which is constant on each Im.

x 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 x 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0

slide-51
SLIDE 51

Strong relations between the geometry of the system,

the correlations between symbols and the probabilistic properties of the source. Two geometric characteristics of the system: – The position of the branches T(Ik) w.r.t Im – The shape of the branches defined by the derivative of hm

Particular cases: simple sources and affine branches A memoryless source

:= a complete system with affine branches and uniform initial density

A Markov chain

:= a Markovian system with affine branches, with an initial density which is constant on each Im.

x 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 x 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0

slide-52
SLIDE 52

General case of interest = the Good Class gathers – Complete systems: T(Im) = I – with a possible infinite denumerable alphabet – with expansive branches : |T ′(x)| ≥ ρ > 1.

slide-53
SLIDE 53

General case of interest = the Good Class gathers – Complete systems: T(Im) = I – with a possible infinite denumerable alphabet – with expansive branches : |T ′(x)| ≥ ρ > 1. Main instance: the Euclidean source defined with T(x) := 1 x − 1 x

slide-54
SLIDE 54

A main analytical object related to any source: the Dirichlet series of probabilities, Λ(s) :=

  • w∈Σ⋆

ps

w

slide-55
SLIDE 55

A main analytical object related to any source: the Dirichlet series of probabilities, Λ(s) :=

  • w∈Σ⋆

ps

w

Memoryless sources, with probabilities (pi) Λ(s) = 1 1 − λ(s) with λ(s) =

r

  • i=1

ps

i

slide-56
SLIDE 56

A main analytical object related to any source: the Dirichlet series of probabilities, Λ(s) :=

  • w∈Σ⋆

ps

w

Memoryless sources, with probabilities (pi) Λ(s) = 1 1 − λ(s) with λ(s) =

r

  • i=1

ps

i

Markov chains, defined by – the vector R of initial probabilities (ri) – and the transition matrix P := (pi,j) Λ(s) = 1 + t1(I − P(s))−1R(s) with P(s) = (ps

i,j),

R(s) = (rs

i ).

slide-57
SLIDE 57

A main analytical object related to any source: the Dirichlet series of probabilities, Λ(s) :=

  • w∈Σ⋆

ps

w

Memoryless sources, with probabilities (pi) Λ(s) = 1 1 − λ(s) with λ(s) =

r

  • i=1

ps

i

Markov chains, defined by – the vector R of initial probabilities (ri) – and the transition matrix P := (pi,j) Λ(s) = 1 + t1(I − P(s))−1R(s) with P(s) = (ps

i,j),

R(s) = (rs

i ).

A general dynamical source Λ(s) closely related to (I − Hs)−1 where Hs is the (secant) transfer operator of the dynamical system.

slide-58
SLIDE 58

The density transformer and the transfer operators

The operator H :=

  • m∈Σ

H[m] with H[m][f](x) = |h′

m(x)| · f ◦ hm(x)

is the density transformer of the dynamical system. It describes the evolution of the density. For a density f on [0, 1], H[f] is the density on [0, 1] after one iteration.

slide-59
SLIDE 59

The density transformer and the transfer operators

The operator H :=

  • m∈Σ

H[m] with H[m][f](x) = |h′

m(x)| · f ◦ hm(x)

is the density transformer of the dynamical system. It describes the evolution of the density. For a density f on [0, 1], H[f] is the density on [0, 1] after one iteration.

Transfer operator (Ruelle) [tangent version] Hs :=

  • m∈Σ

Hs,[m] with Hs,[m][f](x) = |h′

m(x)|s f ◦ ha(x).

slide-60
SLIDE 60

The density transformer and the transfer operators

The operator H :=

  • m∈Σ

H[m] with H[m][f](x) = |h′

m(x)| · f ◦ hm(x)

is the density transformer of the dynamical system. It describes the evolution of the density. For a density f on [0, 1], H[f] is the density on [0, 1] after one iteration.

Transfer operator (Ruelle) [tangent version] Hs :=

  • m∈Σ

Hs,[m] with Hs,[m][f](x) = |h′

m(x)|s f ◦ ha(x).

Transfer operator (Vall´ ee, 2000) [secant version] Hs :=

  • m∈Σ

Hs,[m] with Hs,[m][F](x, y) =

  • hm(x) − hm(y)

x − y

  • s

F(hm(x), hm(y))

slide-61
SLIDE 61

Alternative expression of Λ(s) in the dynamical case.

slide-62
SLIDE 62

Alternative expression of Λ(s) in the dynamical case. The Dirichlet series Λk(s) :=

  • w∈Σk

ps

w,

Λ(s) :=

  • w∈Σ⋆

ps

w

are “generated” by the secant transfer operator Hs [V. 2000] Λk(s) = Hk

s[Ls](0, 1),

Λ(s) = (I − Hs)−1[Ls](0, 1) with L the secant of the distribution function F.

slide-63
SLIDE 63

Alternative expression of Λ(s) in the dynamical case. The Dirichlet series Λk(s) :=

  • w∈Σk

ps

w,

Λ(s) :=

  • w∈Σ⋆

ps

w

are “generated” by the secant transfer operator Hs [V. 2000] Λk(s) = Hk

s[Ls](0, 1),

Λ(s) = (I − Hs)−1[Ls](0, 1) with L the secant of the distribution function F. Singularities of s → Λ(s) are essential in the analysis. Singularities of (I − Hs)−1 are related to spectral properties of Hs.

slide-64
SLIDE 64

Alternative expression of Λ(s) in the dynamical case. The Dirichlet series Λk(s) :=

  • w∈Σk

ps

w,

Λ(s) :=

  • w∈Σ⋆

ps

w

are “generated” by the secant transfer operator Hs [V. 2000] Λk(s) = Hk

s[Ls](0, 1),

Λ(s) = (I − Hs)−1[Ls](0, 1) with L the secant of the distribution function F. Singularities of s → Λ(s) are essential in the analysis. Singularities of (I − Hs)−1 are related to spectral properties of Hs. For s = 1, H1 is an extension of H and has an eigenvalue equal to 1. For a system of the Good Class, s → Λ(s) has a simple pole at s = 1

slide-65
SLIDE 65

Plan of the talk. – General motivations: Dirichlet generating functions and tameness – An important class of sources: dynamical sources. – Tameness of dynamical sources – Conclusion and possible extensions.

slide-66
SLIDE 66

What happens on the left of the vertical line ℜs = 1? It is important for the analysis to deal with a region R where Λ(s) is tame – it is analytic (except for s = 1) and of polynomial growth (ℑs → ∞)

slide-67
SLIDE 67

What happens on the left of the vertical line ℜs = 1? It is important for the analysis to deal with a region R where Λ(s) is tame – it is analytic (except for s = 1) and of polynomial growth (ℑs → ∞) Different possible regions R on the left of ℜs = 1 where Λ(s) is tame.

slide-68
SLIDE 68

What happens on the left of the vertical line ℜs = 1? It is important for the analysis to deal with a region R where Λ(s) is tame – it is analytic (except for s = 1) and of polynomial growth (ℑs → ∞) Different possible regions R on the left of ℜs = 1 where Λ(s) is tame.

Situation 1 Situation 2 Situation 3 Vertical strip Hyperbolic region Vertical strip with holes 1 − σ ≤ a 1 − σ ≤ t−α

slide-69
SLIDE 69

Different possible regions on the left of ℜs = 1 where Λ(s) is tame.

Situation 1 Situation 2 Situation 3 Vertical strip Hyperbolic region Vertical strip with holes

slide-70
SLIDE 70

Different possible regions on the left of ℜs = 1 where Λ(s) is tame.

Situation 1 Situation 2 Situation 3 Vertical strip Hyperbolic region Vertical strip with holes

For which simple sources do these different situations occur?

slide-71
SLIDE 71

Different possible regions on the left of ℜs = 1 where Λ(s) is tame.

Situation 1 Situation 2 Situation 3 Vertical strip Hyperbolic region Vertical strip with holes

For which simple sources do these different situations occur? For memoryless sources relative to probabilities (p1, p2, . . . , pr) – S1 is impossible – S3 occurs when all the ratios log pi/log pj are rational – S2 occurs if there exists a ratio log pi/log pj which is “diophantine” [badly approximable by rationals]

slide-72
SLIDE 72

Memoryless sources Λ(s) = 1 1 − λ(s) with λ(s) = ps

1 + ps 2

[r = 2]

slide-73
SLIDE 73

Memoryless sources Λ(s) = 1 1 − λ(s) with λ(s) = ps

1 + ps 2

[r = 2] The tameness of Λ depends on arithmetical properties of log p2/log p1 which influence Z := the set of poles on ℜs = 1, s = 1

slide-74
SLIDE 74

Memoryless sources Λ(s) = 1 1 − λ(s) with λ(s) = ps

1 + ps 2

[r = 2] The tameness of Λ depends on arithmetical properties of log p2/log p1 which influence Z := the set of poles on ℜs = 1, s = 1

(i) Z = ∅ ⇐ ⇒ log p2/log p1 is rational (ii) If Z = ∅, then the poles of Λ(s) close to ℜs = 1 are created by good rational approximations of log p2/log p1

slide-75
SLIDE 75

Memoryless sources Λ(s) = 1 1 − λ(s) with λ(s) = ps

1 + ps 2

[r = 2] The tameness of Λ depends on arithmetical properties of log p2/log p1 which influence Z := the set of poles on ℜs = 1, s = 1

(i) Z = ∅ ⇐ ⇒ log p2/log p1 is rational (ii) If Z = ∅, then the poles of Λ(s) close to ℜs = 1 are created by good rational approximations of log p2/log p1 The irrationality exponent µ(x) of a number x equals µ if, for any ν > µ, the set of pairs (a, b) ∈ Z2 for which

  • x − a

b

  • ≤ 1

bν is finite x diophantine ⇐ ⇒ µ(x) < ∞

slide-76
SLIDE 76

Memoryless sources Λ(s) = 1 1 − λ(s) with λ(s) = ps

1 + ps 2

[r = 2] The tameness of Λ depends on arithmetical properties of log p2/log p1 which influence Z := the set of poles on ℜs = 1, s = 1

(i) Z = ∅ ⇐ ⇒ log p2/log p1 is rational (ii) If Z = ∅, then the poles of Λ(s) close to ℜs = 1 are created by good rational approximations of log p2/log p1 The irrationality exponent µ(x) of a number x equals µ if, for any ν > µ, the set of pairs (a, b) ∈ Z2 for which

  • x − a

b

  • ≤ 1

bν is finite x diophantine ⇐ ⇒ µ(x) < ∞ The shape of the tameness region is related to µ(log p2/ log p1). If µ(log p2/ log p1) = µ then, for any θ, ν with θ < µ < ν, the tameness region is as shown: [Flajolet-Roux-V. 2010]

slide-77
SLIDE 77

Different possible regions on the left of ℜs = 1 where Λ(s) is tame.

slide-78
SLIDE 78

Different possible regions on the left of ℜs = 1 where Λ(s) is tame.

Situation 1 Situation 2 Situation 3 Vertical strip Hyperbolic region Vertical strip with holes Geometric condition Arithmetic condition Periodicity condition

For which general dynamical sources do these different situations occur?

slide-79
SLIDE 79

Different possible regions on the left of ℜs = 1 where Λ(s) is tame.

Situation 1 Situation 2 Situation 3 Vertical strip Hyperbolic region Vertical strip with holes Geometric condition Arithmetic condition Periodicity condition

For which general dynamical sources do these different situations occur? – S1 occurs when “the branches are not too often of the same shape”. – S3 occurs only if the source is conjugated to a simple source. – S2 occurs if a extension of the following condition holds: “there exists a ratio log pi/log pj which is “diophantine”

slide-80
SLIDE 80

Situation 1- Existence of a vertical strip where Λ(s) is tame The condition UNI expresses that “the branches of the dynamical system are not too often of the same shape”

slide-81
SLIDE 81

Situation 1- Existence of a vertical strip where Λ(s) is tame The condition UNI expresses that “the branches of the dynamical system are not too often of the same shape” Theorem [Dolgopyat-Baladi-Cesaratto-V]. For a good dynamical system which satisfies the condition UNI, there exists a vertical strip where Λ(s) is tame.

slide-82
SLIDE 82

Situation 1- Existence of a vertical strip where Λ(s) is tame The condition UNI expresses that “the branches of the dynamical system are not too often of the same shape” Theorem [Dolgopyat-Baladi-Cesaratto-V]. For a good dynamical system which satisfies the condition UNI, there exists a vertical strip where Λ(s) is tame. Dolgopyat (98) proves the result for the plain transfer operator, in the case

  • f a finite number of branches

– Baladi and V. (03) extend the result for an infinite number of branches – Cesaratto and V. (09) extend the result to the secant transfer operator.

slide-83
SLIDE 83

Situation 2- Existence of a hyperbolic region where Λ(s) is tame The condition DIOP extends the arithmetic condition “There exists a ratio log pi/ log pj which is diophantine”

For a complete system, each branch h has a fixed point denoted by h⋆. The derivatives |h′(h⋆)| replace the probabilities of the memoryless case.

slide-84
SLIDE 84

Situation 2- Existence of a hyperbolic region where Λ(s) is tame The condition DIOP extends the arithmetic condition “There exists a ratio log pi/ log pj which is diophantine”

For a complete system, each branch h has a fixed point denoted by h⋆. The derivatives |h′(h⋆)| replace the probabilities of the memoryless case.

DIOP: There exists a ratio c(h, k) := log |h′(h⋆)| log |k′(k⋆)| which is diophantine.

slide-85
SLIDE 85

Situation 2- Existence of a hyperbolic region where Λ(s) is tame The condition DIOP extends the arithmetic condition “There exists a ratio log pi/ log pj which is diophantine”

For a complete system, each branch h has a fixed point denoted by h⋆. The derivatives |h′(h⋆)| replace the probabilities of the memoryless case.

DIOP: There exists a ratio c(h, k) := log |h′(h⋆)| log |k′(k⋆)| which is diophantine. Theorem [Dolgopyat-Roux-V.] For a good dynamical system which satisfies the condition DIOP, there exists an hyperbolic region where Λ(s) is tame.

slide-86
SLIDE 86

Situation 2- Existence of a hyperbolic region where Λ(s) is tame The condition DIOP extends the arithmetic condition “There exists a ratio log pi/ log pj which is diophantine”

For a complete system, each branch h has a fixed point denoted by h⋆. The derivatives |h′(h⋆)| replace the probabilities of the memoryless case.

DIOP: There exists a ratio c(h, k) := log |h′(h⋆)| log |k′(k⋆)| which is diophantine. Theorem [Dolgopyat-Roux-V.] For a good dynamical system which satisfies the condition DIOP, there exists an hyperbolic region where Λ(s) is tame. Dolgopyat (98) proves the result for the plain transfer operator, in the case

  • f a finite number of branches – Roux and V. (2010) extend the result :

for an infinite number of branches and for the secant transfer operator.

slide-87
SLIDE 87

Plan of the talk. – General motivations: Dirichlet generating functions and ttameness – An important class of sources: dynamical sources. – Tameness of dynamical sources – Conclusion and possible extensions.

slide-88
SLIDE 88

Conclusions. Description of a framework which – unifies the analyses for text algorithms and searching/sorting algorithms

slide-89
SLIDE 89

Conclusions. Description of a framework which – unifies the analyses for text algorithms and searching/sorting algorithms – provides a general model for sources – shows the importance of the Dirichlet generating functions – explains the importance of tameness for sources

slide-90
SLIDE 90

Conclusions. Description of a framework which – unifies the analyses for text algorithms and searching/sorting algorithms – provides a general model for sources – shows the importance of the Dirichlet generating functions – explains the importance of tameness for sources – defines a natural subclass of sources, the dynamical sources – provides sufficient conditions for tameness of dynamical sources

slide-91
SLIDE 91

Conclusions. Description of a framework which – unifies the analyses for text algorithms and searching/sorting algorithms – provides a general model for sources – shows the importance of the Dirichlet generating functions – explains the importance of tameness for sources – defines a natural subclass of sources, the dynamical sources – provides sufficient conditions for tameness of dynamical sources – provides probabilistic analyses for algorithms built on tame sources.

slide-92
SLIDE 92

Possible extensions and work in progress I– Classification of sources

slide-93
SLIDE 93

Possible extensions and work in progress I– Classification of sources – Place of dynamical sources amongst general sources: – A dynamical source = limit of Markov chains with increasing order? – Comparing dynamical sources with Markov chains of variable length

slide-94
SLIDE 94

Possible extensions and work in progress I– Classification of sources – Place of dynamical sources amongst general sources: – A dynamical source = limit of Markov chains with increasing order? – Comparing dynamical sources with Markov chains of variable length

slide-95
SLIDE 95

Possible extensions and work in progress II– Realistic analyses of other algorithms and other structures – Analysis of other sorting algorithms – Analysis of Insertion Sort easy – Analysis of QuickSelect already done – And Selection algorithm ?

slide-96
SLIDE 96

Possible extensions and work in progress II– Realistic analyses of other algorithms and other structures – Analysis of other sorting algorithms – Analysis of Insertion Sort easy – Analysis of QuickSelect already done – And Selection algorithm ? – Analysis of the DST structure?

bbabb abbaa babba aaaab ababb abbba

1: 3: 5: 4: 6: 2: