Kolmogorov complexity as a language Alexander Shen LIF CNRS, - - PowerPoint PPT Presentation

kolmogorov complexity as a language
SMART_READER_LITE
LIVE PREVIEW

Kolmogorov complexity as a language Alexander Shen LIF CNRS, - - PowerPoint PPT Presentation

CSR-2011 Kolmogorov complexity as a language Alexander Shen LIF CNRS, Marseille; on leave from , . . . . . . A powerful tool Just a way to reformulate arguments three languages:


slide-1
SLIDE 1

. . . . . .

CSR-2011

Kolmogorov complexity as a language

Alexander Shen LIF CNRS, Marseille; on leave from ИППИ РАН, Москва

slide-2
SLIDE 2

. . . . . .

Kolmogorov complexity

A powerful tool Just a way to reformulate arguments three languages: combinatorial/algorithmic/probabilistic

slide-3
SLIDE 3

. . . . . .

Kolmogorov complexity

◮ A powerful tool

Just a way to reformulate arguments three languages: combinatorial/algorithmic/probabilistic

slide-4
SLIDE 4

. . . . . .

Kolmogorov complexity

◮ A powerful tool ◮ Just a way to reformulate arguments

three languages: combinatorial/algorithmic/probabilistic

slide-5
SLIDE 5

. . . . . .

Kolmogorov complexity

◮ A powerful tool ◮ Just a way to reformulate arguments ◮ three languages: combinatorial/algorithmic/probabilistic

slide-6
SLIDE 6

. . . . . .

Kolmogorov complexity

◮ A powerful tool ◮ Just a way to reformulate arguments ◮ three languages: combinatorial/algorithmic/probabilistic

slide-7
SLIDE 7

. . . . . .

Reminder and notation

K x minimal length of a program that produces x KD x min p D p x depends on the interpreter D

  • ptimal D makes it minimal up to O

additive term Variations: p = string p = prefix of a sequence x = string plain prefix K x , C x KP x , K x x = prefix decision monotone

  • f a sequence

KR x , KD x KM x , Km x Conditional complexity C x y : minimal length of a program p y x.

There is also a priori probability (in two versions: discrete, on strings; continuous, on prefixes)

slide-8
SLIDE 8

. . . . . .

Reminder and notation

◮ K(x) = minimal length of a program that produces x

KD x min p D p x depends on the interpreter D

  • ptimal D makes it minimal up to O

additive term Variations: p = string p = prefix of a sequence x = string plain prefix K x , C x KP x , K x x = prefix decision monotone

  • f a sequence

KR x , KD x KM x , Km x Conditional complexity C x y : minimal length of a program p y x.

There is also a priori probability (in two versions: discrete, on strings; continuous, on prefixes)

slide-9
SLIDE 9

. . . . . .

Reminder and notation

◮ K(x) = minimal length of a program that produces x ◮ KD(x) = min{|p| : D(p) = x}

depends on the interpreter D

  • ptimal D makes it minimal up to O

additive term Variations: p = string p = prefix of a sequence x = string plain prefix K x , C x KP x , K x x = prefix decision monotone

  • f a sequence

KR x , KD x KM x , Km x Conditional complexity C x y : minimal length of a program p y x.

There is also a priori probability (in two versions: discrete, on strings; continuous, on prefixes)

slide-10
SLIDE 10

. . . . . .

Reminder and notation

◮ K(x) = minimal length of a program that produces x ◮ KD(x) = min{|p| : D(p) = x} ◮ depends on the interpreter D

  • ptimal D makes it minimal up to O

additive term Variations: p = string p = prefix of a sequence x = string plain prefix K x , C x KP x , K x x = prefix decision monotone

  • f a sequence

KR x , KD x KM x , Km x Conditional complexity C x y : minimal length of a program p y x.

There is also a priori probability (in two versions: discrete, on strings; continuous, on prefixes)

slide-11
SLIDE 11

. . . . . .

Reminder and notation

◮ K(x) = minimal length of a program that produces x ◮ KD(x) = min{|p| : D(p) = x} ◮ depends on the interpreter D ◮ optimal D makes it minimal up to O(1) additive term

Variations: p = string p = prefix of a sequence x = string plain prefix K x , C x KP x , K x x = prefix decision monotone

  • f a sequence

KR x , KD x KM x , Km x Conditional complexity C x y : minimal length of a program p y x.

There is also a priori probability (in two versions: discrete, on strings; continuous, on prefixes)

slide-12
SLIDE 12

. . . . . .

Reminder and notation

◮ K(x) = minimal length of a program that produces x ◮ KD(x) = min{|p| : D(p) = x} ◮ depends on the interpreter D ◮ optimal D makes it minimal up to O(1) additive term ◮ Variations:

p = string p = prefix of a sequence x = string plain prefix K x , C x KP x , K x x = prefix decision monotone

  • f a sequence

KR x , KD x KM x , Km x Conditional complexity C x y : minimal length of a program p y x.

There is also a priori probability (in two versions: discrete, on strings; continuous, on prefixes)

slide-13
SLIDE 13

. . . . . .

Reminder and notation

◮ K(x) = minimal length of a program that produces x ◮ KD(x) = min{|p| : D(p) = x} ◮ depends on the interpreter D ◮ optimal D makes it minimal up to O(1) additive term ◮ Variations:

p = string p = prefix of a sequence x = string plain prefix K(x), C(x) KP(x), K(x) x = prefix decision monotone

  • f a sequence

KR(x), KD(x) KM(x), Km(x) Conditional complexity C x y : minimal length of a program p y x.

There is also a priori probability (in two versions: discrete, on strings; continuous, on prefixes)

slide-14
SLIDE 14

. . . . . .

Reminder and notation

◮ K(x) = minimal length of a program that produces x ◮ KD(x) = min{|p| : D(p) = x} ◮ depends on the interpreter D ◮ optimal D makes it minimal up to O(1) additive term ◮ Variations:

p = string p = prefix of a sequence x = string plain prefix K(x), C(x) KP(x), K(x) x = prefix decision monotone

  • f a sequence

KR(x), KD(x) KM(x), Km(x) Conditional complexity C(x|y): minimal length of a program p : y → x.

There is also a priori probability (in two versions: discrete, on strings; continuous, on prefixes)

slide-15
SLIDE 15

. . . . . .

Reminder and notation

◮ K(x) = minimal length of a program that produces x ◮ KD(x) = min{|p| : D(p) = x} ◮ depends on the interpreter D ◮ optimal D makes it minimal up to O(1) additive term ◮ Variations:

p = string p = prefix of a sequence x = string plain prefix K(x), C(x) KP(x), K(x) x = prefix decision monotone

  • f a sequence

KR(x), KD(x) KM(x), Km(x) Conditional complexity C(x|y): minimal length of a program p : y → x.

There is also a priori probability (in two versions: discrete, on strings; continuous, on prefixes)

slide-16
SLIDE 16

. . . . . .

Foundations of probability theory

Random object or random process? “well shuffled deck of cards”: any meaning? [xkcd cartoon] randomness = incompressibility (maximal complexity) is random iff KM

n

n O Classical probability theory: random sequence satisfies the Strong Law of Large Numbers with probability 1 Algorithmic version: every (algorithmically) random sequence satisfies SLLN algorithmic classical: Martin-Löf random sequences form a set of measure .

slide-17
SLIDE 17

. . . . . .

Foundations of probability theory

◮ Random object or random process?

“well shuffled deck of cards”: any meaning? [xkcd cartoon] randomness = incompressibility (maximal complexity) is random iff KM

n

n O Classical probability theory: random sequence satisfies the Strong Law of Large Numbers with probability 1 Algorithmic version: every (algorithmically) random sequence satisfies SLLN algorithmic classical: Martin-Löf random sequences form a set of measure .

slide-18
SLIDE 18

. . . . . .

Foundations of probability theory

◮ Random object or random process? ◮ “well shuffled deck of cards”: any meaning?

[xkcd cartoon] randomness = incompressibility (maximal complexity) is random iff KM

n

n O Classical probability theory: random sequence satisfies the Strong Law of Large Numbers with probability 1 Algorithmic version: every (algorithmically) random sequence satisfies SLLN algorithmic classical: Martin-Löf random sequences form a set of measure .

slide-19
SLIDE 19

. . . . . .

Foundations of probability theory

◮ Random object or random process? ◮ “well shuffled deck of cards”: any meaning?

[xkcd cartoon] randomness = incompressibility (maximal complexity) is random iff KM

n

n O Classical probability theory: random sequence satisfies the Strong Law of Large Numbers with probability 1 Algorithmic version: every (algorithmically) random sequence satisfies SLLN algorithmic classical: Martin-Löf random sequences form a set of measure .

slide-20
SLIDE 20

. . . . . .

Foundations of probability theory

◮ Random object or random process? ◮ “well shuffled deck of cards”: any meaning?

[xkcd cartoon]

◮ randomness = incompressibility (maximal complexity)

is random iff KM

n

n O Classical probability theory: random sequence satisfies the Strong Law of Large Numbers with probability 1 Algorithmic version: every (algorithmically) random sequence satisfies SLLN algorithmic classical: Martin-Löf random sequences form a set of measure .

slide-21
SLIDE 21

. . . . . .

Foundations of probability theory

◮ Random object or random process? ◮ “well shuffled deck of cards”: any meaning?

[xkcd cartoon]

◮ randomness = incompressibility (maximal complexity) ◮ ω = ω1ω2 . . . is random iff KM(ω1 . . . ωn) ≥ n − O(1)

Classical probability theory: random sequence satisfies the Strong Law of Large Numbers with probability 1 Algorithmic version: every (algorithmically) random sequence satisfies SLLN algorithmic classical: Martin-Löf random sequences form a set of measure .

slide-22
SLIDE 22

. . . . . .

Foundations of probability theory

◮ Random object or random process? ◮ “well shuffled deck of cards”: any meaning?

[xkcd cartoon]

◮ randomness = incompressibility (maximal complexity) ◮ ω = ω1ω2 . . . is random iff KM(ω1 . . . ωn) ≥ n − O(1) ◮ Classical probability theory: random sequence satisfies the

Strong Law of Large Numbers with probability 1 Algorithmic version: every (algorithmically) random sequence satisfies SLLN algorithmic classical: Martin-Löf random sequences form a set of measure .

slide-23
SLIDE 23

. . . . . .

Foundations of probability theory

◮ Random object or random process? ◮ “well shuffled deck of cards”: any meaning?

[xkcd cartoon]

◮ randomness = incompressibility (maximal complexity) ◮ ω = ω1ω2 . . . is random iff KM(ω1 . . . ωn) ≥ n − O(1) ◮ Classical probability theory: random sequence satisfies the

Strong Law of Large Numbers with probability 1

◮ Algorithmic version: every (algorithmically) random

sequence satisfies SLLN algorithmic classical: Martin-Löf random sequences form a set of measure .

slide-24
SLIDE 24

. . . . . .

Foundations of probability theory

◮ Random object or random process? ◮ “well shuffled deck of cards”: any meaning?

[xkcd cartoon]

◮ randomness = incompressibility (maximal complexity) ◮ ω = ω1ω2 . . . is random iff KM(ω1 . . . ωn) ≥ n − O(1) ◮ Classical probability theory: random sequence satisfies the

Strong Law of Large Numbers with probability 1

◮ Algorithmic version: every (algorithmically) random

sequence satisfies SLLN

◮ algorithmic ⇒ classical: Martin-Löf random sequences form

a set of measure 1.

slide-25
SLIDE 25

. . . . . .

Sampling random strings (S.Aaronson)

A device that (being switched on) produces N-bit string and stops “The device produces a random string”: what does it mean? classical: the output distribution is close to the uniform one effective: with high probability the output string is incompressible not equivalent if no assumptions about the device but are related under some assumptions

slide-26
SLIDE 26

. . . . . .

Sampling random strings (S.Aaronson)

◮ A device that (being switched on) produces N-bit string and

stops “The device produces a random string”: what does it mean? classical: the output distribution is close to the uniform one effective: with high probability the output string is incompressible not equivalent if no assumptions about the device but are related under some assumptions

slide-27
SLIDE 27

. . . . . .

Sampling random strings (S.Aaronson)

◮ A device that (being switched on) produces N-bit string and

stops

◮ “The device produces a random string”: what does it mean?

classical: the output distribution is close to the uniform one effective: with high probability the output string is incompressible not equivalent if no assumptions about the device but are related under some assumptions

slide-28
SLIDE 28

. . . . . .

Sampling random strings (S.Aaronson)

◮ A device that (being switched on) produces N-bit string and

stops

◮ “The device produces a random string”: what does it mean? ◮ classical: the output distribution is close to the uniform one

effective: with high probability the output string is incompressible not equivalent if no assumptions about the device but are related under some assumptions

slide-29
SLIDE 29

. . . . . .

Sampling random strings (S.Aaronson)

◮ A device that (being switched on) produces N-bit string and

stops

◮ “The device produces a random string”: what does it mean? ◮ classical: the output distribution is close to the uniform one ◮ effective: with high probability the output string is

incompressible not equivalent if no assumptions about the device but are related under some assumptions

slide-30
SLIDE 30

. . . . . .

Sampling random strings (S.Aaronson)

◮ A device that (being switched on) produces N-bit string and

stops

◮ “The device produces a random string”: what does it mean? ◮ classical: the output distribution is close to the uniform one ◮ effective: with high probability the output string is

incompressible

◮ not equivalent if no assumptions about the device

but are related under some assumptions

slide-31
SLIDE 31

. . . . . .

Sampling random strings (S.Aaronson)

◮ A device that (being switched on) produces N-bit string and

stops

◮ “The device produces a random string”: what does it mean? ◮ classical: the output distribution is close to the uniform one ◮ effective: with high probability the output string is

incompressible

◮ not equivalent if no assumptions about the device ◮ but are related under some assumptions

slide-32
SLIDE 32

. . . . . .

Example: matrices without uniform minors

k k minor of n n Boolean matrix: select k rows and k columns minor is uniform if it is all-0 or all-1. claim: there is a n n bit matrix without k k uniform minors for k log n.

slide-33
SLIDE 33

. . . . . .

Example: matrices without uniform minors

◮ k × k minor of n × n Boolean matrix: select k rows and k

columns minor is uniform if it is all-0 or all-1. claim: there is a n n bit matrix without k k uniform minors for k log n.

slide-34
SLIDE 34

. . . . . .

Example: matrices without uniform minors

◮ k × k minor of n × n Boolean matrix: select k rows and k

columns

◮ minor is uniform if it is all-0 or all-1. ◮ claim: there is a n × n bit matrix without k × k uniform

minors for k = 3 log n.

slide-35
SLIDE 35

. . . . . .

Counting argument and complexity reformulation

nk nk positions of the minor [k log n] types of uniform minors (0/1)

n k possibilities for the rest

n k

n k = log n log n n log n n .

log n bits to specify a column or row: k log n bits in total

  • ne additional bit to specify the type of minor ( / )

n k bits to specify the rest of the matrix k log n n k log n n log n n .

slide-36
SLIDE 36

. . . . . .

Counting argument and complexity reformulation

◮ ≤ nk × nk positions of the minor [k = 3 log n]

types of uniform minors (0/1)

n k possibilities for the rest

n k

n k = log n log n n log n n .

log n bits to specify a column or row: k log n bits in total

  • ne additional bit to specify the type of minor ( / )

n k bits to specify the rest of the matrix k log n n k log n n log n n .

slide-37
SLIDE 37

. . . . . .

Counting argument and complexity reformulation

◮ ≤ nk × nk positions of the minor [k = 3 log n] ◮ 2 types of uniform minors (0/1) n k possibilities for the rest

n k

n k = log n log n n log n n .

log n bits to specify a column or row: k log n bits in total

  • ne additional bit to specify the type of minor ( / )

n k bits to specify the rest of the matrix k log n n k log n n log n n .

slide-38
SLIDE 38

. . . . . .

Counting argument and complexity reformulation

◮ ≤ nk × nk positions of the minor [k = 3 log n] ◮ 2 types of uniform minors (0/1) ◮ 2n2−k2 possibilities for the rest

n k

n k = log n log n n log n n .

log n bits to specify a column or row: k log n bits in total

  • ne additional bit to specify the type of minor ( / )

n k bits to specify the rest of the matrix k log n n k log n n log n n .

slide-39
SLIDE 39

. . . . . .

Counting argument and complexity reformulation

◮ ≤ nk × nk positions of the minor [k = 3 log n] ◮ 2 types of uniform minors (0/1) ◮ 2n2−k2 possibilities for the rest ◮ n2k × 2 × 2n2−k2 = log n log n n log n n .

log n bits to specify a column or row: k log n bits in total

  • ne additional bit to specify the type of minor ( / )

n k bits to specify the rest of the matrix k log n n k log n n log n n .

slide-40
SLIDE 40

. . . . . .

Counting argument and complexity reformulation

◮ ≤ nk × nk positions of the minor [k = 3 log n] ◮ 2 types of uniform minors (0/1) ◮ 2n2−k2 possibilities for the rest ◮ n2k × 2 × 2n2−k2 = 2log n×2×3 log n+1+(n2−9 log2 n) < 2n2.

log n bits to specify a column or row: k log n bits in total

  • ne additional bit to specify the type of minor ( / )

n k bits to specify the rest of the matrix k log n n k log n n log n n .

slide-41
SLIDE 41

. . . . . .

Counting argument and complexity reformulation

◮ ≤ nk × nk positions of the minor [k = 3 log n] ◮ 2 types of uniform minors (0/1) ◮ 2n2−k2 possibilities for the rest ◮ n2k × 2 × 2n2−k2 = 2log n×2×3 log n+1+(n2−9 log2 n) < 2n2. ◮ log n bits to specify a column or row: 2k log n bits in total

  • ne additional bit to specify the type of minor ( / )

n k bits to specify the rest of the matrix k log n n k log n n log n n .

slide-42
SLIDE 42

. . . . . .

Counting argument and complexity reformulation

◮ ≤ nk × nk positions of the minor [k = 3 log n] ◮ 2 types of uniform minors (0/1) ◮ 2n2−k2 possibilities for the rest ◮ n2k × 2 × 2n2−k2 = 2log n×2×3 log n+1+(n2−9 log2 n) < 2n2. ◮ log n bits to specify a column or row: 2k log n bits in total ◮ one additional bit to specify the type of minor (0/1)

n k bits to specify the rest of the matrix k log n n k log n n log n n .

slide-43
SLIDE 43

. . . . . .

Counting argument and complexity reformulation

◮ ≤ nk × nk positions of the minor [k = 3 log n] ◮ 2 types of uniform minors (0/1) ◮ 2n2−k2 possibilities for the rest ◮ n2k × 2 × 2n2−k2 = 2log n×2×3 log n+1+(n2−9 log2 n) < 2n2. ◮ log n bits to specify a column or row: 2k log n bits in total ◮ one additional bit to specify the type of minor (0/1) ◮ n2 − k2 bits to specify the rest of the matrix

k log n n k log n n log n n .

slide-44
SLIDE 44

. . . . . .

Counting argument and complexity reformulation

◮ ≤ nk × nk positions of the minor [k = 3 log n] ◮ 2 types of uniform minors (0/1) ◮ 2n2−k2 possibilities for the rest ◮ n2k × 2 × 2n2−k2 = 2log n×2×3 log n+1+(n2−9 log2 n) < 2n2. ◮ log n bits to specify a column or row: 2k log n bits in total ◮ one additional bit to specify the type of minor (0/1) ◮ n2 − k2 bits to specify the rest of the matrix ◮ 2k log n + 1 + (n2 − k2) = 6 log2 n + 1 + (n2 − 9 log2 n) < n2.

slide-45
SLIDE 45

. . . . . .

One-tape Turing machines

copying n-bit string on 1-tape TM requires n time complexity version: if initially the tape was empty on the right of the border, then after n steps the complexity of a zone that is d cells far from the border is O n d . K u t O n d

proof: border guards in each cell of the border security zone write down the contents of the head of TM; each of the records is enough to reconstruct u t so the length of it should be K u t ; the sum of lengths does not exceed time

slide-46
SLIDE 46

. . . . . .

One-tape Turing machines

◮ copying n-bit string on 1-tape TM requires Ω(n2) time

complexity version: if initially the tape was empty on the right of the border, then after n steps the complexity of a zone that is d cells far from the border is O n d . K u t O n d

proof: border guards in each cell of the border security zone write down the contents of the head of TM; each of the records is enough to reconstruct u t so the length of it should be K u t ; the sum of lengths does not exceed time

slide-47
SLIDE 47

. . . . . .

One-tape Turing machines

◮ copying n-bit string on 1-tape TM requires Ω(n2) time ◮ complexity version: if initially the tape was empty on the

right of the border, then after n steps the complexity of a zone that is d cells far from the border is O(n/d). K(u(t)) ≤ O(n/d)

proof: border guards in each cell of the border security zone write down the contents of the head of TM; each of the records is enough to reconstruct u t so the length of it should be K u t ; the sum of lengths does not exceed time

slide-48
SLIDE 48

. . . . . .

One-tape Turing machines

◮ copying n-bit string on 1-tape TM requires Ω(n2) time ◮ complexity version: if initially the tape was empty on the

right of the border, then after n steps the complexity of a zone that is d cells far from the border is O(n/d). K(u(t)) ≤ O(n/d)

◮ proof: border guards in each cell of the border security zone write down the contents of the head of TM; each of the records is enough to reconstruct u(t) so the length of it should be Ω(K(u(t)); the sum of lengths does not exceed time

slide-49
SLIDE 49

. . . . . .

Everywhere complex sequences

Random sequence has n-bit prefix of complexity n but some factors (substrings) have small complexity Levin: there exist everywhere complex sequences: every n-bit substring has complexity n O Combinatorial equivalent: Let F be a set of strings that has at most

n strings of length n. Then there is a sequence

s.t. all sufficiently long substrings of are not in F. combinatorial and complexity proofs not just translations of each other (Lovasz lemma, Rumyantsev, Miller, Muchnik)

slide-50
SLIDE 50

. . . . . .

Everywhere complex sequences

◮ Random sequence has n-bit prefix of complexity n

but some factors (substrings) have small complexity Levin: there exist everywhere complex sequences: every n-bit substring has complexity n O Combinatorial equivalent: Let F be a set of strings that has at most

n strings of length n. Then there is a sequence

s.t. all sufficiently long substrings of are not in F. combinatorial and complexity proofs not just translations of each other (Lovasz lemma, Rumyantsev, Miller, Muchnik)

slide-51
SLIDE 51

. . . . . .

Everywhere complex sequences

◮ Random sequence has n-bit prefix of complexity n ◮ but some factors (substrings) have small complexity

Levin: there exist everywhere complex sequences: every n-bit substring has complexity n O Combinatorial equivalent: Let F be a set of strings that has at most

n strings of length n. Then there is a sequence

s.t. all sufficiently long substrings of are not in F. combinatorial and complexity proofs not just translations of each other (Lovasz lemma, Rumyantsev, Miller, Muchnik)

slide-52
SLIDE 52

. . . . . .

Everywhere complex sequences

◮ Random sequence has n-bit prefix of complexity n ◮ but some factors (substrings) have small complexity ◮ Levin: there exist everywhere complex sequences: every

n-bit substring has complexity 0.99n − O(1) Combinatorial equivalent: Let F be a set of strings that has at most

n strings of length n. Then there is a sequence

s.t. all sufficiently long substrings of are not in F. combinatorial and complexity proofs not just translations of each other (Lovasz lemma, Rumyantsev, Miller, Muchnik)

slide-53
SLIDE 53

. . . . . .

Everywhere complex sequences

◮ Random sequence has n-bit prefix of complexity n ◮ but some factors (substrings) have small complexity ◮ Levin: there exist everywhere complex sequences: every

n-bit substring has complexity 0.99n − O(1)

◮ Combinatorial equivalent: Let F be a set of strings that has at

most 20.99n strings of length n. Then there is a sequence ω s.t. all sufficiently long substrings of ω are not in F. combinatorial and complexity proofs not just translations of each other (Lovasz lemma, Rumyantsev, Miller, Muchnik)

slide-54
SLIDE 54

. . . . . .

Everywhere complex sequences

◮ Random sequence has n-bit prefix of complexity n ◮ but some factors (substrings) have small complexity ◮ Levin: there exist everywhere complex sequences: every

n-bit substring has complexity 0.99n − O(1)

◮ Combinatorial equivalent: Let F be a set of strings that has at

most 20.99n strings of length n. Then there is a sequence ω s.t. all sufficiently long substrings of ω are not in F.

◮ combinatorial and complexity proofs not just translations of

each other (Lovasz lemma, Rumyantsev, Miller, Muchnik)

slide-55
SLIDE 55

. . . . . .

Gilbert-Varshamov complexity bound

coding theory: how many n-bit strings x xk one can find if Hamming distance between every two is at least d lower bound (Gilbert–Varshamov) then d changed bits are harmless but bit insertion or deletions could be general requirement: C xi xj d generalization of GV bound: d-separated family of size

n d

slide-56
SLIDE 56

. . . . . .

Gilbert-Varshamov complexity bound

◮ coding theory: how many n-bit strings x1, . . . , xk one can find

if Hamming distance between every two is at least d lower bound (Gilbert–Varshamov) then d changed bits are harmless but bit insertion or deletions could be general requirement: C xi xj d generalization of GV bound: d-separated family of size

n d

slide-57
SLIDE 57

. . . . . .

Gilbert-Varshamov complexity bound

◮ coding theory: how many n-bit strings x1, . . . , xk one can find

if Hamming distance between every two is at least d

◮ lower bound (Gilbert–Varshamov)

then d changed bits are harmless but bit insertion or deletions could be general requirement: C xi xj d generalization of GV bound: d-separated family of size

n d

slide-58
SLIDE 58

. . . . . .

Gilbert-Varshamov complexity bound

◮ coding theory: how many n-bit strings x1, . . . , xk one can find

if Hamming distance between every two is at least d

◮ lower bound (Gilbert–Varshamov) ◮ then < d/2 changed bits are harmless

but bit insertion or deletions could be general requirement: C xi xj d generalization of GV bound: d-separated family of size

n d

slide-59
SLIDE 59

. . . . . .

Gilbert-Varshamov complexity bound

◮ coding theory: how many n-bit strings x1, . . . , xk one can find

if Hamming distance between every two is at least d

◮ lower bound (Gilbert–Varshamov) ◮ then < d/2 changed bits are harmless ◮ but bit insertion or deletions could be

general requirement: C xi xj d generalization of GV bound: d-separated family of size

n d

slide-60
SLIDE 60

. . . . . .

Gilbert-Varshamov complexity bound

◮ coding theory: how many n-bit strings x1, . . . , xk one can find

if Hamming distance between every two is at least d

◮ lower bound (Gilbert–Varshamov) ◮ then < d/2 changed bits are harmless ◮ but bit insertion or deletions could be ◮ general requirement: C(xi|xj) ≥ d

generalization of GV bound: d-separated family of size

n d

slide-61
SLIDE 61

. . . . . .

Gilbert-Varshamov complexity bound

◮ coding theory: how many n-bit strings x1, . . . , xk one can find

if Hamming distance between every two is at least d

◮ lower bound (Gilbert–Varshamov) ◮ then < d/2 changed bits are harmless ◮ but bit insertion or deletions could be ◮ general requirement: C(xi|xj) ≥ d ◮ generalization of GV bound: d-separated family of size

Ω(2n−d)

slide-62
SLIDE 62

. . . . . .

Inequalities for complexities and combinatorial interpretation

C x y C x C y x O log C x y C x C y x O log C x y k l C x k O log or C y x l O log every set A of size

k l can be split into two parts

A A A such that w A

k and h A l

slide-63
SLIDE 63

. . . . . .

Inequalities for complexities and combinatorial interpretation

◮ C(x, y) ≤ C(x) + C(y|x) + O(log)

C x y C x C y x O log C x y k l C x k O log or C y x l O log every set A of size

k l can be split into two parts

A A A such that w A

k and h A l

slide-64
SLIDE 64

. . . . . .

Inequalities for complexities and combinatorial interpretation

◮ C(x, y) ≤ C(x) + C(y|x) + O(log)

C x y C x C y x O log C x y k l C x k O log or C y x l O log every set A of size

k l can be split into two parts

A A A such that w A

k and h A l

slide-65
SLIDE 65

. . . . . .

Inequalities for complexities and combinatorial interpretation

◮ C(x, y) ≤ C(x) + C(y|x) + O(log) ◮ C(x, y) ≥ C(x) + C(y|x) + O(log)

C x y k l C x k O log or C y x l O log every set A of size

k l can be split into two parts

A A A such that w A

k and h A l

slide-66
SLIDE 66

. . . . . .

Inequalities for complexities and combinatorial interpretation

◮ C(x, y) ≤ C(x) + C(y|x) + O(log) ◮ C(x, y) ≥ C(x) + C(y|x) + O(log) ◮ C(x, y) < k + l ⇒ C(x) < k + O(log) or C(y|x) < l + O(log)

every set A of size

k l can be split into two parts

A A A such that w A

k and h A l

slide-67
SLIDE 67

. . . . . .

Inequalities for complexities and combinatorial interpretation

◮ C(x, y) ≤ C(x) + C(y|x) + O(log) ◮ C(x, y) ≥ C(x) + C(y|x) + O(log) ◮ C(x, y) < k + l ⇒ C(x) < k + O(log) or C(y|x) < l + O(log) ◮ every set A of size < 2k+l can be split into two parts

A = A1 ∪ A2 such that w(A1) ≤ 2k and h(A2) ≤ 2l

slide-68
SLIDE 68

. . . . . .

One more inequality

C x y z C x y C y z C x z V S S S Also for Shannon entropies; special case of Shearer lemma

slide-69
SLIDE 69

. . . . . .

One more inequality

◮ 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z)

V S S S Also for Shannon entropies; special case of Shearer lemma

slide-70
SLIDE 70

. . . . . .

One more inequality

◮ 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) ◮ V2 ≤ S1 × S2 × S3

Also for Shannon entropies; special case of Shearer lemma

slide-71
SLIDE 71

. . . . . .

One more inequality

◮ 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) ◮ V2 ≤ S1 × S2 × S3 ◮ Also for Shannon entropies; special case of Shearer lemma

slide-72
SLIDE 72

. . . . . .

Common information and graph minors

mutual information: I a b C a C b C a b common information: combinatorial: graph minors can the graph be covered by minors of size ?

slide-73
SLIDE 73

. . . . . .

Common information and graph minors

◮ mutual information: I(a : b) = C(a) + C(b) − C(a, b)

common information: combinatorial: graph minors can the graph be covered by minors of size ?

slide-74
SLIDE 74

. . . . . .

Common information and graph minors

◮ mutual information: I(a : b) = C(a) + C(b) − C(a, b) ◮ common information:

combinatorial: graph minors can the graph be covered by minors of size ?

slide-75
SLIDE 75

. . . . . .

Common information and graph minors

◮ mutual information: I(a : b) = C(a) + C(b) − C(a, b) ◮ common information: ◮ combinatorial: graph minors

can the graph be covered by 2δ minors of size 2α−δ × 2β−δ?

slide-76
SLIDE 76

. . . . . .

Almost uniform sets

nonuniformity= (maximal section)/(average section) Theorem: every set of N elements can be represented as union of polylog N sets whose nonuniformity is polylog N . multidimensional version how to construct parts using Kolmogorov complexity: take strings with given complexity bounds so simple that it is not clear what is the combinatorial translation but combinatorial argument exists (and gives even a stronger result)

slide-77
SLIDE 77

. . . . . .

Almost uniform sets

◮ nonuniformity= (maximal section)/(average section)

Theorem: every set of N elements can be represented as union of polylog N sets whose nonuniformity is polylog N . multidimensional version how to construct parts using Kolmogorov complexity: take strings with given complexity bounds so simple that it is not clear what is the combinatorial translation but combinatorial argument exists (and gives even a stronger result)

slide-78
SLIDE 78

. . . . . .

Almost uniform sets

◮ nonuniformity= (maximal section)/(average section) ◮ Theorem: every set of N elements can be represented as

union of polylog(N) sets whose nonuniformity is polylog(N). multidimensional version how to construct parts using Kolmogorov complexity: take strings with given complexity bounds so simple that it is not clear what is the combinatorial translation but combinatorial argument exists (and gives even a stronger result)

slide-79
SLIDE 79

. . . . . .

Almost uniform sets

◮ nonuniformity= (maximal section)/(average section) ◮ Theorem: every set of N elements can be represented as

union of polylog(N) sets whose nonuniformity is polylog(N).

◮ multidimensional version

how to construct parts using Kolmogorov complexity: take strings with given complexity bounds so simple that it is not clear what is the combinatorial translation but combinatorial argument exists (and gives even a stronger result)

slide-80
SLIDE 80

. . . . . .

Almost uniform sets

◮ nonuniformity= (maximal section)/(average section) ◮ Theorem: every set of N elements can be represented as

union of polylog(N) sets whose nonuniformity is polylog(N).

◮ multidimensional version ◮ how to construct parts using Kolmogorov complexity: take

strings with given complexity bounds so simple that it is not clear what is the combinatorial translation but combinatorial argument exists (and gives even a stronger result)

slide-81
SLIDE 81

. . . . . .

Almost uniform sets

◮ nonuniformity= (maximal section)/(average section) ◮ Theorem: every set of N elements can be represented as

union of polylog(N) sets whose nonuniformity is polylog(N).

◮ multidimensional version ◮ how to construct parts using Kolmogorov complexity: take

strings with given complexity bounds

◮ so simple that it is not clear what is the combinatorial

translation but combinatorial argument exists (and gives even a stronger result)

slide-82
SLIDE 82

. . . . . .

Almost uniform sets

◮ nonuniformity= (maximal section)/(average section) ◮ Theorem: every set of N elements can be represented as

union of polylog(N) sets whose nonuniformity is polylog(N).

◮ multidimensional version ◮ how to construct parts using Kolmogorov complexity: take

strings with given complexity bounds

◮ so simple that it is not clear what is the combinatorial

translation

◮ but combinatorial argument exists (and gives even a stronger

result)

slide-83
SLIDE 83

. . . . . .

Shannon coding theorem

is a random variable; k values, probabilities p pk

N: N independent trials of

Shannon’s informal question: how many bits are needed to encode a “typical” value of

N?

Shannon’s answer: NH , where H p log p pn log pn formal statement is a bit complicated Complexity version: with high probablity the value of

N has

complexity close to NH .

slide-84
SLIDE 84

. . . . . .

Shannon coding theorem

◮ ξ is a random variable; k values, probabilities p1, . . . , pk N: N independent trials of

Shannon’s informal question: how many bits are needed to encode a “typical” value of

N?

Shannon’s answer: NH , where H p log p pn log pn formal statement is a bit complicated Complexity version: with high probablity the value of

N has

complexity close to NH .

slide-85
SLIDE 85

. . . . . .

Shannon coding theorem

◮ ξ is a random variable; k values, probabilities p1, . . . , pk ◮ ξN: N independent trials of ξ

Shannon’s informal question: how many bits are needed to encode a “typical” value of

N?

Shannon’s answer: NH , where H p log p pn log pn formal statement is a bit complicated Complexity version: with high probablity the value of

N has

complexity close to NH .

slide-86
SLIDE 86

. . . . . .

Shannon coding theorem

◮ ξ is a random variable; k values, probabilities p1, . . . , pk ◮ ξN: N independent trials of ξ ◮ Shannon’s informal question: how many bits are needed to

encode a “typical” value of ξN? Shannon’s answer: NH , where H p log p pn log pn formal statement is a bit complicated Complexity version: with high probablity the value of

N has

complexity close to NH .

slide-87
SLIDE 87

. . . . . .

Shannon coding theorem

◮ ξ is a random variable; k values, probabilities p1, . . . , pk ◮ ξN: N independent trials of ξ ◮ Shannon’s informal question: how many bits are needed to

encode a “typical” value of ξN?

◮ Shannon’s answer: NH(ξ), where

H(ξ) = p1 log(1/p1) + . . . + pn log(1/pn). formal statement is a bit complicated Complexity version: with high probablity the value of

N has

complexity close to NH .

slide-88
SLIDE 88

. . . . . .

Shannon coding theorem

◮ ξ is a random variable; k values, probabilities p1, . . . , pk ◮ ξN: N independent trials of ξ ◮ Shannon’s informal question: how many bits are needed to

encode a “typical” value of ξN?

◮ Shannon’s answer: NH(ξ), where

H(ξ) = p1 log(1/p1) + . . . + pn log(1/pn).

◮ formal statement is a bit complicated

Complexity version: with high probablity the value of

N has

complexity close to NH .

slide-89
SLIDE 89

. . . . . .

Shannon coding theorem

◮ ξ is a random variable; k values, probabilities p1, . . . , pk ◮ ξN: N independent trials of ξ ◮ Shannon’s informal question: how many bits are needed to

encode a “typical” value of ξN?

◮ Shannon’s answer: NH(ξ), where

H(ξ) = p1 log(1/p1) + . . . + pn log(1/pn).

◮ formal statement is a bit complicated ◮ Complexity version: with high probablity the value of ξN has

complexity close to NH(ξ).

slide-90
SLIDE 90

. . . . . .

Complexity, entropy and group size

C x y z C x y C y z C x z O log The same for entropy: H H H H …and even for the sizes of subgroups U V W of some finite group G: log G U V W log G U V log G U W log G V W . in all three cases inequalities are the same (Romashchenko, Chan, Yeung) some of them are quite strange: I a b I a b c I a b d I c d I a b e I a e b I b e a Related to Romashchenko’s theorem: if three last terms are zeros, one can extract common information from a b e.

slide-91
SLIDE 91

. . . . . .

Complexity, entropy and group size

◮ 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log)

The same for entropy: H H H H …and even for the sizes of subgroups U V W of some finite group G: log G U V W log G U V log G U W log G V W . in all three cases inequalities are the same (Romashchenko, Chan, Yeung) some of them are quite strange: I a b I a b c I a b d I c d I a b e I a e b I b e a Related to Romashchenko’s theorem: if three last terms are zeros, one can extract common information from a b e.

slide-92
SLIDE 92

. . . . . .

Complexity, entropy and group size

◮ 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log) ◮ The same for entropy:

2H(ξ, η, τ) ≤ H(ξ, η) + H(ξ, τ) + H(η, τ) …and even for the sizes of subgroups U V W of some finite group G: log G U V W log G U V log G U W log G V W . in all three cases inequalities are the same (Romashchenko, Chan, Yeung) some of them are quite strange: I a b I a b c I a b d I c d I a b e I a e b I b e a Related to Romashchenko’s theorem: if three last terms are zeros, one can extract common information from a b e.

slide-93
SLIDE 93

. . . . . .

Complexity, entropy and group size

◮ 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log) ◮ The same for entropy:

2H(ξ, η, τ) ≤ H(ξ, η) + H(ξ, τ) + H(η, τ)

◮ …and even for the sizes of subgroups U, V, W of some finite

group G: 2 log(|G|/|U ∩ V ∩ W|) ≤ log(|G|/|U ∩ V|) + log(|G|/|U ∩ W|) + log(|G|/|V ∩ W|). in all three cases inequalities are the same (Romashchenko, Chan, Yeung) some of them are quite strange: I a b I a b c I a b d I c d I a b e I a e b I b e a Related to Romashchenko’s theorem: if three last terms are zeros, one can extract common information from a b e.

slide-94
SLIDE 94

. . . . . .

Complexity, entropy and group size

◮ 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log) ◮ The same for entropy:

2H(ξ, η, τ) ≤ H(ξ, η) + H(ξ, τ) + H(η, τ)

◮ …and even for the sizes of subgroups U, V, W of some finite

group G: 2 log(|G|/|U ∩ V ∩ W|) ≤ log(|G|/|U ∩ V|) + log(|G|/|U ∩ W|) + log(|G|/|V ∩ W|).

◮ in all three cases inequalities are the same (Romashchenko,

Chan, Yeung) some of them are quite strange: I a b I a b c I a b d I c d I a b e I a e b I b e a Related to Romashchenko’s theorem: if three last terms are zeros, one can extract common information from a b e.

slide-95
SLIDE 95

. . . . . .

Complexity, entropy and group size

◮ 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log) ◮ The same for entropy:

2H(ξ, η, τ) ≤ H(ξ, η) + H(ξ, τ) + H(η, τ)

◮ …and even for the sizes of subgroups U, V, W of some finite

group G: 2 log(|G|/|U ∩ V ∩ W|) ≤ log(|G|/|U ∩ V|) + log(|G|/|U ∩ W|) + log(|G|/|V ∩ W|).

◮ in all three cases inequalities are the same (Romashchenko,

Chan, Yeung)

◮ some of them are quite strange:

I(a : b) ≤ ≤ I(a : b|c)+I(a : b|d)+I(c : d)+I(a : b|e)+I(a : e|b)+I(b : e|a) Related to Romashchenko’s theorem: if three last terms are zeros, one can extract common information from a b e.

slide-96
SLIDE 96

. . . . . .

Complexity, entropy and group size

◮ 2C(x, y, z) ≤ C(x, y) + C(y, z) + C(x, z) + O(log) ◮ The same for entropy:

2H(ξ, η, τ) ≤ H(ξ, η) + H(ξ, τ) + H(η, τ)

◮ …and even for the sizes of subgroups U, V, W of some finite

group G: 2 log(|G|/|U ∩ V ∩ W|) ≤ log(|G|/|U ∩ V|) + log(|G|/|U ∩ W|) + log(|G|/|V ∩ W|).

◮ in all three cases inequalities are the same (Romashchenko,

Chan, Yeung)

◮ some of them are quite strange:

I(a : b) ≤ ≤ I(a : b|c)+I(a : b|d)+I(c : d)+I(a : b|e)+I(a : e|b)+I(b : e|a)

◮ Related to Romashchenko’s theorem: if three last terms are

zeros, one can extract common information from a, b, e.

slide-97
SLIDE 97

. . . . . .

Muchnik and Slepian–Wolf

a b: two strings we look for a program p that maps a to b by definition C p is at least C b a but could be higher there exist p a b that is simple relative to b, e.g., “map everything to b” Muchnik theorem: it is possible to combine these two conditions: there exists p a b such that C p C b a and C p b information theory analog: Wolf–Slepian similar technique was developed by Fortnow and Laplante (randomness extractors) (Romashchenko, Musatov): how to use explicit extractors and derandomization to get space-bounded versions

slide-98
SLIDE 98

. . . . . .

Muchnik and Slepian–Wolf

◮ a, b: two strings

we look for a program p that maps a to b by definition C p is at least C b a but could be higher there exist p a b that is simple relative to b, e.g., “map everything to b” Muchnik theorem: it is possible to combine these two conditions: there exists p a b such that C p C b a and C p b information theory analog: Wolf–Slepian similar technique was developed by Fortnow and Laplante (randomness extractors) (Romashchenko, Musatov): how to use explicit extractors and derandomization to get space-bounded versions

slide-99
SLIDE 99

. . . . . .

Muchnik and Slepian–Wolf

◮ a, b: two strings ◮ we look for a program p that maps a to b

by definition C p is at least C b a but could be higher there exist p a b that is simple relative to b, e.g., “map everything to b” Muchnik theorem: it is possible to combine these two conditions: there exists p a b such that C p C b a and C p b information theory analog: Wolf–Slepian similar technique was developed by Fortnow and Laplante (randomness extractors) (Romashchenko, Musatov): how to use explicit extractors and derandomization to get space-bounded versions

slide-100
SLIDE 100

. . . . . .

Muchnik and Slepian–Wolf

◮ a, b: two strings ◮ we look for a program p that maps a to b ◮ by definition C(p) is at least C(b|a) but could be higher

there exist p a b that is simple relative to b, e.g., “map everything to b” Muchnik theorem: it is possible to combine these two conditions: there exists p a b such that C p C b a and C p b information theory analog: Wolf–Slepian similar technique was developed by Fortnow and Laplante (randomness extractors) (Romashchenko, Musatov): how to use explicit extractors and derandomization to get space-bounded versions

slide-101
SLIDE 101

. . . . . .

Muchnik and Slepian–Wolf

◮ a, b: two strings ◮ we look for a program p that maps a to b ◮ by definition C(p) is at least C(b|a) but could be higher ◮ there exist p: a → b that is simple relative to b, e.g., “map

everything to b” Muchnik theorem: it is possible to combine these two conditions: there exists p a b such that C p C b a and C p b information theory analog: Wolf–Slepian similar technique was developed by Fortnow and Laplante (randomness extractors) (Romashchenko, Musatov): how to use explicit extractors and derandomization to get space-bounded versions

slide-102
SLIDE 102

. . . . . .

Muchnik and Slepian–Wolf

◮ a, b: two strings ◮ we look for a program p that maps a to b ◮ by definition C(p) is at least C(b|a) but could be higher ◮ there exist p: a → b that is simple relative to b, e.g., “map

everything to b”

◮ Muchnik theorem: it is possible to combine these two

conditions: there exists p: a → b such that C(p) ≈ C(b|a) and C(p|b) ≈ 0 information theory analog: Wolf–Slepian similar technique was developed by Fortnow and Laplante (randomness extractors) (Romashchenko, Musatov): how to use explicit extractors and derandomization to get space-bounded versions

slide-103
SLIDE 103

. . . . . .

Muchnik and Slepian–Wolf

◮ a, b: two strings ◮ we look for a program p that maps a to b ◮ by definition C(p) is at least C(b|a) but could be higher ◮ there exist p: a → b that is simple relative to b, e.g., “map

everything to b”

◮ Muchnik theorem: it is possible to combine these two

conditions: there exists p: a → b such that C(p) ≈ C(b|a) and C(p|b) ≈ 0

◮ information theory analog: Wolf–Slepian

similar technique was developed by Fortnow and Laplante (randomness extractors) (Romashchenko, Musatov): how to use explicit extractors and derandomization to get space-bounded versions

slide-104
SLIDE 104

. . . . . .

Muchnik and Slepian–Wolf

◮ a, b: two strings ◮ we look for a program p that maps a to b ◮ by definition C(p) is at least C(b|a) but could be higher ◮ there exist p: a → b that is simple relative to b, e.g., “map

everything to b”

◮ Muchnik theorem: it is possible to combine these two

conditions: there exists p: a → b such that C(p) ≈ C(b|a) and C(p|b) ≈ 0

◮ information theory analog: Wolf–Slepian ◮ similar technique was developed by Fortnow and Laplante

(randomness extractors) (Romashchenko, Musatov): how to use explicit extractors and derandomization to get space-bounded versions

slide-105
SLIDE 105

. . . . . .

Muchnik and Slepian–Wolf

◮ a, b: two strings ◮ we look for a program p that maps a to b ◮ by definition C(p) is at least C(b|a) but could be higher ◮ there exist p: a → b that is simple relative to b, e.g., “map

everything to b”

◮ Muchnik theorem: it is possible to combine these two

conditions: there exists p: a → b such that C(p) ≈ C(b|a) and C(p|b) ≈ 0

◮ information theory analog: Wolf–Slepian ◮ similar technique was developed by Fortnow and Laplante

(randomness extractors)

◮ (Romashchenko, Musatov): how to use explicit extractors and

derandomization to get space-bounded versions

slide-106
SLIDE 106

. . . . . .

Computability theory: simple sets

Simple set: enumerable set with infinite complement, but no algorithm can generate infinitely many elements from the complement Construction using Kolmogorov complexity: a simple string x has C x x . Most strings are not simple infinite complement Let x x be a computable sequence of different non-simple strings May assume wlog that xi i and therefore C xi i but to specify xi we need O log i bits only “Minimal integer that cannot be described in ten English words” (Berry)

slide-107
SLIDE 107

. . . . . .

Computability theory: simple sets

◮ Simple set: enumerable set with infinite complement, but no

algorithm can generate infinitely many elements from the complement Construction using Kolmogorov complexity: a simple string x has C x x . Most strings are not simple infinite complement Let x x be a computable sequence of different non-simple strings May assume wlog that xi i and therefore C xi i but to specify xi we need O log i bits only “Minimal integer that cannot be described in ten English words” (Berry)

slide-108
SLIDE 108

. . . . . .

Computability theory: simple sets

◮ Simple set: enumerable set with infinite complement, but no

algorithm can generate infinitely many elements from the complement

◮ Construction using Kolmogorov complexity: a simple string x

has C(x) ≤ |x|/2. Most strings are not simple infinite complement Let x x be a computable sequence of different non-simple strings May assume wlog that xi i and therefore C xi i but to specify xi we need O log i bits only “Minimal integer that cannot be described in ten English words” (Berry)

slide-109
SLIDE 109

. . . . . .

Computability theory: simple sets

◮ Simple set: enumerable set with infinite complement, but no

algorithm can generate infinitely many elements from the complement

◮ Construction using Kolmogorov complexity: a simple string x

has C(x) ≤ |x|/2.

◮ Most strings are not simple ⇒ infinite complement

Let x x be a computable sequence of different non-simple strings May assume wlog that xi i and therefore C xi i but to specify xi we need O log i bits only “Minimal integer that cannot be described in ten English words” (Berry)

slide-110
SLIDE 110

. . . . . .

Computability theory: simple sets

◮ Simple set: enumerable set with infinite complement, but no

algorithm can generate infinitely many elements from the complement

◮ Construction using Kolmogorov complexity: a simple string x

has C(x) ≤ |x|/2.

◮ Most strings are not simple ⇒ infinite complement ◮ Let x1, x2, . . . be a computable sequence of different

non-simple strings May assume wlog that xi i and therefore C xi i but to specify xi we need O log i bits only “Minimal integer that cannot be described in ten English words” (Berry)

slide-111
SLIDE 111

. . . . . .

Computability theory: simple sets

◮ Simple set: enumerable set with infinite complement, but no

algorithm can generate infinitely many elements from the complement

◮ Construction using Kolmogorov complexity: a simple string x

has C(x) ≤ |x|/2.

◮ Most strings are not simple ⇒ infinite complement ◮ Let x1, x2, . . . be a computable sequence of different

non-simple strings

◮ May assume wlog that |xi| > i and therefore C(xi) > i/2

but to specify xi we need O log i bits only “Minimal integer that cannot be described in ten English words” (Berry)

slide-112
SLIDE 112

. . . . . .

Computability theory: simple sets

◮ Simple set: enumerable set with infinite complement, but no

algorithm can generate infinitely many elements from the complement

◮ Construction using Kolmogorov complexity: a simple string x

has C(x) ≤ |x|/2.

◮ Most strings are not simple ⇒ infinite complement ◮ Let x1, x2, . . . be a computable sequence of different

non-simple strings

◮ May assume wlog that |xi| > i and therefore C(xi) > i/2 ◮ but to specify xi we need O(log i) bits only

“Minimal integer that cannot be described in ten English words” (Berry)

slide-113
SLIDE 113

. . . . . .

Computability theory: simple sets

◮ Simple set: enumerable set with infinite complement, but no

algorithm can generate infinitely many elements from the complement

◮ Construction using Kolmogorov complexity: a simple string x

has C(x) ≤ |x|/2.

◮ Most strings are not simple ⇒ infinite complement ◮ Let x1, x2, . . . be a computable sequence of different

non-simple strings

◮ May assume wlog that |xi| > i and therefore C(xi) > i/2 ◮ but to specify xi we need O(log i) bits only ◮ “Minimal integer that cannot be described in ten English

words” (Berry)

slide-114
SLIDE 114

. . . . . .

Lower semicomputable random reals

ai computable converging series with rational terms is ai computable (

  • approximation)?

not necessarily (Specker example) lower semicomputable reals Solovay classification: if -approximation to can be effectively converted to O

  • approximation to

There are maximal elements = random lower semicomputable reals = slowly converging series modulus of convergence: N = how many terms are needed for -precision maximal elements: N

n

BP n O , where BP k is the maximal integer whose prefix complexity is k or less.

slide-115
SLIDE 115

. . . . . .

Lower semicomputable random reals

◮ ∑ ai computable converging series with rational terms

is ai computable (

  • approximation)?

not necessarily (Specker example) lower semicomputable reals Solovay classification: if -approximation to can be effectively converted to O

  • approximation to

There are maximal elements = random lower semicomputable reals = slowly converging series modulus of convergence: N = how many terms are needed for -precision maximal elements: N

n

BP n O , where BP k is the maximal integer whose prefix complexity is k or less.

slide-116
SLIDE 116

. . . . . .

Lower semicomputable random reals

◮ ∑ ai computable converging series with rational terms ◮ is α = ∑ ai computable (ε → ε-approximation)?

not necessarily (Specker example) lower semicomputable reals Solovay classification: if -approximation to can be effectively converted to O

  • approximation to

There are maximal elements = random lower semicomputable reals = slowly converging series modulus of convergence: N = how many terms are needed for -precision maximal elements: N

n

BP n O , where BP k is the maximal integer whose prefix complexity is k or less.

slide-117
SLIDE 117

. . . . . .

Lower semicomputable random reals

◮ ∑ ai computable converging series with rational terms ◮ is α = ∑ ai computable (ε → ε-approximation)? ◮ not necessarily (Specker example)

lower semicomputable reals Solovay classification: if -approximation to can be effectively converted to O

  • approximation to

There are maximal elements = random lower semicomputable reals = slowly converging series modulus of convergence: N = how many terms are needed for -precision maximal elements: N

n

BP n O , where BP k is the maximal integer whose prefix complexity is k or less.

slide-118
SLIDE 118

. . . . . .

Lower semicomputable random reals

◮ ∑ ai computable converging series with rational terms ◮ is α = ∑ ai computable (ε → ε-approximation)? ◮ not necessarily (Specker example) ◮ lower semicomputable reals

Solovay classification: if -approximation to can be effectively converted to O

  • approximation to

There are maximal elements = random lower semicomputable reals = slowly converging series modulus of convergence: N = how many terms are needed for -precision maximal elements: N

n

BP n O , where BP k is the maximal integer whose prefix complexity is k or less.

slide-119
SLIDE 119

. . . . . .

Lower semicomputable random reals

◮ ∑ ai computable converging series with rational terms ◮ is α = ∑ ai computable (ε → ε-approximation)? ◮ not necessarily (Specker example) ◮ lower semicomputable reals ◮ Solovay classification: α β if ε-approximation to β can be

effectively converted to O(ε)-approximation to α There are maximal elements = random lower semicomputable reals = slowly converging series modulus of convergence: N = how many terms are needed for -precision maximal elements: N

n

BP n O , where BP k is the maximal integer whose prefix complexity is k or less.

slide-120
SLIDE 120

. . . . . .

Lower semicomputable random reals

◮ ∑ ai computable converging series with rational terms ◮ is α = ∑ ai computable (ε → ε-approximation)? ◮ not necessarily (Specker example) ◮ lower semicomputable reals ◮ Solovay classification: α β if ε-approximation to β can be

effectively converted to O(ε)-approximation to α

◮ There are maximal elements

= random lower semicomputable reals = slowly converging series modulus of convergence: N = how many terms are needed for -precision maximal elements: N

n

BP n O , where BP k is the maximal integer whose prefix complexity is k or less.

slide-121
SLIDE 121

. . . . . .

Lower semicomputable random reals

◮ ∑ ai computable converging series with rational terms ◮ is α = ∑ ai computable (ε → ε-approximation)? ◮ not necessarily (Specker example) ◮ lower semicomputable reals ◮ Solovay classification: α β if ε-approximation to β can be

effectively converted to O(ε)-approximation to α

◮ There are maximal elements = random lower

semicomputable reals = slowly converging series modulus of convergence: N = how many terms are needed for -precision maximal elements: N

n

BP n O , where BP k is the maximal integer whose prefix complexity is k or less.

slide-122
SLIDE 122

. . . . . .

Lower semicomputable random reals

◮ ∑ ai computable converging series with rational terms ◮ is α = ∑ ai computable (ε → ε-approximation)? ◮ not necessarily (Specker example) ◮ lower semicomputable reals ◮ Solovay classification: α β if ε-approximation to β can be

effectively converted to O(ε)-approximation to α

◮ There are maximal elements = random lower

semicomputable reals

◮ = slowly converging series

modulus of convergence: N = how many terms are needed for -precision maximal elements: N

n

BP n O , where BP k is the maximal integer whose prefix complexity is k or less.

slide-123
SLIDE 123

. . . . . .

Lower semicomputable random reals

◮ ∑ ai computable converging series with rational terms ◮ is α = ∑ ai computable (ε → ε-approximation)? ◮ not necessarily (Specker example) ◮ lower semicomputable reals ◮ Solovay classification: α β if ε-approximation to β can be

effectively converted to O(ε)-approximation to α

◮ There are maximal elements = random lower

semicomputable reals

◮ = slowly converging series ◮ modulus of convergence: ε → N(ε) = how many terms are

needed for ε-precision maximal elements: N

n

BP n O , where BP k is the maximal integer whose prefix complexity is k or less.

slide-124
SLIDE 124

. . . . . .

Lower semicomputable random reals

◮ ∑ ai computable converging series with rational terms ◮ is α = ∑ ai computable (ε → ε-approximation)? ◮ not necessarily (Specker example) ◮ lower semicomputable reals ◮ Solovay classification: α β if ε-approximation to β can be

effectively converted to O(ε)-approximation to α

◮ There are maximal elements = random lower

semicomputable reals

◮ = slowly converging series ◮ modulus of convergence: ε → N(ε) = how many terms are

needed for ε-precision

◮ maximal elements: N(2−n) > BP(n − O(1)), where BP(k) is

the maximal integer whose prefix complexity is k or less.

slide-125
SLIDE 125

. . . . . .

Lovasz local lemma: constructive proof

CNF: a b c e neighbors: clauses having common variables several clauses with k literals in each each clause has o

k neighbors

CNF is satisfiable Non-constructive proof: lower bound for probability, Lovasz local lemma Naïve algorithm: just resample false clause (while they exist) Recent breakthrough (Moser): this algorithm with high probability terminates quickly Explanation: if not, the sequence of resampled clauses would encode the random bits used in resampling making them compressible

slide-126
SLIDE 126

. . . . . .

Lovasz local lemma: constructive proof

◮ CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . .

neighbors: clauses having common variables several clauses with k literals in each each clause has o

k neighbors

CNF is satisfiable Non-constructive proof: lower bound for probability, Lovasz local lemma Naïve algorithm: just resample false clause (while they exist) Recent breakthrough (Moser): this algorithm with high probability terminates quickly Explanation: if not, the sequence of resampled clauses would encode the random bits used in resampling making them compressible

slide-127
SLIDE 127

. . . . . .

Lovasz local lemma: constructive proof

◮ CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . . ◮ neighbors: clauses having common variables

several clauses with k literals in each each clause has o

k neighbors

CNF is satisfiable Non-constructive proof: lower bound for probability, Lovasz local lemma Naïve algorithm: just resample false clause (while they exist) Recent breakthrough (Moser): this algorithm with high probability terminates quickly Explanation: if not, the sequence of resampled clauses would encode the random bits used in resampling making them compressible

slide-128
SLIDE 128

. . . . . .

Lovasz local lemma: constructive proof

◮ CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . . ◮ neighbors: clauses having common variables ◮ several clauses with k literals in each

each clause has o

k neighbors

CNF is satisfiable Non-constructive proof: lower bound for probability, Lovasz local lemma Naïve algorithm: just resample false clause (while they exist) Recent breakthrough (Moser): this algorithm with high probability terminates quickly Explanation: if not, the sequence of resampled clauses would encode the random bits used in resampling making them compressible

slide-129
SLIDE 129

. . . . . .

Lovasz local lemma: constructive proof

◮ CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . . ◮ neighbors: clauses having common variables ◮ several clauses with k literals in each ◮ each clause has o(2k) neighbors

CNF is satisfiable Non-constructive proof: lower bound for probability, Lovasz local lemma Naïve algorithm: just resample false clause (while they exist) Recent breakthrough (Moser): this algorithm with high probability terminates quickly Explanation: if not, the sequence of resampled clauses would encode the random bits used in resampling making them compressible

slide-130
SLIDE 130

. . . . . .

Lovasz local lemma: constructive proof

◮ CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . . ◮ neighbors: clauses having common variables ◮ several clauses with k literals in each ◮ each clause has o(2k) neighbors ◮ ⇒ CNF is satisfiable

Non-constructive proof: lower bound for probability, Lovasz local lemma Naïve algorithm: just resample false clause (while they exist) Recent breakthrough (Moser): this algorithm with high probability terminates quickly Explanation: if not, the sequence of resampled clauses would encode the random bits used in resampling making them compressible

slide-131
SLIDE 131

. . . . . .

Lovasz local lemma: constructive proof

◮ CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . . ◮ neighbors: clauses having common variables ◮ several clauses with k literals in each ◮ each clause has o(2k) neighbors ◮ ⇒ CNF is satisfiable ◮ Non-constructive proof: lower bound for probability, Lovasz

local lemma Naïve algorithm: just resample false clause (while they exist) Recent breakthrough (Moser): this algorithm with high probability terminates quickly Explanation: if not, the sequence of resampled clauses would encode the random bits used in resampling making them compressible

slide-132
SLIDE 132

. . . . . .

Lovasz local lemma: constructive proof

◮ CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . . ◮ neighbors: clauses having common variables ◮ several clauses with k literals in each ◮ each clause has o(2k) neighbors ◮ ⇒ CNF is satisfiable ◮ Non-constructive proof: lower bound for probability, Lovasz

local lemma

◮ Naïve algorithm: just resample false clause (while they exist)

Recent breakthrough (Moser): this algorithm with high probability terminates quickly Explanation: if not, the sequence of resampled clauses would encode the random bits used in resampling making them compressible

slide-133
SLIDE 133

. . . . . .

Lovasz local lemma: constructive proof

◮ CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . . ◮ neighbors: clauses having common variables ◮ several clauses with k literals in each ◮ each clause has o(2k) neighbors ◮ ⇒ CNF is satisfiable ◮ Non-constructive proof: lower bound for probability, Lovasz

local lemma

◮ Naïve algorithm: just resample false clause (while they exist) ◮ Recent breakthrough (Moser): this algorithm with high

probability terminates quickly Explanation: if not, the sequence of resampled clauses would encode the random bits used in resampling making them compressible

slide-134
SLIDE 134

. . . . . .

Lovasz local lemma: constructive proof

◮ CNF: (a ∧ ¬b ∧ . . .) ∨ (¬c ∧ e ∧ . . .) ∨ . . . ◮ neighbors: clauses having common variables ◮ several clauses with k literals in each ◮ each clause has o(2k) neighbors ◮ ⇒ CNF is satisfiable ◮ Non-constructive proof: lower bound for probability, Lovasz

local lemma

◮ Naïve algorithm: just resample false clause (while they exist) ◮ Recent breakthrough (Moser): this algorithm with high

probability terminates quickly

◮ Explanation: if not, the sequence of resampled clauses

would encode the random bits used in resampling making them compressible

slide-135
SLIDE 135

. . . . . .

Berry, Gödel, Chaitin, Raz

There are only finitely many strings of complexity n for all strings (except finitely many ones) x the statement K x n is true Can all true statements of this form be provable? No, otherwise we could effectively generate string of complexity n by enumerating all proofs and get Berry’s paradox: the first provable statement C x n for given n gives some x of complexity n that can be described by O log n bits (Gödel theorem in Chaitin form): There are only finitely many n such that C x n is provable for some x. (Note that x C x n is always provable!) (Gödel second theorem, Kritchman–Raz proof): the “unexpected test paradox” (a test will be given next week but it won’t be known before the day of the test)

slide-136
SLIDE 136

. . . . . .

Berry, Gödel, Chaitin, Raz

◮ There are only finitely many strings of complexity < n

for all strings (except finitely many ones) x the statement K x n is true Can all true statements of this form be provable? No, otherwise we could effectively generate string of complexity n by enumerating all proofs and get Berry’s paradox: the first provable statement C x n for given n gives some x of complexity n that can be described by O log n bits (Gödel theorem in Chaitin form): There are only finitely many n such that C x n is provable for some x. (Note that x C x n is always provable!) (Gödel second theorem, Kritchman–Raz proof): the “unexpected test paradox” (a test will be given next week but it won’t be known before the day of the test)

slide-137
SLIDE 137

. . . . . .

Berry, Gödel, Chaitin, Raz

◮ There are only finitely many strings of complexity < n ◮ for all strings (except finitely many ones) x the statement

K(x) > n is true Can all true statements of this form be provable? No, otherwise we could effectively generate string of complexity n by enumerating all proofs and get Berry’s paradox: the first provable statement C x n for given n gives some x of complexity n that can be described by O log n bits (Gödel theorem in Chaitin form): There are only finitely many n such that C x n is provable for some x. (Note that x C x n is always provable!) (Gödel second theorem, Kritchman–Raz proof): the “unexpected test paradox” (a test will be given next week but it won’t be known before the day of the test)

slide-138
SLIDE 138

. . . . . .

Berry, Gödel, Chaitin, Raz

◮ There are only finitely many strings of complexity < n ◮ for all strings (except finitely many ones) x the statement

K(x) > n is true

◮ Can all true statements of this form be provable?

No, otherwise we could effectively generate string of complexity n by enumerating all proofs and get Berry’s paradox: the first provable statement C x n for given n gives some x of complexity n that can be described by O log n bits (Gödel theorem in Chaitin form): There are only finitely many n such that C x n is provable for some x. (Note that x C x n is always provable!) (Gödel second theorem, Kritchman–Raz proof): the “unexpected test paradox” (a test will be given next week but it won’t be known before the day of the test)

slide-139
SLIDE 139

. . . . . .

Berry, Gödel, Chaitin, Raz

◮ There are only finitely many strings of complexity < n ◮ for all strings (except finitely many ones) x the statement

K(x) > n is true

◮ Can all true statements of this form be provable? ◮ No, otherwise we could effectively generate string of

complexity > n by enumerating all proofs and get Berry’s paradox: the first provable statement C x n for given n gives some x of complexity n that can be described by O log n bits (Gödel theorem in Chaitin form): There are only finitely many n such that C x n is provable for some x. (Note that x C x n is always provable!) (Gödel second theorem, Kritchman–Raz proof): the “unexpected test paradox” (a test will be given next week but it won’t be known before the day of the test)

slide-140
SLIDE 140

. . . . . .

Berry, Gödel, Chaitin, Raz

◮ There are only finitely many strings of complexity < n ◮ for all strings (except finitely many ones) x the statement

K(x) > n is true

◮ Can all true statements of this form be provable? ◮ No, otherwise we could effectively generate string of

complexity > n by enumerating all proofs

◮ and get Berry’s paradox: the first provable statement

C(x) > n for given n gives some x of complexity > n that can be described by O(log n) bits (Gödel theorem in Chaitin form): There are only finitely many n such that C x n is provable for some x. (Note that x C x n is always provable!) (Gödel second theorem, Kritchman–Raz proof): the “unexpected test paradox” (a test will be given next week but it won’t be known before the day of the test)

slide-141
SLIDE 141

. . . . . .

Berry, Gödel, Chaitin, Raz

◮ There are only finitely many strings of complexity < n ◮ for all strings (except finitely many ones) x the statement

K(x) > n is true

◮ Can all true statements of this form be provable? ◮ No, otherwise we could effectively generate string of

complexity > n by enumerating all proofs

◮ and get Berry’s paradox: the first provable statement

C(x) > n for given n gives some x of complexity > n that can be described by O(log n) bits

◮ (Gödel theorem in Chaitin form): There are only finitely many

n such that C(x) > n is provable for some x. (Note that x C x n is always provable!) (Gödel second theorem, Kritchman–Raz proof): the “unexpected test paradox” (a test will be given next week but it won’t be known before the day of the test)

slide-142
SLIDE 142

. . . . . .

Berry, Gödel, Chaitin, Raz

◮ There are only finitely many strings of complexity < n ◮ for all strings (except finitely many ones) x the statement

K(x) > n is true

◮ Can all true statements of this form be provable? ◮ No, otherwise we could effectively generate string of

complexity > n by enumerating all proofs

◮ and get Berry’s paradox: the first provable statement

C(x) > n for given n gives some x of complexity > n that can be described by O(log n) bits

◮ (Gödel theorem in Chaitin form): There are only finitely many

n such that C(x) > n is provable for some x. (Note that ∃x C(x) > n is always provable!) (Gödel second theorem, Kritchman–Raz proof): the “unexpected test paradox” (a test will be given next week but it won’t be known before the day of the test)

slide-143
SLIDE 143

. . . . . .

Berry, Gödel, Chaitin, Raz

◮ There are only finitely many strings of complexity < n ◮ for all strings (except finitely many ones) x the statement

K(x) > n is true

◮ Can all true statements of this form be provable? ◮ No, otherwise we could effectively generate string of

complexity > n by enumerating all proofs

◮ and get Berry’s paradox: the first provable statement

C(x) > n for given n gives some x of complexity > n that can be described by O(log n) bits

◮ (Gödel theorem in Chaitin form): There are only finitely many

n such that C(x) > n is provable for some x. (Note that ∃x C(x) > n is always provable!)

◮ (Gödel second theorem, Kritchman–Raz proof): the

“unexpected test paradox” (a test will be given next week but it won’t be known before the day of the test)

slide-144
SLIDE 144

. . . . . .

13th Hilbert’s problem

Function of variables: e.g., solution of a polynomial of degree as function of its coefficients Is it possible to represent this function as a composition of functions of at most variables? Yes, with weird functions (Cantor) Yes, even with continuous functions (Kolmogorov, Arnold) Circuit version: explicit function

n n n n

(polynomial circuit size?) that cannot be represented as composition of O functions

n n

  • n. Not known.

Kolmogorov complexity version: we have three strings a b c

  • n a blackboard. It is allowed to write (add) a new string if it

simple relative to two of the strings on the board. Which strings we can obtain in O steps? Only strings of small complexity relative to a b c, but not all of them (for random a b c)

slide-145
SLIDE 145

. . . . . .

13th Hilbert’s problem

◮ Function of ≥ 3 variables: e.g., solution of a polynomial of

degree 7 as function of its coefficients Is it possible to represent this function as a composition of functions of at most variables? Yes, with weird functions (Cantor) Yes, even with continuous functions (Kolmogorov, Arnold) Circuit version: explicit function

n n n n

(polynomial circuit size?) that cannot be represented as composition of O functions

n n

  • n. Not known.

Kolmogorov complexity version: we have three strings a b c

  • n a blackboard. It is allowed to write (add) a new string if it

simple relative to two of the strings on the board. Which strings we can obtain in O steps? Only strings of small complexity relative to a b c, but not all of them (for random a b c)

slide-146
SLIDE 146

. . . . . .

13th Hilbert’s problem

◮ Function of ≥ 3 variables: e.g., solution of a polynomial of

degree 7 as function of its coefficients

◮ Is it possible to represent this function as a composition of

functions of at most 2 variables? Yes, with weird functions (Cantor) Yes, even with continuous functions (Kolmogorov, Arnold) Circuit version: explicit function

n n n n

(polynomial circuit size?) that cannot be represented as composition of O functions

n n

  • n. Not known.

Kolmogorov complexity version: we have three strings a b c

  • n a blackboard. It is allowed to write (add) a new string if it

simple relative to two of the strings on the board. Which strings we can obtain in O steps? Only strings of small complexity relative to a b c, but not all of them (for random a b c)

slide-147
SLIDE 147

. . . . . .

13th Hilbert’s problem

◮ Function of ≥ 3 variables: e.g., solution of a polynomial of

degree 7 as function of its coefficients

◮ Is it possible to represent this function as a composition of

functions of at most 2 variables?

◮ Yes, with weird functions (Cantor)

Yes, even with continuous functions (Kolmogorov, Arnold) Circuit version: explicit function

n n n n

(polynomial circuit size?) that cannot be represented as composition of O functions

n n

  • n. Not known.

Kolmogorov complexity version: we have three strings a b c

  • n a blackboard. It is allowed to write (add) a new string if it

simple relative to two of the strings on the board. Which strings we can obtain in O steps? Only strings of small complexity relative to a b c, but not all of them (for random a b c)

slide-148
SLIDE 148

. . . . . .

13th Hilbert’s problem

◮ Function of ≥ 3 variables: e.g., solution of a polynomial of

degree 7 as function of its coefficients

◮ Is it possible to represent this function as a composition of

functions of at most 2 variables?

◮ Yes, with weird functions (Cantor) ◮ Yes, even with continuous functions (Kolmogorov, Arnold)

Circuit version: explicit function

n n n n

(polynomial circuit size?) that cannot be represented as composition of O functions

n n

  • n. Not known.

Kolmogorov complexity version: we have three strings a b c

  • n a blackboard. It is allowed to write (add) a new string if it

simple relative to two of the strings on the board. Which strings we can obtain in O steps? Only strings of small complexity relative to a b c, but not all of them (for random a b c)

slide-149
SLIDE 149

. . . . . .

13th Hilbert’s problem

◮ Function of ≥ 3 variables: e.g., solution of a polynomial of

degree 7 as function of its coefficients

◮ Is it possible to represent this function as a composition of

functions of at most 2 variables?

◮ Yes, with weird functions (Cantor) ◮ Yes, even with continuous functions (Kolmogorov, Arnold) ◮ Circuit version: explicit function Bn × Bn × Bn → Bn

(polynomial circuit size?) that cannot be represented as composition of O(1) functions Bn × Bn → Bn. Not known. Kolmogorov complexity version: we have three strings a b c

  • n a blackboard. It is allowed to write (add) a new string if it

simple relative to two of the strings on the board. Which strings we can obtain in O steps? Only strings of small complexity relative to a b c, but not all of them (for random a b c)

slide-150
SLIDE 150

. . . . . .

13th Hilbert’s problem

◮ Function of ≥ 3 variables: e.g., solution of a polynomial of

degree 7 as function of its coefficients

◮ Is it possible to represent this function as a composition of

functions of at most 2 variables?

◮ Yes, with weird functions (Cantor) ◮ Yes, even with continuous functions (Kolmogorov, Arnold) ◮ Circuit version: explicit function Bn × Bn × Bn → Bn

(polynomial circuit size?) that cannot be represented as composition of O(1) functions Bn × Bn → Bn. Not known.

◮ Kolmogorov complexity version: we have three strings a, b, c

  • n a blackboard. It is allowed to write (add) a new string if it

simple relative to two of the strings on the board. Which strings we can obtain in O(1) steps? Only strings of small complexity relative to a, b, c, but not all of them (for random a, b, c)

slide-151
SLIDE 151

. . . . . .

Secret sharing

secret s and three people; any two are able to reconstruct the secret working together; but each one in isolation has no information about it assume s (a field); take random a and tell the people a, a s and a s (assuming in )

  • ther secret sharing schemes; how long should be secrets

(not understood) Kolmogorov complexity settings: for a given s find a b c such that C s a C s b C s c C s C s a b C s a c C s b c Some relation between Kolmogorov and traditional settings (Romashchenko, Kaced)

slide-152
SLIDE 152

. . . . . .

Secret sharing

◮ secret s and three people; any two are able to reconstruct the

secret working together; but each one in isolation has no information about it assume s (a field); take random a and tell the people a, a s and a s (assuming in )

  • ther secret sharing schemes; how long should be secrets

(not understood) Kolmogorov complexity settings: for a given s find a b c such that C s a C s b C s c C s C s a b C s a c C s b c Some relation between Kolmogorov and traditional settings (Romashchenko, Kaced)

slide-153
SLIDE 153

. . . . . .

Secret sharing

◮ secret s and three people; any two are able to reconstruct the

secret working together; but each one in isolation has no information about it

◮ assume s ∈ F (a field); take random a and tell the people a,

a + s and a + 2s (assuming 2 = 0 in F)

  • ther secret sharing schemes; how long should be secrets

(not understood) Kolmogorov complexity settings: for a given s find a b c such that C s a C s b C s c C s C s a b C s a c C s b c Some relation between Kolmogorov and traditional settings (Romashchenko, Kaced)

slide-154
SLIDE 154

. . . . . .

Secret sharing

◮ secret s and three people; any two are able to reconstruct the

secret working together; but each one in isolation has no information about it

◮ assume s ∈ F (a field); take random a and tell the people a,

a + s and a + 2s (assuming 2 = 0 in F)

◮ other secret sharing schemes; how long should be secrets

(not understood) Kolmogorov complexity settings: for a given s find a b c such that C s a C s b C s c C s C s a b C s a c C s b c Some relation between Kolmogorov and traditional settings (Romashchenko, Kaced)

slide-155
SLIDE 155

. . . . . .

Secret sharing

◮ secret s and three people; any two are able to reconstruct the

secret working together; but each one in isolation has no information about it

◮ assume s ∈ F (a field); take random a and tell the people a,

a + s and a + 2s (assuming 2 = 0 in F)

◮ other secret sharing schemes; how long should be secrets

(not understood)

◮ Kolmogorov complexity settings: for a given s find a, b, c

such that C(s|a), C(s|b), C(s|c) ≈ C(s); C(s|a, b), C(s|a, c), C(s|b, c) ≈ 0 Some relation between Kolmogorov and traditional settings (Romashchenko, Kaced)

slide-156
SLIDE 156

. . . . . .

Secret sharing

◮ secret s and three people; any two are able to reconstruct the

secret working together; but each one in isolation has no information about it

◮ assume s ∈ F (a field); take random a and tell the people a,

a + s and a + 2s (assuming 2 = 0 in F)

◮ other secret sharing schemes; how long should be secrets

(not understood)

◮ Kolmogorov complexity settings: for a given s find a, b, c

such that C(s|a), C(s|b), C(s|c) ≈ C(s); C(s|a, b), C(s|a, c), C(s|b, c) ≈ 0

◮ Some relation between Kolmogorov and traditional settings

(Romashchenko, Kaced)

slide-157
SLIDE 157

. . . . . .

Quasi-cryptography

Alice has some information a Bob wants to let her know some b by sending some message f in such a way that Eve gets minimal information about b Formally: for given a, b find f such that C b a f and C b f max. Theorem (Muchnik): it is always possible to have C b f min C b C a Full version: Eve knows some c and we want to send message of a minimal possible length C b a

slide-158
SLIDE 158

. . . . . .

Quasi-cryptography

◮ Alice has some information a

Bob wants to let her know some b by sending some message f in such a way that Eve gets minimal information about b Formally: for given a, b find f such that C b a f and C b f max. Theorem (Muchnik): it is always possible to have C b f min C b C a Full version: Eve knows some c and we want to send message of a minimal possible length C b a

slide-159
SLIDE 159

. . . . . .

Quasi-cryptography

◮ Alice has some information a ◮ Bob wants to let her know some b

by sending some message f in such a way that Eve gets minimal information about b Formally: for given a, b find f such that C b a f and C b f max. Theorem (Muchnik): it is always possible to have C b f min C b C a Full version: Eve knows some c and we want to send message of a minimal possible length C b a

slide-160
SLIDE 160

. . . . . .

Quasi-cryptography

◮ Alice has some information a ◮ Bob wants to let her know some b ◮ by sending some message f

in such a way that Eve gets minimal information about b Formally: for given a, b find f such that C b a f and C b f max. Theorem (Muchnik): it is always possible to have C b f min C b C a Full version: Eve knows some c and we want to send message of a minimal possible length C b a

slide-161
SLIDE 161

. . . . . .

Quasi-cryptography

◮ Alice has some information a ◮ Bob wants to let her know some b ◮ by sending some message f ◮ in such a way that Eve gets minimal information about b

Formally: for given a, b find f such that C b a f and C b f max. Theorem (Muchnik): it is always possible to have C b f min C b C a Full version: Eve knows some c and we want to send message of a minimal possible length C b a

slide-162
SLIDE 162

. . . . . .

Quasi-cryptography

◮ Alice has some information a ◮ Bob wants to let her know some b ◮ by sending some message f ◮ in such a way that Eve gets minimal information about b ◮ Formally: for given a, b find f such that C(b|a, f) ≈ 0 and

C(b|f) → max. Theorem (Muchnik): it is always possible to have C b f min C b C a Full version: Eve knows some c and we want to send message of a minimal possible length C b a

slide-163
SLIDE 163

. . . . . .

Quasi-cryptography

◮ Alice has some information a ◮ Bob wants to let her know some b ◮ by sending some message f ◮ in such a way that Eve gets minimal information about b ◮ Formally: for given a, b find f such that C(b|a, f) ≈ 0 and

C(b|f) → max.

◮ Theorem (Muchnik): it is always possible to have

C(b|f) ≈ min(C(b), C(a)) Full version: Eve knows some c and we want to send message of a minimal possible length C b a

slide-164
SLIDE 164

. . . . . .

Quasi-cryptography

◮ Alice has some information a ◮ Bob wants to let her know some b ◮ by sending some message f ◮ in such a way that Eve gets minimal information about b ◮ Formally: for given a, b find f such that C(b|a, f) ≈ 0 and

C(b|f) → max.

◮ Theorem (Muchnik): it is always possible to have

C(b|f) ≈ min(C(b), C(a))

◮ Full version: Eve knows some c and we want to send

message of a minimal possible length C(b|a)

slide-165
SLIDE 165

. . . . . .

Andrej Muchnik (1958-2007)