Channel Capacity and Isoperimetric Inequality Iftach Haitner Tel - - PowerPoint PPT Presentation

channel capacity and isoperimetric inequality
SMART_READER_LITE
LIVE PREVIEW

Channel Capacity and Isoperimetric Inequality Iftach Haitner Tel - - PowerPoint PPT Presentation

Application of Information Theory, Lecture 5 Channel Capacity and Isoperimetric Inequality Iftach Haitner Tel Aviv University. November 25, 2014 Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 1 / 21 Part


slide-1
SLIDE 1

Application of Information Theory, Lecture 5

Channel Capacity and Isoperimetric Inequality

Iftach Haitner

Tel Aviv University.

November 25, 2014

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 1 / 21

slide-2
SLIDE 2

Part I Channel Capacity

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 2 / 21

slide-3
SLIDE 3

The problem

◮ We want to send a message x = (x1, . . . , xm) ∈ {0, 1}m, but the

communication channel is faulty

slide-4
SLIDE 4

The problem

◮ We want to send a message x = (x1, . . . , xm) ∈ {0, 1}m, but the

communication channel is faulty

◮ Each bit is (independently) flipped w.p. p (e.g., 0.1)

slide-5
SLIDE 5

The problem

◮ We want to send a message x = (x1, . . . , xm) ∈ {0, 1}m, but the

communication channel is faulty

◮ Each bit is (independently) flipped w.p. p (e.g., 0.1) 1 1 1 − p p p 1 − p

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 3 / 21

slide-6
SLIDE 6

The problem

◮ We want to send a message x = (x1, . . . , xm) ∈ {0, 1}m, but the

communication channel is faulty

◮ Each bit is (independently) flipped w.p. p (e.g., 0.1) 1 1 1 − p p p 1 − p ◮ (expected) Error rate is p

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 3 / 21

slide-7
SLIDE 7

The problem

◮ We want to send a message x = (x1, . . . , xm) ∈ {0, 1}m, but the

communication channel is faulty

◮ Each bit is (independently) flipped w.p. p (e.g., 0.1) 1 1 1 − p p p 1 − p ◮ (expected) Error rate is p ◮ Such “channel” is called Binary Symmetric Channel (BSC)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 3 / 21

slide-8
SLIDE 8

The problem

◮ We want to send a message x = (x1, . . . , xm) ∈ {0, 1}m, but the

communication channel is faulty

◮ Each bit is (independently) flipped w.p. p (e.g., 0.1) 1 1 1 − p p p 1 − p ◮ (expected) Error rate is p ◮ Such “channel” is called Binary Symmetric Channel (BSC) ◮ When sending m bits, we have ≈ pm errors

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 3 / 21

slide-9
SLIDE 9

The problem

◮ We want to send a message x = (x1, . . . , xm) ∈ {0, 1}m, but the

communication channel is faulty

◮ Each bit is (independently) flipped w.p. p (e.g., 0.1) 1 1 1 − p p p 1 − p ◮ (expected) Error rate is p ◮ Such “channel” is called Binary Symmetric Channel (BSC) ◮ When sending m bits, we have ≈ pm errors ◮ Can we send bits with smaller error?

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 3 / 21

slide-10
SLIDE 10

Solution

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 4 / 21

slide-11
SLIDE 11

Solution

◮ Obvious solution is “error correction codes (ECC)”

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 4 / 21

slide-12
SLIDE 12

Solution

◮ Obvious solution is “error correction codes (ECC)” ◮ Most simple example: send each bit three times, and take majority

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 4 / 21

slide-13
SLIDE 13

Solution

◮ Obvious solution is “error correction codes (ECC)” ◮ Most simple example: send each bit three times, and take majority ◮ Error happens if the channel errs at least twice

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 4 / 21

slide-14
SLIDE 14

Solution

◮ Obvious solution is “error correction codes (ECC)” ◮ Most simple example: send each bit three times, and take majority ◮ Error happens if the channel errs at least twice ◮ For p = 0.1: happens w.p.

3p2(1 − p) + p3 = 3 · 0.01 · 0.9 + 0.001 = 0.028

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 4 / 21

slide-15
SLIDE 15

Solution

◮ Obvious solution is “error correction codes (ECC)” ◮ Most simple example: send each bit three times, and take majority ◮ Error happens if the channel errs at least twice ◮ For p = 0.1: happens w.p.

3p2(1 − p) + p3 = 3 · 0.01 · 0.9 + 0.001 = 0.028

◮ Error rate: .028

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 4 / 21

slide-16
SLIDE 16

Solution

◮ Obvious solution is “error correction codes (ECC)” ◮ Most simple example: send each bit three times, and take majority ◮ Error happens if the channel errs at least twice ◮ For p = 0.1: happens w.p.

3p2(1 − p) + p3 = 3 · 0.01 · 0.9 + 0.001 = 0.028

◮ Error rate: .028 ◮ Transmission rate: 1/3 (i.e., # of bits recovered / #of bits transmitted)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 4 / 21

slide-17
SLIDE 17

Solution

◮ Obvious solution is “error correction codes (ECC)” ◮ Most simple example: send each bit three times, and take majority ◮ Error happens if the channel errs at least twice ◮ For p = 0.1: happens w.p.

3p2(1 − p) + p3 = 3 · 0.01 · 0.9 + 0.001 = 0.028

◮ Error rate: .028 ◮ Transmission rate: 1/3 (i.e., # of bits recovered / #of bits transmitted) ◮ We reduced the error rate by reducing the transmission rate.

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 4 / 21

slide-18
SLIDE 18

Solution

◮ Obvious solution is “error correction codes (ECC)” ◮ Most simple example: send each bit three times, and take majority ◮ Error happens if the channel errs at least twice ◮ For p = 0.1: happens w.p.

3p2(1 − p) + p3 = 3 · 0.01 · 0.9 + 0.001 = 0.028

◮ Error rate: .028 ◮ Transmission rate: 1/3 (i.e., # of bits recovered / #of bits transmitted) ◮ We reduced the error rate by reducing the transmission rate. ◮ Can we reduce the error rate, without reducing the transmitting rate too

much?

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 4 / 21

slide-19
SLIDE 19

Solution

◮ Obvious solution is “error correction codes (ECC)” ◮ Most simple example: send each bit three times, and take majority ◮ Error happens if the channel errs at least twice ◮ For p = 0.1: happens w.p.

3p2(1 − p) + p3 = 3 · 0.01 · 0.9 + 0.001 = 0.028

◮ Error rate: .028 ◮ Transmission rate: 1/3 (i.e., # of bits recovered / #of bits transmitted) ◮ We reduced the error rate by reducing the transmission rate. ◮ Can we reduce the error rate, without reducing the transmitting rate too

much?

◮ Before Shannon it was believed that very small error rate requires very

small transamination rate.

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 4 / 21

slide-20
SLIDE 20

Shannon’s result

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 5 / 21

slide-21
SLIDE 21

Shannon’s result

◮ Shannon showed that you can reduce the error rate towards 0, without

reducing the transmission rate towards 0

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 5 / 21

slide-22
SLIDE 22

Shannon’s result

◮ Shannon showed that you can reduce the error rate towards 0, without

reducing the transmission rate towards 0

◮ For any c < Cp, exists a code with transmission rate c that is correct w.p.

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 5 / 21

slide-23
SLIDE 23

Shannon’s result

◮ Shannon showed that you can reduce the error rate towards 0, without

reducing the transmission rate towards 0

◮ For any c < Cp, exists a code with transmission rate c that is correct w.p. ◮ Example: for p = .1, Cp > 1

2.

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 5 / 21

slide-24
SLIDE 24

Shannon’s result

◮ Shannon showed that you can reduce the error rate towards 0, without

reducing the transmission rate towards 0

◮ For any c < Cp, exists a code with transmission rate c that is correct w.p. ◮ Example: for p = .1, Cp > 1

2.

Hence, for sending x = (x1, . . . , xm), one can send 2m bits, such that x is recovered w.p. close to 1

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 5 / 21

slide-25
SLIDE 25

Shannon’s result

◮ Shannon showed that you can reduce the error rate towards 0, without

reducing the transmission rate towards 0

◮ For any c < Cp, exists a code with transmission rate c that is correct w.p. ◮ Example: for p = .1, Cp > 1

2.

Hence, for sending x = (x1, . . . , xm), one can send 2m bits, such that x is recovered w.p. close to 1

◮ More generally, ∀p ∈ [0, 1]

∃Cp such that for sending x ∈ {0, 1}m, one can send ≈ m

Cp bits, and x is recovered w.p. close to 1

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 5 / 21

slide-26
SLIDE 26

Shannon’s result

◮ Shannon showed that you can reduce the error rate towards 0, without

reducing the transmission rate towards 0

◮ For any c < Cp, exists a code with transmission rate c that is correct w.p. ◮ Example: for p = .1, Cp > 1

2.

Hence, for sending x = (x1, . . . , xm), one can send 2m bits, such that x is recovered w.p. close to 1

◮ More generally, ∀p ∈ [0, 1]

∃Cp such that for sending x ∈ {0, 1}m, one can send ≈ m

Cp bits, and x is recovered w.p. close to 1

◮ Cp might be 0 (i.e., for p = 1

2)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 5 / 21

slide-27
SLIDE 27

Shannon’s result

◮ Shannon showed that you can reduce the error rate towards 0, without

reducing the transmission rate towards 0

◮ For any c < Cp, exists a code with transmission rate c that is correct w.p. ◮ Example: for p = .1, Cp > 1

2.

Hence, for sending x = (x1, . . . , xm), one can send 2m bits, such that x is recovered w.p. close to 1

◮ More generally, ∀p ∈ [0, 1]

∃Cp such that for sending x ∈ {0, 1}m, one can send ≈ m

Cp bits, and x is recovered w.p. close to 1

◮ Cp might be 0 (i.e., for p = 1

2)

◮ A revolution in EE and the whole world

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 5 / 21

slide-28
SLIDE 28

Error correction code

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 6 / 21

slide-29
SLIDE 29

Error correction code

◮ Message to send x = (x1, . . . , xm) ∈ {0, 1}m

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 6 / 21

slide-30
SLIDE 30

Error correction code

◮ Message to send x = (x1, . . . , xm) ∈ {0, 1}m ◮ Encoding scheme: f : {0, 1}m → {0, 1}n

(n > m)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 6 / 21

slide-31
SLIDE 31

Error correction code

◮ Message to send x = (x1, . . . , xm) ∈ {0, 1}m ◮ Encoding scheme: f : {0, 1}m → {0, 1}n

(n > m)

◮ Decoding scheme: g : {0, 1}n → {0, 1}m

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 6 / 21

slide-32
SLIDE 32

Error correction code

◮ Message to send x = (x1, . . . , xm) ∈ {0, 1}m ◮ Encoding scheme: f : {0, 1}m → {0, 1}n

(n > m)

◮ Decoding scheme: g : {0, 1}n → {0, 1}m ◮

m n — transmission rate

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 6 / 21

slide-33
SLIDE 33

Error correction code

◮ Message to send x = (x1, . . . , xm) ∈ {0, 1}m ◮ Encoding scheme: f : {0, 1}m → {0, 1}n

(n > m)

◮ Decoding scheme: g : {0, 1}n → {0, 1}m ◮

m n — transmission rate

◮ Sender sends f(x) rather than x

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 6 / 21

slide-34
SLIDE 34

Error correction code

◮ Message to send x = (x1, . . . , xm) ∈ {0, 1}m ◮ Encoding scheme: f : {0, 1}m → {0, 1}n

(n > m)

◮ Decoding scheme: g : {0, 1}n → {0, 1}m ◮

m n — transmission rate

◮ Sender sends f(x) rather than x ◮ Receiver decodes the message by applying g

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 6 / 21

slide-35
SLIDE 35

Error correction code

◮ Message to send x = (x1, . . . , xm) ∈ {0, 1}m ◮ Encoding scheme: f : {0, 1}m → {0, 1}n

(n > m)

◮ Decoding scheme: g : {0, 1}n → {0, 1}m ◮

m n — transmission rate

◮ Sender sends f(x) rather than x ◮ Receiver decodes the message by applying g ◮

x

  • m bits

encoding

− → f(x)

  • n bits

channel

− → f(x) ⊕ Z

  • bitwise XOR

decoding

− → g(f(x) ⊕ Z) Z = (Z1, . . . , Zn) where Z1, . . . , Zn iid ∼ (1 − p, p) (i.e., over {0, 1} with Pr [Zi = 1] = p)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 6 / 21

slide-36
SLIDE 36

Error correction code

◮ Message to send x = (x1, . . . , xm) ∈ {0, 1}m ◮ Encoding scheme: f : {0, 1}m → {0, 1}n

(n > m)

◮ Decoding scheme: g : {0, 1}n → {0, 1}m ◮

m n — transmission rate

◮ Sender sends f(x) rather than x ◮ Receiver decodes the message by applying g ◮

x

  • m bits

encoding

− → f(x)

  • n bits

channel

− → f(x) ⊕ Z

  • bitwise XOR

decoding

− → g(f(x) ⊕ Z) Z = (Z1, . . . , Zn) where Z1, . . . , Zn iid ∼ (1 − p, p) (i.e., over {0, 1} with Pr [Zi = 1] = p)

◮ We hope g(f(x) ⊕ Z) = x

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 6 / 21

slide-37
SLIDE 37

Error correction code

◮ Message to send x = (x1, . . . , xm) ∈ {0, 1}m ◮ Encoding scheme: f : {0, 1}m → {0, 1}n

(n > m)

◮ Decoding scheme: g : {0, 1}n → {0, 1}m ◮

m n — transmission rate

◮ Sender sends f(x) rather than x ◮ Receiver decodes the message by applying g ◮

x

  • m bits

encoding

− → f(x)

  • n bits

channel

− → f(x) ⊕ Z

  • bitwise XOR

decoding

− → g(f(x) ⊕ Z) Z = (Z1, . . . , Zn) where Z1, . . . , Zn iid ∼ (1 − p, p) (i.e., over {0, 1} with Pr [Zi = 1] = p)

◮ We hope g(f(x) ⊕ Z) = x ◮ ECCs are everywhere

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 6 / 21

slide-38
SLIDE 38

Error correction code

◮ Message to send x = (x1, . . . , xm) ∈ {0, 1}m ◮ Encoding scheme: f : {0, 1}m → {0, 1}n

(n > m)

◮ Decoding scheme: g : {0, 1}n → {0, 1}m ◮

m n — transmission rate

◮ Sender sends f(x) rather than x ◮ Receiver decodes the message by applying g ◮

x

  • m bits

encoding

− → f(x)

  • n bits

channel

− → f(x) ⊕ Z

  • bitwise XOR

decoding

− → g(f(x) ⊕ Z) Z = (Z1, . . . , Zn) where Z1, . . . , Zn iid ∼ (1 − p, p) (i.e., over {0, 1} with Pr [Zi = 1] = p)

◮ We hope g(f(x) ⊕ Z) = x ◮ ECCs are everywhere ◮ ECC Vs compression

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 6 / 21

slide-39
SLIDE 39

Shannon’s theorem

Theorem 1 ∀p ∃Cp, s.t. ∀ε > 0 ∃mε, s.t. ∀m > mε and n > m( 1

Cp + ε),

∃ f : {0, 1}m → {0, 1}n and g : {0, 1}n → {0, 1}m, s.t. ∀x ∈ {0, 1}m: Pr

z←Z=(Z1,...,Zn) [g(f(x) ⊕ z) = x] ≤ ε

for Z1, . . . , Zn iid ∼ (1 − p, p).

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 7 / 21

slide-40
SLIDE 40

Shannon’s theorem

Theorem 1 ∀p ∃Cp, s.t. ∀ε > 0 ∃mε, s.t. ∀m > mε and n > m( 1

Cp + ε),

∃ f : {0, 1}m → {0, 1}n and g : {0, 1}n → {0, 1}m, s.t. ∀x ∈ {0, 1}m: Pr

z←Z=(Z1,...,Zn) [g(f(x) ⊕ z) = x] ≤ ε

for Z1, . . . , Zn iid ∼ (1 − p, p).

◮ Cp = 1 − h(p) — the channel capacity

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 7 / 21

slide-41
SLIDE 41

Shannon’s theorem

Theorem 1 ∀p ∃Cp, s.t. ∀ε > 0 ∃mε, s.t. ∀m > mε and n > m( 1

Cp + ε),

∃ f : {0, 1}m → {0, 1}n and g : {0, 1}n → {0, 1}m, s.t. ∀x ∈ {0, 1}m: Pr

z←Z=(Z1,...,Zn) [g(f(x) ⊕ z) = x] ≤ ε

for Z1, . . . , Zn iid ∼ (1 − p, p).

◮ Cp = 1 − h(p) — the channel capacity

p = .1 = ⇒ Cp = 0.5310 > 1

2

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 7 / 21

slide-42
SLIDE 42

Shannon’s theorem

Theorem 1 ∀p ∃Cp, s.t. ∀ε > 0 ∃mε, s.t. ∀m > mε and n > m( 1

Cp + ε),

∃ f : {0, 1}m → {0, 1}n and g : {0, 1}n → {0, 1}m, s.t. ∀x ∈ {0, 1}m: Pr

z←Z=(Z1,...,Zn) [g(f(x) ⊕ z) = x] ≤ ε

for Z1, . . . , Zn iid ∼ (1 − p, p).

◮ Cp = 1 − h(p) — the channel capacity

p = .1 = ⇒ Cp = 0.5310 > 1

2

p = .25 = ⇒ Cp ≈ 1

5

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 7 / 21

slide-43
SLIDE 43

Shannon’s theorem

Theorem 1 ∀p ∃Cp, s.t. ∀ε > 0 ∃mε, s.t. ∀m > mε and n > m( 1

Cp + ε),

∃ f : {0, 1}m → {0, 1}n and g : {0, 1}n → {0, 1}m, s.t. ∀x ∈ {0, 1}m: Pr

z←Z=(Z1,...,Zn) [g(f(x) ⊕ z) = x] ≤ ε

for Z1, . . . , Zn iid ∼ (1 − p, p).

◮ Cp = 1 − h(p) — the channel capacity

p = .1 = ⇒ Cp = 0.5310 > 1

2

p = .25 = ⇒ Cp ≈ 1

5

◮ Tight theorem

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 7 / 21

slide-44
SLIDE 44

Shannon’s theorem

Theorem 1 ∀p ∃Cp, s.t. ∀ε > 0 ∃mε, s.t. ∀m > mε and n > m( 1

Cp + ε),

∃ f : {0, 1}m → {0, 1}n and g : {0, 1}n → {0, 1}m, s.t. ∀x ∈ {0, 1}m: Pr

z←Z=(Z1,...,Zn) [g(f(x) ⊕ z) = x] ≤ ε

for Z1, . . . , Zn iid ∼ (1 − p, p).

◮ Cp = 1 − h(p) — the channel capacity

p = .1 = ⇒ Cp = 0.5310 > 1

2

p = .25 = ⇒ Cp ≈ 1

5

◮ Tight theorem ◮ We prove a weaker variant that holds w.h.p. over x ← {0, 1}m

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 7 / 21

slide-45
SLIDE 45

Hamming distance

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 8 / 21

slide-46
SLIDE 46

Hamming distance

◮ For y = (y1, . . . , yn) ∈ {0, 1}n, let |y| =

i yi — Hamming weight of y

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 8 / 21

slide-47
SLIDE 47

Hamming distance

◮ For y = (y1, . . . , yn) ∈ {0, 1}n, let |y| =

i yi — Hamming weight of y

◮ |y − y′| = |y ⊕ y′| — Hamming distance of y from y′; # of places differ.

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 8 / 21

slide-48
SLIDE 48

Proving the theorem

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 9 / 21

slide-49
SLIDE 49

Proving the theorem

◮ Fix p ∈ [0, 1

2) and ε > 0, and let m > mε and n ≥ m( 1 Cp + ε), for mε to be

determined by the analysis.

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 9 / 21

slide-50
SLIDE 50

Proving the theorem

◮ Fix p ∈ [0, 1

2) and ε > 0, and let m > mε and n ≥ m( 1 Cp + ε), for mε to be

determined by the analysis.

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 9 / 21

slide-51
SLIDE 51

Proving the theorem

◮ Fix p ∈ [0, 1

2) and ε > 0, and let m > mε and n ≥ m( 1 Cp + ε), for mε to be

determined by the analysis. (Recall Cp = 1 − h(p)).

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 9 / 21

slide-52
SLIDE 52

Proving the theorem

◮ Fix p ∈ [0, 1

2) and ε > 0, and let m > mε and n ≥ m( 1 Cp + ε), for mε to be

determined by the analysis. (Recall Cp = 1 − h(p)).

◮ We show ∃ f : {0, 1}m → {0, 1}n and g : {0, 1}n → {0, 1}m, s.t.

Prx←{0,1}m [g(f(x) ⊕ Z) = x] ≤ ε

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 9 / 21

slide-53
SLIDE 53

Proving the theorem

◮ Fix p ∈ [0, 1

2) and ε > 0, and let m > mε and n ≥ m( 1 Cp + ε), for mε to be

determined by the analysis. (Recall Cp = 1 − h(p)).

◮ We show ∃ f : {0, 1}m → {0, 1}n and g : {0, 1}n → {0, 1}m, s.t.

Prx←{0,1}m [g(f(x) ⊕ Z) = x] ≤ ε

◮ g(y) returns argminx′∈{0,1}m |y − f(x′)|

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 9 / 21

slide-54
SLIDE 54

Proving the theorem

◮ Fix p ∈ [0, 1

2) and ε > 0, and let m > mε and n ≥ m( 1 Cp + ε), for mε to be

determined by the analysis. (Recall Cp = 1 − h(p)).

◮ We show ∃ f : {0, 1}m → {0, 1}n and g : {0, 1}n → {0, 1}m, s.t.

Prx←{0,1}m [g(f(x) ⊕ Z) = x] ≤ ε

◮ g(y) returns argminx′∈{0,1}m |y − f(x′)| ◮ So it all boils down to finding f s.t.

Prx←{0,1}m;y=f(x)⊕Z

  • |f(x) − y| < minx′∈{0,1}m\{x} |f(x′) − y|
  • ≥ 1 − ε

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 9 / 21

slide-55
SLIDE 55

Proving the theorem

◮ Fix p ∈ [0, 1

2) and ε > 0, and let m > mε and n ≥ m( 1 Cp + ε), for mε to be

determined by the analysis. (Recall Cp = 1 − h(p)).

◮ We show ∃ f : {0, 1}m → {0, 1}n and g : {0, 1}n → {0, 1}m, s.t.

Prx←{0,1}m [g(f(x) ⊕ Z) = x] ≤ ε

◮ g(y) returns argminx′∈{0,1}m |y − f(x′)| ◮ So it all boils down to finding f s.t.

Prx←{0,1}m;y=f(x)⊕Z

  • |f(x) − y| < minx′∈{0,1}m\{x} |f(x′) − y|
  • ≥ 1 − ε

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 9 / 21

slide-56
SLIDE 56

Proving the theorem

◮ Fix p ∈ [0, 1

2) and ε > 0, and let m > mε and n ≥ m( 1 Cp + ε), for mε to be

determined by the analysis. (Recall Cp = 1 − h(p)).

◮ We show ∃ f : {0, 1}m → {0, 1}n and g : {0, 1}n → {0, 1}m, s.t.

Prx←{0,1}m [g(f(x) ⊕ Z) = x] ≤ ε

◮ g(y) returns argminx′∈{0,1}m |y − f(x′)| ◮ So it all boils down to finding f s.t.

Prx←{0,1}m;y=f(x)⊕Z

  • |f(x) − y| < minx′∈{0,1}m\{x} |f(x′) − y|
  • ≥ 1 − ε

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 9 / 21

slide-57
SLIDE 57

Proving the theorem

◮ Fix p ∈ [0, 1

2) and ε > 0, and let m > mε and n ≥ m( 1 Cp + ε), for mε to be

determined by the analysis. (Recall Cp = 1 − h(p)).

◮ We show ∃ f : {0, 1}m → {0, 1}n and g : {0, 1}n → {0, 1}m, s.t.

Prx←{0,1}m [g(f(x) ⊕ Z) = x] ≤ ε

◮ g(y) returns argminx′∈{0,1}m |y − f(x′)| ◮ So it all boils down to finding f s.t.

Prx←{0,1}m;y=f(x)⊕Z

  • |f(x) − y| < minx′∈{0,1}m\{x} |f(x′) − y|
  • ≥ 1 − ε

◮ Idea: for p′ > p to be determined later, find f

s.t. w.h.p. over x and Z: (1) |f(x) ⊕ Z, f(x)| ≤ p′n (2) |f(x) ⊕ Z, f(x′)| > p′n for all x′ = x

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 9 / 21

slide-58
SLIDE 58

Proving the theorem

◮ Fix p ∈ [0, 1

2) and ε > 0, and let m > mε and n ≥ m( 1 Cp + ε), for mε to be

determined by the analysis. (Recall Cp = 1 − h(p)).

◮ We show ∃ f : {0, 1}m → {0, 1}n and g : {0, 1}n → {0, 1}m, s.t.

Prx←{0,1}m [g(f(x) ⊕ Z) = x] ≤ ε

◮ g(y) returns argminx′∈{0,1}m |y − f(x′)| ◮ So it all boils down to finding f s.t.

Prx←{0,1}m;y=f(x)⊕Z

  • |f(x) − y| < minx′∈{0,1}m\{x} |f(x′) − y|
  • ≥ 1 − ε

◮ Idea: for p′ > p to be determined later, find f

s.t. w.h.p. over x and Z: (1) |f(x) ⊕ Z, f(x)| ≤ p′n (2) |f(x) ⊕ Z, f(x′)| > p′n for all x′ = x

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 9 / 21

slide-59
SLIDE 59

Proving the theorem

◮ Fix p ∈ [0, 1

2) and ε > 0, and let m > mε and n ≥ m( 1 Cp + ε), for mε to be

determined by the analysis. (Recall Cp = 1 − h(p)).

◮ We show ∃ f : {0, 1}m → {0, 1}n and g : {0, 1}n → {0, 1}m, s.t.

Prx←{0,1}m [g(f(x) ⊕ Z) = x] ≤ ε

◮ g(y) returns argminx′∈{0,1}m |y − f(x′)| ◮ So it all boils down to finding f s.t.

Prx←{0,1}m;y=f(x)⊕Z

  • |f(x) − y| < minx′∈{0,1}m\{x} |f(x′) − y|
  • ≥ 1 − ε

◮ Idea: for p′ > p to be determined later, find f

s.t. w.h.p. over x and Z: (1) |f(x) ⊕ Z, f(x)| ≤ p′n (2) |f(x) ⊕ Z, f(x′)| > p′n for all x′ = x

◮ We choose f uniformly at random (what does it mean?)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 9 / 21

slide-60
SLIDE 60

Proving the theorem

◮ Fix p ∈ [0, 1

2) and ε > 0, and let m > mε and n ≥ m( 1 Cp + ε), for mε to be

determined by the analysis. (Recall Cp = 1 − h(p)).

◮ We show ∃ f : {0, 1}m → {0, 1}n and g : {0, 1}n → {0, 1}m, s.t.

Prx←{0,1}m [g(f(x) ⊕ Z) = x] ≤ ε

◮ g(y) returns argminx′∈{0,1}m |y − f(x′)| ◮ So it all boils down to finding f s.t.

Prx←{0,1}m;y=f(x)⊕Z

  • |f(x) − y| < minx′∈{0,1}m\{x} |f(x′) − y|
  • ≥ 1 − ε

◮ Idea: for p′ > p to be determined later, find f

s.t. w.h.p. over x and Z: (1) |f(x) ⊕ Z, f(x)| ≤ p′n (2) |f(x) ⊕ Z, f(x′)| > p′n for all x′ = x

◮ We choose f uniformly at random (what does it mean?) ◮ Non-constructive proof

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 9 / 21

slide-61
SLIDE 61

Proving the theorem

◮ Fix p ∈ [0, 1

2) and ε > 0, and let m > mε and n ≥ m( 1 Cp + ε), for mε to be

determined by the analysis. (Recall Cp = 1 − h(p)).

◮ We show ∃ f : {0, 1}m → {0, 1}n and g : {0, 1}n → {0, 1}m, s.t.

Prx←{0,1}m [g(f(x) ⊕ Z) = x] ≤ ε

◮ g(y) returns argminx′∈{0,1}m |y − f(x′)| ◮ So it all boils down to finding f s.t.

Prx←{0,1}m;y=f(x)⊕Z

  • |f(x) − y| < minx′∈{0,1}m\{x} |f(x′) − y|
  • ≥ 1 − ε

◮ Idea: for p′ > p to be determined later, find f

s.t. w.h.p. over x and Z: (1) |f(x) ⊕ Z, f(x)| ≤ p′n (2) |f(x) ⊕ Z, f(x′)| > p′n for all x′ = x

◮ We choose f uniformly at random (what does it mean?) ◮ Non-constructive proof ◮ Probabilistic method

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 9 / 21

slide-62
SLIDE 62

Proving there exists good f

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 10 / 21

slide-63
SLIDE 63

Proving there exists good f

◮ Fix p′ > p such that

1 Cp′ − 1 Cp ≤ ε 2

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 10 / 21

slide-64
SLIDE 64

Proving there exists good f

◮ Fix p′ > p such that

1 Cp′ − 1 Cp ≤ ε 2

◮ For y ∈ {0, 1}n, let Bp′(y) = {y ∈ {0, 1}n : |y′ − y| ≤ p′n}

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 10 / 21

slide-65
SLIDE 65

Proving there exists good f

◮ Fix p′ > p such that

1 Cp′ − 1 Cp ≤ ε 2

◮ For y ∈ {0, 1}n, let Bp′(y) = {y ∈ {0, 1}n : |y′ − y| ≤ p′n}

(1) By weak low of large numbers, ∃n′ ∈ N s.t. ∀n ≥ n′ and ∀x ∈ {0, 1}m: αn := Prz←Z [(f(x) ⊕ z) / ∈ Bp′(f(x))] ≤ ε

2

(for any fixed f)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 10 / 21

slide-66
SLIDE 66

Proving there exists good f

◮ Fix p′ > p such that

1 Cp′ − 1 Cp ≤ ε 2

◮ For y ∈ {0, 1}n, let Bp′(y) = {y ∈ {0, 1}n : |y′ − y| ≤ p′n}

(1) By weak low of large numbers, ∃n′ ∈ N s.t. ∀n ≥ n′ and ∀x ∈ {0, 1}m: αn := Prz←Z [(f(x) ⊕ z) / ∈ Bp′(f(x))] ≤ ε

2

(for any fixed f)

◮ Fact (proved later): b(p′) = |Bp′(y)| ≤ 2n·h(p′)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 10 / 21

slide-67
SLIDE 67

Proving there exists good f

◮ Fix p′ > p such that

1 Cp′ − 1 Cp ≤ ε 2

◮ For y ∈ {0, 1}n, let Bp′(y) = {y ∈ {0, 1}n : |y′ − y| ≤ p′n}

(1) By weak low of large numbers, ∃n′ ∈ N s.t. ∀n ≥ n′ and ∀x ∈ {0, 1}m: αn := Prz←Z [(f(x) ⊕ z) / ∈ Bp′(f(x))] ≤ ε

2

(for any fixed f)

◮ Fact (proved later): b(p′) = |Bp′(y)| ≤ 2n·h(p′)

= ⇒ ∀x = x′ ∈ {0, 1}m: Prf,Z [f(x) ⊕ Z ∈ Bp′(f(x′))] = b(p′)

2n

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 10 / 21

slide-68
SLIDE 68

Proving there exists good f

◮ Fix p′ > p such that

1 Cp′ − 1 Cp ≤ ε 2

◮ For y ∈ {0, 1}n, let Bp′(y) = {y ∈ {0, 1}n : |y′ − y| ≤ p′n}

(1) By weak low of large numbers, ∃n′ ∈ N s.t. ∀n ≥ n′ and ∀x ∈ {0, 1}m: αn := Prz←Z [(f(x) ⊕ z) / ∈ Bp′(f(x))] ≤ ε

2

(for any fixed f)

◮ Fact (proved later): b(p′) = |Bp′(y)| ≤ 2n·h(p′)

= ⇒ ∀x = x′ ∈ {0, 1}m: Prf,Z [f(x) ⊕ Z ∈ Bp′(f(x′))] = b(p′)

2n

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 10 / 21

slide-69
SLIDE 69

Proving there exists good f

◮ Fix p′ > p such that

1 Cp′ − 1 Cp ≤ ε 2

◮ For y ∈ {0, 1}n, let Bp′(y) = {y ∈ {0, 1}n : |y′ − y| ≤ p′n}

(1) By weak low of large numbers, ∃n′ ∈ N s.t. ∀n ≥ n′ and ∀x ∈ {0, 1}m: αn := Prz←Z [(f(x) ⊕ z) / ∈ Bp′(f(x))] ≤ ε

2

(for any fixed f)

◮ Fact (proved later): b(p′) = |Bp′(y)| ≤ 2n·h(p′)

= ⇒ ∀x = x′ ∈ {0, 1}m: Prf,Z [f(x) ⊕ Z ∈ Bp′(f(x′))] = b(p′)

2n

≤ 2n·h(p′)

2n

= 2−nCp′

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 10 / 21

slide-70
SLIDE 70

Proving there exists good f

◮ Fix p′ > p such that

1 Cp′ − 1 Cp ≤ ε 2

◮ For y ∈ {0, 1}n, let Bp′(y) = {y ∈ {0, 1}n : |y′ − y| ≤ p′n}

(1) By weak low of large numbers, ∃n′ ∈ N s.t. ∀n ≥ n′ and ∀x ∈ {0, 1}m: αn := Prz←Z [(f(x) ⊕ z) / ∈ Bp′(f(x))] ≤ ε

2

(for any fixed f)

◮ Fact (proved later): b(p′) = |Bp′(y)| ≤ 2n·h(p′)

= ⇒ ∀x = x′ ∈ {0, 1}m: Prf,Z [f(x) ⊕ Z ∈ Bp′(f(x′))] = b(p′)

2n

≤ 2n·h(p′)

2n

= 2−nCp′ = ⇒ ∀x ∈ {0, 1}m: Prf,Z [∃x′ = x ∈ {0, 1}m : f(x′) ⊕ Z ∈ Bp′(f(x′))] ≤ 2m−nCp′

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 10 / 21

slide-71
SLIDE 71

Proving there exists good f

◮ Fix p′ > p such that

1 Cp′ − 1 Cp ≤ ε 2

◮ For y ∈ {0, 1}n, let Bp′(y) = {y ∈ {0, 1}n : |y′ − y| ≤ p′n}

(1) By weak low of large numbers, ∃n′ ∈ N s.t. ∀n ≥ n′ and ∀x ∈ {0, 1}m: αn := Prz←Z [(f(x) ⊕ z) / ∈ Bp′(f(x))] ≤ ε

2

(for any fixed f)

◮ Fact (proved later): b(p′) = |Bp′(y)| ≤ 2n·h(p′)

= ⇒ ∀x = x′ ∈ {0, 1}m: Prf,Z [f(x) ⊕ Z ∈ Bp′(f(x′))] = b(p′)

2n

≤ 2n·h(p′)

2n

= 2−nCp′ = ⇒ ∀x ∈ {0, 1}m: Prf,Z [∃x′ = x ∈ {0, 1}m : f(x′) ⊕ Z ∈ Bp′(f(x′))] ≤ 2m−nCp′ = ⇒ ∃f s.t. βm,n := Prx←{0,1}m [∃x′ = x ∈ {0, 1}m : f(x′) ⊕ Z ∈ Bp′(f(x′))] ≤ 2m−nCp′

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 10 / 21

slide-72
SLIDE 72

Proving there exists good f

◮ Fix p′ > p such that

1 Cp′ − 1 Cp ≤ ε 2

◮ For y ∈ {0, 1}n, let Bp′(y) = {y ∈ {0, 1}n : |y′ − y| ≤ p′n}

(1) By weak low of large numbers, ∃n′ ∈ N s.t. ∀n ≥ n′ and ∀x ∈ {0, 1}m: αn := Prz←Z [(f(x) ⊕ z) / ∈ Bp′(f(x))] ≤ ε

2

(for any fixed f)

◮ Fact (proved later): b(p′) = |Bp′(y)| ≤ 2n·h(p′)

= ⇒ ∀x = x′ ∈ {0, 1}m: Prf,Z [f(x) ⊕ Z ∈ Bp′(f(x′))] = b(p′)

2n

≤ 2n·h(p′)

2n

= 2−nCp′ = ⇒ ∀x ∈ {0, 1}m: Prf,Z [∃x′ = x ∈ {0, 1}m : f(x′) ⊕ Z ∈ Bp′(f(x′))] ≤ 2m−nCp′ = ⇒ ∃f s.t. βm,n := Prx←{0,1}m [∃x′ = x ∈ {0, 1}m : f(x′) ⊕ Z ∈ Bp′(f(x′))] ≤ 2m−nCp′ = ⇒ βm,n ≤ ε

2, for n ≥ m Cp′ − log ε + 1 = m( 1 Cp′ + 1−log ε m

)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 10 / 21

slide-73
SLIDE 73

Proving there exists good f

◮ Fix p′ > p such that

1 Cp′ − 1 Cp ≤ ε 2

◮ For y ∈ {0, 1}n, let Bp′(y) = {y ∈ {0, 1}n : |y′ − y| ≤ p′n}

(1) By weak low of large numbers, ∃n′ ∈ N s.t. ∀n ≥ n′ and ∀x ∈ {0, 1}m: αn := Prz←Z [(f(x) ⊕ z) / ∈ Bp′(f(x))] ≤ ε

2

(for any fixed f)

◮ Fact (proved later): b(p′) = |Bp′(y)| ≤ 2n·h(p′)

= ⇒ ∀x = x′ ∈ {0, 1}m: Prf,Z [f(x) ⊕ Z ∈ Bp′(f(x′))] = b(p′)

2n

≤ 2n·h(p′)

2n

= 2−nCp′ = ⇒ ∀x ∈ {0, 1}m: Prf,Z [∃x′ = x ∈ {0, 1}m : f(x′) ⊕ Z ∈ Bp′(f(x′))] ≤ 2m−nCp′ = ⇒ ∃f s.t. βm,n := Prx←{0,1}m [∃x′ = x ∈ {0, 1}m : f(x′) ⊕ Z ∈ Bp′(f(x′))] ≤ 2m−nCp′ = ⇒ βm,n ≤ ε

2, for n ≥ m Cp′ − log ε + 1 = m( 1 Cp′ + 1−log ε m

) (2) βm,n ≤ ε

2, for m ≥ m′ = 2(1−log ε) ε

and n ≥ m( 1

Cp + ε 2 + 1−log ε m

) = m( 1

Cp + ε)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 10 / 21

slide-74
SLIDE 74

Proving there exists good f

◮ Fix p′ > p such that

1 Cp′ − 1 Cp ≤ ε 2

◮ For y ∈ {0, 1}n, let Bp′(y) = {y ∈ {0, 1}n : |y′ − y| ≤ p′n}

(1) By weak low of large numbers, ∃n′ ∈ N s.t. ∀n ≥ n′ and ∀x ∈ {0, 1}m: αn := Prz←Z [(f(x) ⊕ z) / ∈ Bp′(f(x))] ≤ ε

2

(for any fixed f)

◮ Fact (proved later): b(p′) = |Bp′(y)| ≤ 2n·h(p′)

= ⇒ ∀x = x′ ∈ {0, 1}m: Prf,Z [f(x) ⊕ Z ∈ Bp′(f(x′))] = b(p′)

2n

≤ 2n·h(p′)

2n

= 2−nCp′ = ⇒ ∀x ∈ {0, 1}m: Prf,Z [∃x′ = x ∈ {0, 1}m : f(x′) ⊕ Z ∈ Bp′(f(x′))] ≤ 2m−nCp′ = ⇒ ∃f s.t. βm,n := Prx←{0,1}m [∃x′ = x ∈ {0, 1}m : f(x′) ⊕ Z ∈ Bp′(f(x′))] ≤ 2m−nCp′ = ⇒ βm,n ≤ ε

2, for n ≥ m Cp′ − log ε + 1 = m( 1 Cp′ + 1−log ε m

) (2) βm,n ≤ ε

2, for m ≥ m′ = 2(1−log ε) ε

and n ≥ m( 1

Cp + ε 2 + 1−log ε m

) = m( 1

Cp + ε)

◮ Hence, for m > mε = max{m′, n′} and n > m( 1

Cp + ε), it holds that

Prx←{0,1}m [g(f(x) ⊕ Z) = x] ≤ αn + βm,n ≤ ε.

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 10 / 21

slide-75
SLIDE 75

Why Cp = 1 − h(p)?

slide-76
SLIDE 76

Why Cp = 1 − h(p)?

◮ Let X ← {0, 1} , Z ∼ (1 − p, p) and Y = X ⊕ Z

slide-77
SLIDE 77

Why Cp = 1 − h(p)?

◮ Let X ← {0, 1} , Z ∼ (1 − p, p) and Y = X ⊕ Z

X Y

1 1 1 − p p p 1 − p

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 11 / 21

slide-78
SLIDE 78

Why Cp = 1 − h(p)?

◮ Let X ← {0, 1} , Z ∼ (1 − p, p) and Y = X ⊕ Z

X Y

1 1 1 − p p p 1 − p ◮ I(X; Y) = H(Y) − H(Y|X) = H(Y) − H(Z) = 1 − h(p) = Cp

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 11 / 21

slide-79
SLIDE 79

Why Cp = 1 − h(p)?

◮ Let X ← {0, 1} , Z ∼ (1 − p, p) and Y = X ⊕ Z

X Y

1 1 1 − p p p 1 − p ◮ I(X; Y) = H(Y) − H(Y|X) = H(Y) − H(Z) = 1 − h(p) = Cp ◮ Received bit “gives" Cp information about transmitted bit

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 11 / 21

slide-80
SLIDE 80

Why Cp = 1 − h(p)?

◮ Let X ← {0, 1} , Z ∼ (1 − p, p) and Y = X ⊕ Z

X Y

1 1 1 − p p p 1 − p ◮ I(X; Y) = H(Y) − H(Y|X) = H(Y) − H(Z) = 1 − h(p) = Cp ◮ Received bit “gives" Cp information about transmitted bit ◮ Hence, to recover m bits, we need to send at least m ·

1 Cp bits

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 11 / 21

slide-81
SLIDE 81

Size of bounding ball

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 12 / 21

slide-82
SLIDE 82

Size of bounding ball

Claim 2 For p ∈ [0, 1

2] and n ∈ N: it holds that ⌊pn⌋ k=0

n

k

  • ≤ 2n·h(p)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 12 / 21

slide-83
SLIDE 83

Size of bounding ball

Claim 2 For p ∈ [0, 1

2] and n ∈ N: it holds that ⌊pn⌋ k=0

n

k

  • ≤ 2n·h(p)

Proof in a few slides (we already saw that n

pn

  • ≈ 2n·h(p))

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 12 / 21

slide-84
SLIDE 84

Size of bounding ball

Claim 2 For p ∈ [0, 1

2] and n ∈ N: it holds that ⌊pn⌋ k=0

n

k

  • ≤ 2n·h(p)

Proof in a few slides (we already saw that n

pn

  • ≈ 2n·h(p))

Corollary 3 For y ∈ {0, 1}n and p ∈ [0, 1

2], let Bp(y) = {y ∈ {0, 1}n : |y′ − y| ≤ pn}. Then

|Bp(y)| = ⌊pn⌋

k=0

n

k

  • ≤ 2n·h(p)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 12 / 21

slide-85
SLIDE 85

Size of bounding ball

Claim 2 For p ∈ [0, 1

2] and n ∈ N: it holds that ⌊pn⌋ k=0

n

k

  • ≤ 2n·h(p)

Proof in a few slides (we already saw that n

pn

  • ≈ 2n·h(p))

Corollary 3 For y ∈ {0, 1}n and p ∈ [0, 1

2], let Bp(y) = {y ∈ {0, 1}n : |y′ − y| ≤ pn}. Then

|Bp(y)| = ⌊pn⌋

k=0

n

k

  • ≤ 2n·h(p)

Very useful estimation. Weaker variants follows by AEP or Stirling,

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 12 / 21

slide-86
SLIDE 86

Tightness

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 13 / 21

slide-87
SLIDE 87

Tightness

◮ X ← {0, 1}m, Z = (Z1, . . . , Zn) where Z1, . . . , Zn iid ∼ (1 − p, p)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 13 / 21

slide-88
SLIDE 88

Tightness

◮ X ← {0, 1}m, Z = (Z1, . . . , Zn) where Z1, . . . , Zn iid ∼ (1 − p, p) ◮

X

  • m bits

− → f(X)

  • n bits

− → f(X) ⊕ Z

  • Y

− → g(f(X) ⊕ Z)

  • g(Y)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 13 / 21

slide-89
SLIDE 89

Tightness

◮ X ← {0, 1}m, Z = (Z1, . . . , Zn) where Z1, . . . , Zn iid ∼ (1 − p, p) ◮

X

  • m bits

− → f(X)

  • n bits

− → f(X) ⊕ Z

  • Y

− → g(f(X) ⊕ Z)

  • g(Y)

◮ Assuming Pr [g(Y) = X] ≥ 1 − ε, we show nCp ≥ m(1 − ε) − 1

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 13 / 21

slide-90
SLIDE 90

Tightness

◮ X ← {0, 1}m, Z = (Z1, . . . , Zn) where Z1, . . . , Zn iid ∼ (1 − p, p) ◮

X

  • m bits

− → f(X)

  • n bits

− → f(X) ⊕ Z

  • Y

− → g(f(X) ⊕ Z)

  • g(Y)

◮ Assuming Pr [g(Y) = X] ≥ 1 − ε, we show nCp ≥ m(1 − ε) − 1 ◮ Compare to nCp > m(1 + εCp) in Thm 1

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 13 / 21

slide-91
SLIDE 91

Tightness

◮ X ← {0, 1}m, Z = (Z1, . . . , Zn) where Z1, . . . , Zn iid ∼ (1 − p, p) ◮

X

  • m bits

− → f(X)

  • n bits

− → f(X) ⊕ Z

  • Y

− → g(f(X) ⊕ Z)

  • g(Y)

◮ Assuming Pr [g(Y) = X] ≥ 1 − ε, we show nCp ≥ m(1 − ε) − 1 ◮ Compare to nCp > m(1 + εCp) in Thm 1 ◮ Hence, limε→0 m

n = Cp

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 13 / 21

slide-92
SLIDE 92

Tightness

◮ X ← {0, 1}m, Z = (Z1, . . . , Zn) where Z1, . . . , Zn iid ∼ (1 − p, p) ◮

X

  • m bits

− → f(X)

  • n bits

− → f(X) ⊕ Z

  • Y

− → g(f(X) ⊕ Z)

  • g(Y)

◮ Assuming Pr [g(Y) = X] ≥ 1 − ε, we show nCp ≥ m(1 − ε) − 1 ◮ Compare to nCp > m(1 + εCp) in Thm 1 ◮ Hence, limε→0 m

n = Cp

◮ By Fano, H(X|Y) ≤ h(ε) + ε log(2m − 1) ≤ 1 + εm

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 13 / 21

slide-93
SLIDE 93

Tightness

◮ X ← {0, 1}m, Z = (Z1, . . . , Zn) where Z1, . . . , Zn iid ∼ (1 − p, p) ◮

X

  • m bits

− → f(X)

  • n bits

− → f(X) ⊕ Z

  • Y

− → g(f(X) ⊕ Z)

  • g(Y)

◮ Assuming Pr [g(Y) = X] ≥ 1 − ε, we show nCp ≥ m(1 − ε) − 1 ◮ Compare to nCp > m(1 + εCp) in Thm 1 ◮ Hence, limε→0 m

n = Cp

◮ By Fano, H(X|Y) ≤ h(ε) + ε log(2m − 1) ≤ 1 + εm ◮ I(X; Y) = H(X) − H(X|Y) ≥ m − εm − 1 = m(1 − ε) − 1

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 13 / 21

slide-94
SLIDE 94

Tightness

◮ X ← {0, 1}m, Z = (Z1, . . . , Zn) where Z1, . . . , Zn iid ∼ (1 − p, p) ◮

X

  • m bits

− → f(X)

  • n bits

− → f(X) ⊕ Z

  • Y

− → g(f(X) ⊕ Z)

  • g(Y)

◮ Assuming Pr [g(Y) = X] ≥ 1 − ε, we show nCp ≥ m(1 − ε) − 1 ◮ Compare to nCp > m(1 + εCp) in Thm 1 ◮ Hence, limε→0 m

n = Cp

◮ By Fano, H(X|Y) ≤ h(ε) + ε log(2m − 1) ≤ 1 + εm ◮ I(X; Y) = H(X) − H(X|Y) ≥ m − εm − 1 = m(1 − ε) − 1 ◮ H(Y|X) = H(X, Y) − H(X) = H(X, Z) − H(X) = H(Z) = nh(p)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 13 / 21

slide-95
SLIDE 95

Tightness

◮ X ← {0, 1}m, Z = (Z1, . . . , Zn) where Z1, . . . , Zn iid ∼ (1 − p, p) ◮

X

  • m bits

− → f(X)

  • n bits

− → f(X) ⊕ Z

  • Y

− → g(f(X) ⊕ Z)

  • g(Y)

◮ Assuming Pr [g(Y) = X] ≥ 1 − ε, we show nCp ≥ m(1 − ε) − 1 ◮ Compare to nCp > m(1 + εCp) in Thm 1 ◮ Hence, limε→0 m

n = Cp

◮ By Fano, H(X|Y) ≤ h(ε) + ε log(2m − 1) ≤ 1 + εm ◮ I(X; Y) = H(X) − H(X|Y) ≥ m − εm − 1 = m(1 − ε) − 1 ◮ H(Y|X) = H(X, Y) − H(X) = H(X, Z) − H(X) = H(Z) = nh(p) ◮ I(X; Y) = H(Y) − H(Y|X) = n − nh(p)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 13 / 21

slide-96
SLIDE 96

Tightness

◮ X ← {0, 1}m, Z = (Z1, . . . , Zn) where Z1, . . . , Zn iid ∼ (1 − p, p) ◮

X

  • m bits

− → f(X)

  • n bits

− → f(X) ⊕ Z

  • Y

− → g(f(X) ⊕ Z)

  • g(Y)

◮ Assuming Pr [g(Y) = X] ≥ 1 − ε, we show nCp ≥ m(1 − ε) − 1 ◮ Compare to nCp > m(1 + εCp) in Thm 1 ◮ Hence, limε→0 m

n = Cp

◮ By Fano, H(X|Y) ≤ h(ε) + ε log(2m − 1) ≤ 1 + εm ◮ I(X; Y) = H(X) − H(X|Y) ≥ m − εm − 1 = m(1 − ε) − 1 ◮ H(Y|X) = H(X, Y) − H(X) = H(X, Z) − H(X) = H(Z) = nh(p) ◮ I(X; Y) = H(Y) − H(Y|X) = n − nh(p) ◮ Hence, m(1 − ε) ≤ I(X; Y) + 1 = n(1 − h(p)) + 1 = nCp + 1

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 13 / 21

slide-97
SLIDE 97

Tightness

◮ X ← {0, 1}m, Z = (Z1, . . . , Zn) where Z1, . . . , Zn iid ∼ (1 − p, p) ◮

X

  • m bits

− → f(X)

  • n bits

− → f(X) ⊕ Z

  • Y

− → g(f(X) ⊕ Z)

  • g(Y)

◮ Assuming Pr [g(Y) = X] ≥ 1 − ε, we show nCp ≥ m(1 − ε) − 1 ◮ Compare to nCp > m(1 + εCp) in Thm 1 ◮ Hence, limε→0 m

n = Cp

◮ By Fano, H(X|Y) ≤ h(ε) + ε log(2m − 1) ≤ 1 + εm ◮ I(X; Y) = H(X) − H(X|Y) ≥ m − εm − 1 = m(1 − ε) − 1 ◮ H(Y|X) = H(X, Y) − H(X) = H(X, Z) − H(X) = H(Z) = nh(p) ◮ I(X; Y) = H(Y) − H(Y|X) = n − nh(p) ◮ Hence, m(1 − ε) ≤ I(X; Y) + 1 = n(1 − h(p)) + 1 = nCp + 1 ◮ . . .

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 13 / 21

slide-98
SLIDE 98

General communication channel

slide-99
SLIDE 99

General communication channel

Q : [k] → [k] that channel (a probabilistic function) pi,j = Pr [Q(i) = j] X Y = Q(X)

1 2 . . . k 1 2 . . . k p1,1 p1,4 p2,4 pk,2

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 14 / 21

slide-100
SLIDE 100

General communication channel

Q : [k] → [k] that channel (a probabilistic function) pi,j = Pr [Q(i) = j] X Y = Q(X)

1 2 . . . k 1 2 . . . k p1,1 p1,4 p2,4 pk,2 ◮ x = (x1, . . . , xm) ∈ {0, 1}m

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 14 / 21

slide-101
SLIDE 101

General communication channel

Q : [k] → [k] that channel (a probabilistic function) pi,j = Pr [Q(i) = j] X Y = Q(X)

1 2 . . . k 1 2 . . . k p1,1 p1,4 p2,4 pk,2 ◮ x = (x1, . . . , xm) ∈ {0, 1}m ◮ Encoding function f : {0, 1}m → {1, . . . , k}n

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 14 / 21

slide-102
SLIDE 102

General communication channel

Q : [k] → [k] that channel (a probabilistic function) pi,j = Pr [Q(i) = j] X Y = Q(X)

1 2 . . . k 1 2 . . . k p1,1 p1,4 p2,4 pk,2 ◮ x = (x1, . . . , xm) ∈ {0, 1}m ◮ Encoding function f : {0, 1}m → {1, . . . , k}n ◮ Decoding function g{1, . . . , k}n → {0, 1}m

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 14 / 21

slide-103
SLIDE 103

General communication channel

Q : [k] → [k] that channel (a probabilistic function) pi,j = Pr [Q(i) = j] X Y = Q(X)

1 2 . . . k 1 2 . . . k p1,1 p1,4 p2,4 pk,2 ◮ x = (x1, . . . , xm) ∈ {0, 1}m ◮ Encoding function f : {0, 1}m → {1, . . . , k}n ◮ Decoding function g{1, . . . , k}n → {0, 1}m ◮ x

encoding

− → f(x)

channel

− → Q(f(x))

decoding

− → g(Q(f(x)))

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 14 / 21

slide-104
SLIDE 104

General communication channel

Q : [k] → [k] that channel (a probabilistic function) pi,j = Pr [Q(i) = j] X Y = Q(X)

1 2 . . . k 1 2 . . . k p1,1 p1,4 p2,4 pk,2 ◮ x = (x1, . . . , xm) ∈ {0, 1}m ◮ Encoding function f : {0, 1}m → {1, . . . , k}n ◮ Decoding function g{1, . . . , k}n → {0, 1}m ◮ x

encoding

− → f(x)

channel

− → Q(f(x))

decoding

− → g(Q(f(x)))

◮ We hope for g(Q(f(x))) = x

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 14 / 21

slide-105
SLIDE 105

General communication channel

Q : [k] → [k] that channel (a probabilistic function) pi,j = Pr [Q(i) = j] X Y = Q(X)

1 2 . . . k 1 2 . . . k p1,1 p1,4 p2,4 pk,2 ◮ x = (x1, . . . , xm) ∈ {0, 1}m ◮ Encoding function f : {0, 1}m → {1, . . . , k}n ◮ Decoding function g{1, . . . , k}n → {0, 1}m ◮ x

encoding

− → f(x)

channel

− → Q(f(x))

decoding

− → g(Q(f(x)))

◮ We hope for g(Q(f(x))) = x ◮ Channel capacity CQ = maxX I(X; Y)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 14 / 21

slide-106
SLIDE 106

General communication channel

Q : [k] → [k] that channel (a probabilistic function) pi,j = Pr [Q(i) = j] X Y = Q(X)

1 2 . . . k 1 2 . . . k p1,1 p1,4 p2,4 pk,2 ◮ x = (x1, . . . , xm) ∈ {0, 1}m ◮ Encoding function f : {0, 1}m → {1, . . . , k}n ◮ Decoding function g{1, . . . , k}n → {0, 1}m ◮ x

encoding

− → f(x)

channel

− → Q(f(x))

decoding

− → g(Q(f(x)))

◮ We hope for g(Q(f(x))) = x ◮ Channel capacity CQ = maxX I(X; Y) ◮ The maximal information Y gives on X

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 14 / 21

slide-107
SLIDE 107

General communication channel

Q : [k] → [k] that channel (a probabilistic function) pi,j = Pr [Q(i) = j] X Y = Q(X)

1 2 . . . k 1 2 . . . k p1,1 p1,4 p2,4 pk,2 ◮ x = (x1, . . . , xm) ∈ {0, 1}m ◮ Encoding function f : {0, 1}m → {1, . . . , k}n ◮ Decoding function g{1, . . . , k}n → {0, 1}m ◮ x

encoding

− → f(x)

channel

− → Q(f(x))

decoding

− → g(Q(f(x)))

◮ We hope for g(Q(f(x))) = x ◮ Channel capacity CQ = maxX I(X; Y) ◮ The maximal information Y gives on X ◮ Shannon theorem: ∀Q and ∀ε > 0, ∃mε: ∀m > mε and ∀n > m( 1

CQ + ε) :

∃f, g as above s.t. PrQ [g(Q(f(x))) = x] ≤ ε, for all x ∈ {0, 1}m.

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 14 / 21

slide-108
SLIDE 108

General communication channel

Q : [k] → [k] that channel (a probabilistic function) pi,j = Pr [Q(i) = j] X Y = Q(X)

1 2 . . . k 1 2 . . . k p1,1 p1,4 p2,4 pk,2 ◮ x = (x1, . . . , xm) ∈ {0, 1}m ◮ Encoding function f : {0, 1}m → {1, . . . , k}n ◮ Decoding function g{1, . . . , k}n → {0, 1}m ◮ x

encoding

− → f(x)

channel

− → Q(f(x))

decoding

− → g(Q(f(x)))

◮ We hope for g(Q(f(x))) = x ◮ Channel capacity CQ = maxX I(X; Y) ◮ The maximal information Y gives on X ◮ Shannon theorem: ∀Q and ∀ε > 0, ∃mε: ∀m > mε and ∀n > m( 1

CQ + ε) :

∃f, g as above s.t. PrQ [g(Q(f(x))) = x] ≤ ε, for all x ∈ {0, 1}m.

◮ Proof: similar lines to the binary case, but more subtle distribution for f

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 14 / 21

slide-109
SLIDE 109

Discussion

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 15 / 21

slide-110
SLIDE 110

Discussion

◮ Tight result

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 15 / 21

slide-111
SLIDE 111

Discussion

◮ Tight result ◮ Non-constructive

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 15 / 21

slide-112
SLIDE 112

Discussion

◮ Tight result ◮ Non-constructive ◮ Coding theory: design explicit (and efficient) code achieving the same

bounds

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 15 / 21

slide-113
SLIDE 113

Discussion

◮ Tight result ◮ Non-constructive ◮ Coding theory: design explicit (and efficient) code achieving the same

bounds

◮ Application: faulty communication, storage

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 15 / 21

slide-114
SLIDE 114

Discussion

◮ Tight result ◮ Non-constructive ◮ Coding theory: design explicit (and efficient) code achieving the same

bounds

◮ Application: faulty communication, storage ◮ Combination of data compression and ECC

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 15 / 21

slide-115
SLIDE 115

Part II Combinatorial Applications

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 16 / 21

slide-116
SLIDE 116

Movies

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 17 / 21

slide-117
SLIDE 117

Movies

◮ 2n people, m = 3n movies.

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 17 / 21

slide-118
SLIDE 118

Movies

◮ 2n people, m = 3n movies. ◮ Every pair of movies was seen by at least 90% of the people

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 17 / 21

slide-119
SLIDE 119

Movies

◮ 2n people, m = 3n movies. ◮ Every pair of movies was seen by at least 90% of the people ◮ Claim: there exist two people who saw exactly the same set of movies

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 17 / 21

slide-120
SLIDE 120

Movies

◮ 2n people, m = 3n movies. ◮ Every pair of movies was seen by at least 90% of the people ◮ Claim: there exist two people who saw exactly the same set of movies ◮ X ← [2n] — a random person

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 17 / 21

slide-121
SLIDE 121

Movies

◮ 2n people, m = 3n movies. ◮ Every pair of movies was seen by at least 90% of the people ◮ Claim: there exist two people who saw exactly the same set of movies ◮ X ← [2n] — a random person ◮ gi(x) =

1, x saw movie i; 0,

  • therwise.

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 17 / 21

slide-122
SLIDE 122

Movies

◮ 2n people, m = 3n movies. ◮ Every pair of movies was seen by at least 90% of the people ◮ Claim: there exist two people who saw exactly the same set of movies ◮ X ← [2n] — a random person ◮ gi(x) =

1, x saw movie i; 0,

  • therwise.

◮ Yi = gi(X)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 17 / 21

slide-123
SLIDE 123

Movies

◮ 2n people, m = 3n movies. ◮ Every pair of movies was seen by at least 90% of the people ◮ Claim: there exist two people who saw exactly the same set of movies ◮ X ← [2n] — a random person ◮ gi(x) =

1, x saw movie i; 0,

  • therwise.

◮ Yi = gi(X) ◮ ∀i, j: H(Yi, Yj) ≤ H(0.9, 0.1

3 , 0.1 3 , 0.1 3 ) ≤ 2 3

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 17 / 21

slide-124
SLIDE 124

Movies

◮ 2n people, m = 3n movies. ◮ Every pair of movies was seen by at least 90% of the people ◮ Claim: there exist two people who saw exactly the same set of movies ◮ X ← [2n] — a random person ◮ gi(x) =

1, x saw movie i; 0,

  • therwise.

◮ Yi = gi(X) ◮ ∀i, j: H(Yi, Yj) ≤ H(0.9, 0.1

3 , 0.1 3 , 0.1 3 ) ≤ 2 3

◮ H(Y = (Y1, . . . , Ym)) ≤ m/2

i

H(Yi, Yi+ m

2 )

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 17 / 21

slide-125
SLIDE 125

Movies

◮ 2n people, m = 3n movies. ◮ Every pair of movies was seen by at least 90% of the people ◮ Claim: there exist two people who saw exactly the same set of movies ◮ X ← [2n] — a random person ◮ gi(x) =

1, x saw movie i; 0,

  • therwise.

◮ Yi = gi(X) ◮ ∀i, j: H(Yi, Yj) ≤ H(0.9, 0.1

3 , 0.1 3 , 0.1 3 ) ≤ 2 3

◮ H(Y = (Y1, . . . , Ym)) ≤ m/2

i

H(Yi, Yi+ m

2 )

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 17 / 21

slide-126
SLIDE 126

Movies

◮ 2n people, m = 3n movies. ◮ Every pair of movies was seen by at least 90% of the people ◮ Claim: there exist two people who saw exactly the same set of movies ◮ X ← [2n] — a random person ◮ gi(x) =

1, x saw movie i; 0,

  • therwise.

◮ Yi = gi(X) ◮ ∀i, j: H(Yi, Yj) ≤ H(0.9, 0.1

3 , 0.1 3 , 0.1 3 ) ≤ 2 3

◮ H(Y = (Y1, . . . , Ym)) ≤ m/2

i

H(Yi, Yi+ m

2 ) < 3n

2 · 2 3

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 17 / 21

slide-127
SLIDE 127

Movies

◮ 2n people, m = 3n movies. ◮ Every pair of movies was seen by at least 90% of the people ◮ Claim: there exist two people who saw exactly the same set of movies ◮ X ← [2n] — a random person ◮ gi(x) =

1, x saw movie i; 0,

  • therwise.

◮ Yi = gi(X) ◮ ∀i, j: H(Yi, Yj) ≤ H(0.9, 0.1

3 , 0.1 3 , 0.1 3 ) ≤ 2 3

◮ H(Y = (Y1, . . . , Ym)) ≤ m/2

i

H(Yi, Yi+ m

2 ) < 3n

2 · 2 3 = n

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 17 / 21

slide-128
SLIDE 128

Movies

◮ 2n people, m = 3n movies. ◮ Every pair of movies was seen by at least 90% of the people ◮ Claim: there exist two people who saw exactly the same set of movies ◮ X ← [2n] — a random person ◮ gi(x) =

1, x saw movie i; 0,

  • therwise.

◮ Yi = gi(X) ◮ ∀i, j: H(Yi, Yj) ≤ H(0.9, 0.1

3 , 0.1 3 , 0.1 3 ) ≤ 2 3

◮ H(Y = (Y1, . . . , Ym)) ≤ m/2

i

H(Yi, Yi+ m

2 ) < 3n

2 · 2 3 = n = H(X)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 17 / 21

slide-129
SLIDE 129

Movies

◮ 2n people, m = 3n movies. ◮ Every pair of movies was seen by at least 90% of the people ◮ Claim: there exist two people who saw exactly the same set of movies ◮ X ← [2n] — a random person ◮ gi(x) =

1, x saw movie i; 0,

  • therwise.

◮ Yi = gi(X) ◮ ∀i, j: H(Yi, Yj) ≤ H(0.9, 0.1

3 , 0.1 3 , 0.1 3 ) ≤ 2 3

◮ H(Y = (Y1, . . . , Ym)) ≤ m/2

i

H(Yi, Yi+ m

2 ) < 3n

2 · 2 3 = n = H(X)

◮ Hence, X is not determined by Y

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 17 / 21

slide-130
SLIDE 130

Why H(X1, . . . , Xn) ≤

i H(Xi) so useful?

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 18 / 21

slide-131
SLIDE 131

Why H(X1, . . . , Xn) ≤

i H(Xi) so useful? ◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 18 / 21

slide-132
SLIDE 132

Why H(X1, . . . , Xn) ≤

i H(Xi) so useful? ◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S ◮ log |S| = H(X) ≤

i(Xi) implies |S| ≤ 2

  • i(Xi)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 18 / 21

slide-133
SLIDE 133

Why H(X1, . . . , Xn) ≤

i H(Xi) so useful? ◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S ◮ log |S| = H(X) ≤

i(Xi) implies |S| ≤ 2

  • i(Xi)

◮ If

i H(Xi) is small, then S is small

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 18 / 21

slide-134
SLIDE 134

Why H(X1, . . . , Xn) ≤

i H(Xi) so useful? ◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S ◮ log |S| = H(X) ≤

i(Xi) implies |S| ≤ 2

  • i(Xi)

◮ If

i H(Xi) is small, then S is small

Xi are unbalanced, e.g., ∼ (0.1, 0.9), implies |S| ≤ 2n·h(0.1) ≤ 2n/2

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 18 / 21

slide-135
SLIDE 135

Why H(X1, . . . , Xn) ≤

i H(Xi) so useful? ◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S ◮ log |S| = H(X) ≤

i(Xi) implies |S| ≤ 2

  • i(Xi)

◮ If

i H(Xi) is small, then S is small

Xi are unbalanced, e.g., ∼ (0.1, 0.9), implies |S| ≤ 2n·h(0.1) ≤ 2n/2

◮ S is large implies

i H(Xi) is large, hence most Xi are almost balanced

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 18 / 21

slide-136
SLIDE 136

Why H(X1, . . . , Xn) ≤

i H(Xi) so useful? ◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S ◮ log |S| = H(X) ≤

i(Xi) implies |S| ≤ 2

  • i(Xi)

◮ If

i H(Xi) is small, then S is small

Xi are unbalanced, e.g., ∼ (0.1, 0.9), implies |S| ≤ 2n·h(0.1) ≤ 2n/2

◮ S is large implies

i H(Xi) is large, hence most Xi are almost balanced

◮ |S| ≥ 2n/2 implies Ei←[n] [H(Xi)] ≥ 1 − 1

n

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 18 / 21

slide-137
SLIDE 137

Why H(X1, . . . , Xn) ≤

i H(Xi) so useful? ◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S ◮ log |S| = H(X) ≤

i(Xi) implies |S| ≤ 2

  • i(Xi)

◮ If

i H(Xi) is small, then S is small

Xi are unbalanced, e.g., ∼ (0.1, 0.9), implies |S| ≤ 2n·h(0.1) ≤ 2n/2

◮ S is large implies

i H(Xi) is large, hence most Xi are almost balanced

◮ |S| ≥ 2n/2 implies Ei←[n] [H(Xi)] ≥ 1 − 1

n

◮ Most Xi are close to uniform

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 18 / 21

slide-138
SLIDE 138

Hamming ball

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 19 / 21

slide-139
SLIDE 139

Hamming ball

◮ p ≤ 1

2; S = {(a1, . . . , an) ∈ {0, 1}n : i ai ≤ pn}

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 19 / 21

slide-140
SLIDE 140

Hamming ball

◮ p ≤ 1

2; S = {(a1, . . . , an) ∈ {0, 1}n : i ai ≤ pn}

◮ |S| = ⌊pn⌋

k=0

n

k

  • Iftach Haitner (TAU)

Application of Information Theory, Lecture 5 November 25, 2014 19 / 21

slide-141
SLIDE 141

Hamming ball

◮ p ≤ 1

2; S = {(a1, . . . , an) ∈ {0, 1}n : i ai ≤ pn}

◮ |S| = ⌊pn⌋

k=0

n

k

  • ◮ X = (X1, . . . , Xn) ← S

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 19 / 21

slide-142
SLIDE 142

Hamming ball

◮ p ≤ 1

2; S = {(a1, . . . , an) ∈ {0, 1}n : i ai ≤ pn}

◮ |S| = ⌊pn⌋

k=0

n

k

  • ◮ X = (X1, . . . , Xn) ← S

i Xi ≤ pn =

⇒ E [ Xi] ≤ pn, and by symmetry E [Xi] ≤ p for every i

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 19 / 21

slide-143
SLIDE 143

Hamming ball

◮ p ≤ 1

2; S = {(a1, . . . , an) ∈ {0, 1}n : i ai ≤ pn}

◮ |S| = ⌊pn⌋

k=0

n

k

  • ◮ X = (X1, . . . , Xn) ← S

i Xi ≤ pn =

⇒ E [ Xi] ≤ pn, and by symmetry E [Xi] ≤ p for every i

◮ ∀i, j: Pr[Xi = 1] = Pr[Xj = 1] ≤ p

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 19 / 21

slide-144
SLIDE 144

Hamming ball

◮ p ≤ 1

2; S = {(a1, . . . , an) ∈ {0, 1}n : i ai ≤ pn}

◮ |S| = ⌊pn⌋

k=0

n

k

  • ◮ X = (X1, . . . , Xn) ← S

i Xi ≤ pn =

⇒ E [ Xi] ≤ pn, and by symmetry E [Xi] ≤ p for every i

◮ ∀i, j: Pr[Xi = 1] = Pr[Xj = 1] ≤ p

= ⇒ H(Xi) ≤ h(p) for every i

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 19 / 21

slide-145
SLIDE 145

Hamming ball

◮ p ≤ 1

2; S = {(a1, . . . , an) ∈ {0, 1}n : i ai ≤ pn}

◮ |S| = ⌊pn⌋

k=0

n

k

  • ◮ X = (X1, . . . , Xn) ← S

i Xi ≤ pn =

⇒ E [ Xi] ≤ pn, and by symmetry E [Xi] ≤ p for every i

◮ ∀i, j: Pr[Xi = 1] = Pr[Xj = 1] ≤ p

= ⇒ H(Xi) ≤ h(p) for every i = ⇒ |S| ≤ 2nh(p)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 19 / 21

slide-146
SLIDE 146

Hamming ball

◮ p ≤ 1

2; S = {(a1, . . . , an) ∈ {0, 1}n : i ai ≤ pn}

◮ |S| = ⌊pn⌋

k=0

n

k

  • ◮ X = (X1, . . . , Xn) ← S

i Xi ≤ pn =

⇒ E [ Xi] ≤ pn, and by symmetry E [Xi] ≤ p for every i

◮ ∀i, j: Pr[Xi = 1] = Pr[Xj = 1] ≤ p

= ⇒ H(Xi) ≤ h(p) for every i = ⇒ |S| ≤ 2nh(p) = ⇒ ⌊pn⌋

k=0

n

k

  • ≤ 2nh(p)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 19 / 21

slide-147
SLIDE 147

Hamming ball

◮ p ≤ 1

2; S = {(a1, . . . , an) ∈ {0, 1}n : i ai ≤ pn}

◮ |S| = ⌊pn⌋

k=0

n

k

  • ◮ X = (X1, . . . , Xn) ← S

i Xi ≤ pn =

⇒ E [ Xi] ≤ pn, and by symmetry E [Xi] ≤ p for every i

◮ ∀i, j: Pr[Xi = 1] = Pr[Xj = 1] ≤ p

= ⇒ H(Xi) ≤ h(p) for every i = ⇒ |S| ≤ 2nh(p) = ⇒ ⌊pn⌋

k=0

n

k

  • ≤ 2nh(p)

◮ . . .

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 19 / 21

slide-148
SLIDE 148

Hamming ball

◮ p ≤ 1

2; S = {(a1, . . . , an) ∈ {0, 1}n : i ai ≤ pn}

◮ |S| = ⌊pn⌋

k=0

n

k

  • ◮ X = (X1, . . . , Xn) ← S

i Xi ≤ pn =

⇒ E [ Xi] ≤ pn, and by symmetry E [Xi] ≤ p for every i

◮ ∀i, j: Pr[Xi = 1] = Pr[Xj = 1] ≤ p

= ⇒ H(Xi) ≤ h(p) for every i = ⇒ |S| ≤ 2nh(p) = ⇒ ⌊pn⌋

k=0

n

k

  • ≤ 2nh(p)

◮ . . .

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 19 / 21

slide-149
SLIDE 149

Hamming ball

◮ p ≤ 1

2; S = {(a1, . . . , an) ∈ {0, 1}n : i ai ≤ pn}

◮ |S| = ⌊pn⌋

k=0

n

k

  • ◮ X = (X1, . . . , Xn) ← S

i Xi ≤ pn =

⇒ E [ Xi] ≤ pn, and by symmetry E [Xi] ≤ p for every i

◮ ∀i, j: Pr[Xi = 1] = Pr[Xj = 1] ≤ p

= ⇒ H(Xi) ≤ h(p) for every i = ⇒ |S| ≤ 2nh(p) = ⇒ ⌊pn⌋

k=0

n

k

  • ≤ 2nh(p)

◮ . . .

Application

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 19 / 21

slide-150
SLIDE 150

Hamming ball

◮ p ≤ 1

2; S = {(a1, . . . , an) ∈ {0, 1}n : i ai ≤ pn}

◮ |S| = ⌊pn⌋

k=0

n

k

  • ◮ X = (X1, . . . , Xn) ← S

i Xi ≤ pn =

⇒ E [ Xi] ≤ pn, and by symmetry E [Xi] ≤ p for every i

◮ ∀i, j: Pr[Xi = 1] = Pr[Xj = 1] ≤ p

= ⇒ H(Xi) ≤ h(p) for every i = ⇒ |S| ≤ 2nh(p) = ⇒ ⌊pn⌋

k=0

n

k

  • ≤ 2nh(p)

◮ . . .

Application

◮ X1, . . . , Xn iid uniform bits (i.e., ∼ ( 1

2, 1 2))

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 19 / 21

slide-151
SLIDE 151

Hamming ball

◮ p ≤ 1

2; S = {(a1, . . . , an) ∈ {0, 1}n : i ai ≤ pn}

◮ |S| = ⌊pn⌋

k=0

n

k

  • ◮ X = (X1, . . . , Xn) ← S

i Xi ≤ pn =

⇒ E [ Xi] ≤ pn, and by symmetry E [Xi] ≤ p for every i

◮ ∀i, j: Pr[Xi = 1] = Pr[Xj = 1] ≤ p

= ⇒ H(Xi) ≤ h(p) for every i = ⇒ |S| ≤ 2nh(p) = ⇒ ⌊pn⌋

k=0

n

k

  • ≤ 2nh(p)

◮ . . .

Application

◮ X1, . . . , Xn iid uniform bits (i.e., ∼ ( 1

2, 1 2))

◮ Pr [

i Xi ≤ pn] = Pr [(X1, . . . , Xn) ∈ S] ≤ 2nh(p) · 2−n = 2−n(1−h(p))

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 19 / 21

slide-152
SLIDE 152

Hamming ball

◮ p ≤ 1

2; S = {(a1, . . . , an) ∈ {0, 1}n : i ai ≤ pn}

◮ |S| = ⌊pn⌋

k=0

n

k

  • ◮ X = (X1, . . . , Xn) ← S

i Xi ≤ pn =

⇒ E [ Xi] ≤ pn, and by symmetry E [Xi] ≤ p for every i

◮ ∀i, j: Pr[Xi = 1] = Pr[Xj = 1] ≤ p

= ⇒ H(Xi) ≤ h(p) for every i = ⇒ |S| ≤ 2nh(p) = ⇒ ⌊pn⌋

k=0

n

k

  • ≤ 2nh(p)

◮ . . .

Application

◮ X1, . . . , Xn iid uniform bits (i.e., ∼ ( 1

2, 1 2))

◮ Pr [

i Xi ≤ pn] = Pr [(X1, . . . , Xn) ∈ S] ≤ 2nh(p) · 2−n = 2−n(1−h(p))

◮ Very useful inequality. No Chernoff, just IT

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 19 / 21

slide-153
SLIDE 153

Isoperimetric inequality

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-154
SLIDE 154

Isoperimetric inequality

◮ S ⊆ {0, 1}n

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-155
SLIDE 155

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-156
SLIDE 156

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-157
SLIDE 157

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Theorem 4 |E| ≤ 1

2 · |S| · log |S|

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-158
SLIDE 158

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Theorem 4 |E| ≤ 1

2 · |S| · log |S|

◮ Equality if S is “face" : S = {(x, y): y ∈ {0, 1}d} for some x ∈ {0, 1}n−d

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-159
SLIDE 159

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Theorem 4 |E| ≤ 1

2 · |S| · log |S|

◮ Equality if S is “face" : S = {(x, y): y ∈ {0, 1}d} for some x ∈ {0, 1}n−d ◮ Example: S is a face of the 3-dimensional cube

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-160
SLIDE 160

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Theorem 4 |E| ≤ 1

2 · |S| · log |S|

◮ Equality if S is “face" : S = {(x, y): y ∈ {0, 1}d} for some x ∈ {0, 1}n−d ◮ Example: S is a face of the 3-dimensional cube

n = 3, |S| = 4, implies |E| ≤ 1

2 · 4 · log 4 = 4

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-161
SLIDE 161

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Theorem 4 |E| ≤ 1

2 · |S| · log |S|

◮ Equality if S is “face" : S = {(x, y): y ∈ {0, 1}d} for some x ∈ {0, 1}n−d ◮ Example: S is a face of the 3-dimensional cube

n = 3, |S| = 4, implies |E| ≤ 1

2 · 4 · log 4 = 4

◮ Ei — edges of E in direction i

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-162
SLIDE 162

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Theorem 4 |E| ≤ 1

2 · |S| · log |S|

◮ Equality if S is “face" : S = {(x, y): y ∈ {0, 1}d} for some x ∈ {0, 1}n−d ◮ Example: S is a face of the 3-dimensional cube

n = 3, |S| = 4, implies |E| ≤ 1

2 · 4 · log 4 = 4

◮ Ei — edges of E in direction i

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-163
SLIDE 163

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Theorem 4 |E| ≤ 1

2 · |S| · log |S|

◮ Equality if S is “face" : S = {(x, y): y ∈ {0, 1}d} for some x ∈ {0, 1}n−d ◮ Example: S is a face of the 3-dimensional cube

n = 3, |S| = 4, implies |E| ≤ 1

2 · 4 · log 4 = 4

◮ Ei — edges of E in direction i

(E =

i∈[n] Ei)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-164
SLIDE 164

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Theorem 4 |E| ≤ 1

2 · |S| · log |S|

◮ Equality if S is “face" : S = {(x, y): y ∈ {0, 1}d} for some x ∈ {0, 1}n−d ◮ Example: S is a face of the 3-dimensional cube

n = 3, |S| = 4, implies |E| ≤ 1

2 · 4 · log 4 = 4

◮ Ei — edges of E in direction i

(E =

i∈[n] Ei)

◮ X = (X1, . . . , Xn) ← S and X−i = (X1, X2, . . . , Xi−1, Xi+1, . . . , Xn)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-165
SLIDE 165

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Theorem 4 |E| ≤ 1

2 · |S| · log |S|

◮ Equality if S is “face" : S = {(x, y): y ∈ {0, 1}d} for some x ∈ {0, 1}n−d ◮ Example: S is a face of the 3-dimensional cube

n = 3, |S| = 4, implies |E| ≤ 1

2 · 4 · log 4 = 4

◮ Ei — edges of E in direction i

(E =

i∈[n] Ei)

◮ X = (X1, . . . , Xn) ← S and X−i = (X1, X2, . . . , Xi−1, Xi+1, . . . , Xn)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-166
SLIDE 166

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Theorem 4 |E| ≤ 1

2 · |S| · log |S|

◮ Equality if S is “face" : S = {(x, y): y ∈ {0, 1}d} for some x ∈ {0, 1}n−d ◮ Example: S is a face of the 3-dimensional cube

n = 3, |S| = 4, implies |E| ≤ 1

2 · 4 · log 4 = 4

◮ Ei — edges of E in direction i

(E =

i∈[n] Ei)

◮ X = (X1, . . . , Xn) ← S and X−i = (X1, X2, . . . , Xi−1, Xi+1, . . . , Xn)

Lemma 5 H(Xi|X−i) = 2|Ei|

|S|

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-167
SLIDE 167

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Theorem 4 |E| ≤ 1

2 · |S| · log |S|

◮ Equality if S is “face" : S = {(x, y): y ∈ {0, 1}d} for some x ∈ {0, 1}n−d ◮ Example: S is a face of the 3-dimensional cube

n = 3, |S| = 4, implies |E| ≤ 1

2 · 4 · log 4 = 4

◮ Ei — edges of E in direction i

(E =

i∈[n] Ei)

◮ X = (X1, . . . , Xn) ← S and X−i = (X1, X2, . . . , Xi−1, Xi+1, . . . , Xn)

Lemma 5 H(Xi|X−i) = 2|Ei|

|S|

Proving Thm 4:

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-168
SLIDE 168

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Theorem 4 |E| ≤ 1

2 · |S| · log |S|

◮ Equality if S is “face" : S = {(x, y): y ∈ {0, 1}d} for some x ∈ {0, 1}n−d ◮ Example: S is a face of the 3-dimensional cube

n = 3, |S| = 4, implies |E| ≤ 1

2 · 4 · log 4 = 4

◮ Ei — edges of E in direction i

(E =

i∈[n] Ei)

◮ X = (X1, . . . , Xn) ← S and X−i = (X1, X2, . . . , Xi−1, Xi+1, . . . , Xn)

Lemma 5 H(Xi|X−i) = 2|Ei|

|S|

Proving Thm 4: log |S| = H(X1, . . . , Xn) = H(X1) + H(X2|X1) + . . . + H(Xn|X1, X2, . . . , Xn−1)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-169
SLIDE 169

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Theorem 4 |E| ≤ 1

2 · |S| · log |S|

◮ Equality if S is “face" : S = {(x, y): y ∈ {0, 1}d} for some x ∈ {0, 1}n−d ◮ Example: S is a face of the 3-dimensional cube

n = 3, |S| = 4, implies |E| ≤ 1

2 · 4 · log 4 = 4

◮ Ei — edges of E in direction i

(E =

i∈[n] Ei)

◮ X = (X1, . . . , Xn) ← S and X−i = (X1, X2, . . . , Xi−1, Xi+1, . . . , Xn)

Lemma 5 H(Xi|X−i) = 2|Ei|

|S|

Proving Thm 4: log |S| = H(X1, . . . , Xn) = H(X1) + H(X2|X1) + . . . + H(Xn|X1, X2, . . . , Xn−1) ≥ H(X1|X−1) + H(X2|X−2) + . . . H(Xn|X−n)

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-170
SLIDE 170

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Theorem 4 |E| ≤ 1

2 · |S| · log |S|

◮ Equality if S is “face" : S = {(x, y): y ∈ {0, 1}d} for some x ∈ {0, 1}n−d ◮ Example: S is a face of the 3-dimensional cube

n = 3, |S| = 4, implies |E| ≤ 1

2 · 4 · log 4 = 4

◮ Ei — edges of E in direction i

(E =

i∈[n] Ei)

◮ X = (X1, . . . , Xn) ← S and X−i = (X1, X2, . . . , Xi−1, Xi+1, . . . , Xn)

Lemma 5 H(Xi|X−i) = 2|Ei|

|S|

Proving Thm 4: log |S| = H(X1, . . . , Xn) = H(X1) + H(X2|X1) + . . . + H(Xn|X1, X2, . . . , Xn−1) ≥ H(X1|X−1) + H(X2|X−2) + . . . H(Xn|X−n) =

  • i

2 |Ei| |S|

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-171
SLIDE 171

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Theorem 4 |E| ≤ 1

2 · |S| · log |S|

◮ Equality if S is “face" : S = {(x, y): y ∈ {0, 1}d} for some x ∈ {0, 1}n−d ◮ Example: S is a face of the 3-dimensional cube

n = 3, |S| = 4, implies |E| ≤ 1

2 · 4 · log 4 = 4

◮ Ei — edges of E in direction i

(E =

i∈[n] Ei)

◮ X = (X1, . . . , Xn) ← S and X−i = (X1, X2, . . . , Xi−1, Xi+1, . . . , Xn)

Lemma 5 H(Xi|X−i) = 2|Ei|

|S|

Proving Thm 4: log |S| = H(X1, . . . , Xn) = H(X1) + H(X2|X1) + . . . + H(Xn|X1, X2, . . . , Xn−1) ≥ H(X1|X−1) + H(X2|X−2) + . . . H(Xn|X−n) =

  • i

2 |Ei| |S| = 2 |E| |S|

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-172
SLIDE 172

Isoperimetric inequality

◮ S ⊆ {0, 1}n ◮ Edges of S — E = {(u, v) ∈ S : |u − v| = 1}

Theorem 4 |E| ≤ 1

2 · |S| · log |S|

◮ Equality if S is “face" : S = {(x, y): y ∈ {0, 1}d} for some x ∈ {0, 1}n−d ◮ Example: S is a face of the 3-dimensional cube

n = 3, |S| = 4, implies |E| ≤ 1

2 · 4 · log 4 = 4

◮ Ei — edges of E in direction i

(E =

i∈[n] Ei)

◮ X = (X1, . . . , Xn) ← S and X−i = (X1, X2, . . . , Xi−1, Xi+1, . . . , Xn)

Lemma 5 H(Xi|X−i) = 2|Ei|

|S|

Proving Thm 4: log |S| = H(X1, . . . , Xn) = H(X1) + H(X2|X1) + . . . + H(Xn|X1, X2, . . . , Xn−1) ≥ H(X1|X−1) + H(X2|X−2) + . . . H(Xn|X−n) =

  • i

2 |Ei| |S| = 2 |E| |S| .

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 20 / 21

slide-173
SLIDE 173

Proving Lemma 5

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 21 / 21

slide-174
SLIDE 174

Proving Lemma 5

We prove for i = 1

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 21 / 21

slide-175
SLIDE 175

Proving Lemma 5

We prove for i = 1

◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 21 / 21

slide-176
SLIDE 176

Proving Lemma 5

We prove for i = 1

◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S ◮ E = {(u, v) ∈ S : |u − v| = 1} and E1 contains edges of E in direction 1

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 21 / 21

slide-177
SLIDE 177

Proving Lemma 5

We prove for i = 1

◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S ◮ E = {(u, v) ∈ S : |u − v| = 1} and E1 contains edges of E in direction 1 ◮ S−1 = {y ∈ {0, 1}n−1 : ∃x ∈ {0, 1} s.t. (x, y) ∈ S}.

(S projected on (2, . . . , n))

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 21 / 21

slide-178
SLIDE 178

Proving Lemma 5

We prove for i = 1

◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S ◮ E = {(u, v) ∈ S : |u − v| = 1} and E1 contains edges of E in direction 1 ◮ S−1 = {y ∈ {0, 1}n−1 : ∃x ∈ {0, 1} s.t. (x, y) ∈ S}.

(S projected on (2, . . . , n))

◮ Se

−1 = {y ∈ {0, 1}n−1 : (0, y), (1, y) ∈ S} and S¬e −1 = S−1 \ Se −1

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 21 / 21

slide-179
SLIDE 179

Proving Lemma 5

We prove for i = 1

◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S ◮ E = {(u, v) ∈ S : |u − v| = 1} and E1 contains edges of E in direction 1 ◮ S−1 = {y ∈ {0, 1}n−1 : ∃x ∈ {0, 1} s.t. (x, y) ∈ S}.

(S projected on (2, . . . , n))

◮ Se

−1 = {y ∈ {0, 1}n−1 : (0, y), (1, y) ∈ S} and S¬e −1 = S−1 \ Se −1

◮ |S| = 2

  • Se

−1

  • +
  • S¬e

−1

  • Iftach Haitner (TAU)

Application of Information Theory, Lecture 5 November 25, 2014 21 / 21

slide-180
SLIDE 180

Proving Lemma 5

We prove for i = 1

◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S ◮ E = {(u, v) ∈ S : |u − v| = 1} and E1 contains edges of E in direction 1 ◮ S−1 = {y ∈ {0, 1}n−1 : ∃x ∈ {0, 1} s.t. (x, y) ∈ S}.

(S projected on (2, . . . , n))

◮ Se

−1 = {y ∈ {0, 1}n−1 : (0, y), (1, y) ∈ S} and S¬e −1 = S−1 \ Se −1

◮ |S| = 2

  • Se

−1

  • +
  • S¬e

−1

  • ◮ |E1| =
  • Se

−1

  • Iftach Haitner (TAU)

Application of Information Theory, Lecture 5 November 25, 2014 21 / 21

slide-181
SLIDE 181

Proving Lemma 5

We prove for i = 1

◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S ◮ E = {(u, v) ∈ S : |u − v| = 1} and E1 contains edges of E in direction 1 ◮ S−1 = {y ∈ {0, 1}n−1 : ∃x ∈ {0, 1} s.t. (x, y) ∈ S}.

(S projected on (2, . . . , n))

◮ Se

−1 = {y ∈ {0, 1}n−1 : (0, y), (1, y) ∈ S} and S¬e −1 = S−1 \ Se −1

◮ |S| = 2

  • Se

−1

  • +
  • S¬e

−1

  • ◮ |E1| =
  • Se

−1

  • ◮ H(X|X−1) = Pr
  • X−1 ∈ Se

−1

  • · 1

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 21 / 21

slide-182
SLIDE 182

Proving Lemma 5

We prove for i = 1

◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S ◮ E = {(u, v) ∈ S : |u − v| = 1} and E1 contains edges of E in direction 1 ◮ S−1 = {y ∈ {0, 1}n−1 : ∃x ∈ {0, 1} s.t. (x, y) ∈ S}.

(S projected on (2, . . . , n))

◮ Se

−1 = {y ∈ {0, 1}n−1 : (0, y), (1, y) ∈ S} and S¬e −1 = S−1 \ Se −1

◮ |S| = 2

  • Se

−1

  • +
  • S¬e

−1

  • ◮ |E1| =
  • Se

−1

  • ◮ H(X|X−1) = Pr
  • X−1 ∈ Se

−1

  • · 1

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 21 / 21

slide-183
SLIDE 183

Proving Lemma 5

We prove for i = 1

◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S ◮ E = {(u, v) ∈ S : |u − v| = 1} and E1 contains edges of E in direction 1 ◮ S−1 = {y ∈ {0, 1}n−1 : ∃x ∈ {0, 1} s.t. (x, y) ∈ S}.

(S projected on (2, . . . , n))

◮ Se

−1 = {y ∈ {0, 1}n−1 : (0, y), (1, y) ∈ S} and S¬e −1 = S−1 \ Se −1

◮ |S| = 2

  • Se

−1

  • +
  • S¬e

−1

  • ◮ |E1| =
  • Se

−1

  • ◮ H(X|X−1) = Pr
  • X−1 ∈ Se

−1

  • · 1 =

2|Se

−1|

2|Se

−1|+|S¬e −1|

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 21 / 21

slide-184
SLIDE 184

Proving Lemma 5

We prove for i = 1

◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S ◮ E = {(u, v) ∈ S : |u − v| = 1} and E1 contains edges of E in direction 1 ◮ S−1 = {y ∈ {0, 1}n−1 : ∃x ∈ {0, 1} s.t. (x, y) ∈ S}.

(S projected on (2, . . . , n))

◮ Se

−1 = {y ∈ {0, 1}n−1 : (0, y), (1, y) ∈ S} and S¬e −1 = S−1 \ Se −1

◮ |S| = 2

  • Se

−1

  • +
  • S¬e

−1

  • ◮ |E1| =
  • Se

−1

  • ◮ H(X|X−1) = Pr
  • X−1 ∈ Se

−1

  • · 1 =

2|Se

−1|

2|Se

−1|+|S¬e −1| = 2|E1|

|S|

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 21 / 21

slide-185
SLIDE 185

Proving Lemma 5

We prove for i = 1

◮ S ⊆ {0, 1}n; X = (X1, . . . , Xn) ← S ◮ E = {(u, v) ∈ S : |u − v| = 1} and E1 contains edges of E in direction 1 ◮ S−1 = {y ∈ {0, 1}n−1 : ∃x ∈ {0, 1} s.t. (x, y) ∈ S}.

(S projected on (2, . . . , n))

◮ Se

−1 = {y ∈ {0, 1}n−1 : (0, y), (1, y) ∈ S} and S¬e −1 = S−1 \ Se −1

◮ |S| = 2

  • Se

−1

  • +
  • S¬e

−1

  • ◮ |E1| =
  • Se

−1

  • ◮ H(X|X−1) = Pr
  • X−1 ∈ Se

−1

  • · 1 =

2|Se

−1|

2|Se

−1|+|S¬e −1| = 2|E1|

|S|

◮ . . .

Iftach Haitner (TAU) Application of Information Theory, Lecture 5 November 25, 2014 21 / 21