Communication Complexity David P. Woodruff IBM Almaden Talk - - PowerPoint PPT Presentation

communication
SMART_READER_LITE
LIVE PREVIEW

Communication Complexity David P. Woodruff IBM Almaden Talk - - PowerPoint PPT Presentation

Information Theory for Communication Complexity David P. Woodruff IBM Almaden Talk Outline 1. Information Theory Concepts 2. Distances Between Distributions 3. An Example Communication Lower Bound Randomized 1-way Communication


slide-1
SLIDE 1

Information Theory for Communication Complexity

David P. Woodruff IBM Almaden

slide-2
SLIDE 2

Talk Outline

1. Information Theory Concepts 2. Distances Between Distributions 3. An Example Communication Lower Bound – Randomized 1-way Communication Complexity of the INDEX problem 4. Communication Lower Bounds imply space lower bounds for data stream algorithms 5. Techniques for Multi-Player Communication

slide-3
SLIDE 3

Discrete Distributions

slide-4
SLIDE 4

Entropy

(symmetric)

slide-5
SLIDE 5

Conditional and Joint Entropy

slide-6
SLIDE 6

Chain Rule for Entropy

slide-7
SLIDE 7

Conditioning Cannot Increase Entropy

continuous

slide-8
SLIDE 8

Conditioning Cannot Increase Entropy

slide-9
SLIDE 9

Mutual Information

  • (Mutual Information) I(X ; Y) = H(X) – H(X | Y)

= H(Y) – H(Y | X) = I(Y ; X) Note: I(X ; X) = H(X) – H(X | X) = H(X)

  • (Conditional Mutual Information)

I(X ; Y | Z) = H(X | Z) – H(X | Y, Z)

slide-10
SLIDE 10

Chain Rule for Mutual Information

slide-11
SLIDE 11

Fano’s Inequality

Here X -> Y -> X’ is a Markov Chain, meaning X’ and X are independent given Y. “Past and future are conditionally independent given the present” To prove Fano’s Inequality, we need the data processing inequality

slide-12
SLIDE 12

Data Processing Inequality

  • Suppose X -> Y -> Z is a Markov Chain. Then,

𝐽 𝑌 ; 𝑍 ≥ 𝐽(𝑌; 𝑎)

  • That is, no clever combination of the data can improve estimation
  • I(X ; Y, Z) = I(X ; Z) + I(X ; Y | Z) = I(X ; Y) + I(X ; Z | Y)
  • So, it suffices to show I(X ; Z | Y) = 0
  • I(X ; Z | Y) = H(X | Y) – H(X | Y, Z)
  • But given Y, then X and Z are independent, so H(X | Y, Z) = H(X | Y).
  • Data Processing Inequality implies H(X | Y) ≤ 𝐼 𝑌 𝑎)
slide-13
SLIDE 13

Proof of Fano’s Inequality

  • For any estimator X’ such that X-> Y -> X’ with 𝑄

𝑓 = Pr 𝑌 ≠ 𝑌′ ,

we have 𝐼 𝑌 𝑍) ≤ 𝐼 𝑄

𝑓 + 𝑄 𝑓(log2 𝑌 − 1) .

Proof: Let E = 1 if X’ is not equal to X, and E = 0 otherwise. H(E, X | X’) = H(X | X’) + H(E | X, X’) = H(X | X’) H(E, X | X’) = H(E | X’) + H(X | E, X’) ≤ 𝐼 𝑄

𝑓 + H(X | E, X’)

But H(X | E, X’) = Pr(E = 0)H(X | X’, E = 0) + Pr(E = 1)H(X | X’, E = 1) ≤ (1 − 𝑄

𝑓) ⋅ 0 + 𝑄 𝑓 ⋅ log2 𝑌 − 1

Combining the above, H(X | X’) ≤ 𝐼 𝑄

𝑓 + 𝑄 𝑓 ⋅ log2 𝑌 − 1

By Data Processing, H(X | Y) ≤ 𝐼 𝑌 𝑌′) ≤ 𝐼 𝑄

𝑓 + 𝑄 𝑓 ⋅ log2 𝑌 − 1

slide-14
SLIDE 14

Tightness of Fano’s Inequality

slide-15
SLIDE 15

Talk Outline

1. Information Theory Concepts 2. Distances Between Distributions 3. An Example Communication Lower Bound – Randomized 1-way Communication Complexity of the INDEX problem 4. Communication Lower Bounds imply space lower bounds for data stream algorithms 5. Techniques for Multi-Player Communication

slide-16
SLIDE 16

Distances Between Distributions

slide-17
SLIDE 17

Why Hellinger Distance?

slide-18
SLIDE 18

Product Property of Hellinger Distance

slide-19
SLIDE 19

Jensen-Shannon Distance

l

slide-20
SLIDE 20

Relations Between Distance Measures

slide-21
SLIDE 21

Talk Outline

1. Information Theory Concepts 2. Distances Between Distributions 3. An Example Communication Lower Bound – Randomized 1-way Communication Complexity of the INDEX problem 4. Communication Lower Bounds imply space lower bounds for data stream algorithms 5. Techniques for Multi-Player Communication

slide-22
SLIDE 22

Randomized 1-Way Communication Complexity

x 2 {0,1}n j 2 {1, 2, 3, …, n}

INDEX PROBLEM

slide-23
SLIDE 23

1-Way Communication Complexity of Index

  • Consider a uniform distribution μ on X
  • Alice sends a single message M to Bob
  • We can think of Bob’s output as a guess 𝑌

𝑘 ′𝑢𝑝 𝑌 𝑘

  • For all j, Pr 𝑌

𝑘 ′ = 𝑌 𝑘 ≥ 2 3

  • By Fano’s inequality, for all j,

𝐼 𝑌

𝑘 𝑁) ≤ 𝐼 2 3 + 1 3 (log2 2 − 1) = 𝐼( 1 3)

slide-24
SLIDE 24

1-Way Communication of Index Continued

So, 𝐽 𝑌 ; 𝑁 ≥ 𝑜 − 𝑗 𝐼 𝑌𝑗 𝑁) ≥ 𝑜 − 𝐼

1 3 𝑜

So, 𝑁 ≥ 𝐼 𝑁 ≥ 𝐽 𝑌 ; 𝑁 = Ω 𝑜

slide-25
SLIDE 25

Talk Outline

1. Information Theory Concepts 2. Distances Between Distributions 3. An Example Communication Lower Bound – Randomized 1-way Communication Complexity of the INDEX problem 4. Communication Lower Bounds imply space lower bounds for data stream algorithms 5. Techniques for Multi-Player Communication