Sub-optimality of superposition coding for three or more receivers - - PowerPoint PPT Presentation

sub optimality of superposition coding for three or more
SMART_READER_LITE
LIVE PREVIEW

Sub-optimality of superposition coding for three or more receivers - - PowerPoint PPT Presentation

Sub-optimality of superposition coding for three or more receivers Chandra Nair, & Mehdi Yazdanpanah The Chinese University of Hong Kong CISS 2018 23 Mar, 2018 The story of superposition coding is the story of broadcast channels with


slide-1
SLIDE 1

Sub-optimality of superposition coding for three or more receivers

Chandra Nair, & Mehdi Yazdanpanah

The Chinese University of Hong Kong CISS 2018 23 Mar, 2018

slide-2
SLIDE 2

The story of superposition coding is the story of broadcast channels with degradation (receivers or message sets)

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 2 / 17

slide-3
SLIDE 3

The story of superposition coding is the story of broadcast channels with degradation (receivers or message sets) It is also the story of auxiliary random variables

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 2 / 17

slide-4
SLIDE 4

In the beginning

◮ Cover (1972) proposed the superposition coding achievable region for degraded

broadcast channels

He used an auxiliary variable to represent the message for the weaker of two receivers.

◮ Bergmans (1973, Gaussian) and Gallager (1974, discrete-memoryless) established

the optimality of superposition coding for the degraded broadcast channel

Gallager’s proof of optimality of superposition coding naturally extended to a sequence of degraded receivers

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 3 / 17

slide-5
SLIDE 5

In the beginning

◮ Cover (1972) proposed the superposition coding achievable region for degraded

broadcast channels

He used an auxiliary variable to represent the message for the weaker of two receivers.

◮ Bergmans (1973, Gaussian) and Gallager (1974, discrete-memoryless) established

the optimality of superposition coding for the degraded broadcast channel

Gallager’s proof of optimality of superposition coding naturally extended to a sequence of degraded receivers

The optimality of superposition coding region was then established for

◮ Weaker notions of weaker receiver in a two-receiver broadcast channel

Less Noisy (Korner-Marton 75), More Capable (El Gamal 79)

◮ Degraded message sets (Korner-Marton 77)

Comparison of the sizes of the images of a set via two channels

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 3 / 17

slide-6
SLIDE 6

In the beginning

◮ Cover (1972) proposed the superposition coding achievable region for degraded

broadcast channels

He used an auxiliary variable to represent the message for the weaker of two receivers.

◮ Bergmans (1973, Gaussian) and Gallager (1974, discrete-memoryless) established

the optimality of superposition coding for the degraded broadcast channel

Gallager’s proof of optimality of superposition coding naturally extended to a sequence of degraded receivers

The optimality of superposition coding region was then established for

◮ Weaker notions of weaker receiver in a two-receiver broadcast channel

Less Noisy (Korner-Marton 75), More Capable (El Gamal 79)

◮ Degraded message sets (Korner-Marton 77)

Comparison of the sizes of the images of a set via two channels

A (seemingly) natural "Implication": Auxiliaries (in superposition coding) captured the coarser message.

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 3 / 17

slide-7
SLIDE 7

More recent times ..

The optimality of superposition coding was established for

◮ Three receiver less-noisy broadcast channel (N-Wang 2010)

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 4 / 17

slide-8
SLIDE 8

More recent times ..

The optimality of superposition coding was established for

◮ Three receiver less-noisy broadcast channel (N-Wang 2010)

However for the following three-receiver broadcast channel setting:

◮ Receivers Y2, Y3 wish to decode message M0 ◮ Receiver Y1 wishes to decode messages M0, M1

superposition coding region was not optimal (El Gamal-N 2009)

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 4 / 17

slide-9
SLIDE 9

More recent times ..

The optimality of superposition coding was established for

◮ Three receiver less-noisy broadcast channel (N-Wang 2010)

However for the following three-receiver broadcast channel setting:

◮ Receivers Y2, Y3 wish to decode message M0 ◮ Receiver Y1 wishes to decode messages M0, M1

superposition coding region was not optimal (El Gamal-N 2009) Demonstrates that auxiliaries do not capture the coarser message.

◮ Associating an auxiliary with information decoded by groups of receivers

improved the achievable region

U123, U12, U13, and U1 = X. The achievable region was no longer a superposition coding region, it also involved the other (old) idea: random binning

A (seemingly) natural "Implication" (Take 2): Auxiliaries (in superposition coding) captured the information decoded by groups of receivers.

◮ Binning and Superposition both present

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 4 / 17

slide-10
SLIDE 10

Consistency

The revised intuition about auxiliaries is consistent with Marton’s achievable scheme for two-receiver broadcast channels with private messages

◮ The region employs three auxiliaries: U1, U2, U12 ◮ It is known that this region is strictly better than the one with only U1, U2 (even

for private messages).

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 5 / 17

slide-11
SLIDE 11

Consistency

The revised intuition about auxiliaries is consistent with Marton’s achievable scheme for two-receiver broadcast channels with private messages

◮ The region employs three auxiliaries: U1, U2, U12 ◮ It is known that this region is strictly better than the one with only U1, U2 (even

for private messages). Open Question Is Marton’s achievable scheme for two-receiver broadcast channels optimal? Three-or-more receivers A three-receiver broadcast channel with private messages would have seven auxiliaries U123, U12, U13, U23, U1, U2, U3 and a natural extension of Marton’s scheme would have two layers of superposition coding and binning between random variables in each layer. A succinct clean representation of the rate constraints is not available

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 5 / 17

slide-12
SLIDE 12

Consistency

The revised intuition about auxiliaries is consistent with Marton’s achievable scheme for two-receiver broadcast channels with private messages

◮ The region employs three auxiliaries: U1, U2, U12 ◮ It is known that this region is strictly better than the one with only U1, U2 (even

for private messages). Open Question Is Marton’s achievable scheme for two-receiver broadcast channels optimal? Three-or-more receivers A three-receiver broadcast channel with private messages would have seven auxiliaries U123, U12, U13, U23, U1, U2, U3 and a natural extension of Marton’s scheme would have two layers of superposition coding and binning between random variables in each layer. A succinct clean representation of the rate constraints is not available However: this region is not optimal (Padlakanda and Pradhan (2015))

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 5 / 17

slide-13
SLIDE 13

Still one fundamental setting remained

Consider the setting:

◮ Receivers Y3 wish to decode message M0 ◮ Receiver Y1, Y2 wishes to decode messages M0, M1

Interpretation of auxiliaries: U123 and U12 = X, and only superposition coding Question: Is superposition coding optimal? (Note: The first layer superposition coding of the previous private message setting)

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 6 / 17

slide-14
SLIDE 14

Still one fundamental setting remained

Consider the setting:

◮ Receivers Y3 wish to decode message M0 ◮ Receiver Y1, Y2 wishes to decode messages M0, M1

Interpretation of auxiliaries: U123 and U12 = X, and only superposition coding Question: Is superposition coding optimal? (Note: The first layer superposition coding of the previous private message setting)

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 6 / 17

slide-15
SLIDE 15

Still one fundamental setting remained

Consider the setting:

◮ Receivers Y3 wish to decode message M0 ◮ Receiver Y1, Y2 wishes to decode messages M0, M1

Interpretation of auxiliaries: U123 and U12 = X, and only superposition coding Question: Is superposition coding optimal? (Note: The first layer superposition coding of the previous private message setting) Csiszar’s open problem is very closely tied to finding the capacity region

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 6 / 17

slide-16
SLIDE 16

A remark

Körner had proposed a region (1984) for the image size characterization over three channels

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 7 / 17

slide-17
SLIDE 17

Suboptimality of superposition coding

Superposition coding is sub-optimal for the setting (N-Yazdanpanah 2017)

◮ Constructed a channel whose 2-letter superposition-coding region was larger than

the 1-letter one

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 8 / 17

slide-18
SLIDE 18

Suboptimality of superposition coding

Superposition coding is sub-optimal for the setting (N-Yazdanpanah 2017)

◮ Constructed a channel whose 2-letter superposition-coding region was larger than

the 1-letter one Note: The same (counter)-example showed that Korner’s region is a proper subset of H(X; Y ; Z|S).

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 8 / 17

slide-19
SLIDE 19

Suboptimality of superposition coding

Superposition coding is sub-optimal for the setting (N-Yazdanpanah 2017)

◮ Constructed a channel whose 2-letter superposition-coding region was larger than

the 1-letter one Note: The same (counter)-example showed that Korner’s region is a proper subset of H(X; Y ; Z|S). Remarks

◮ It took us three years to get counterexamples

The optimization problems involved are non-convex In small dimensions counter-examples lie in a set of very small "size" (random sampling does not work)

◮ Question: Why did we believe that superposition coding was sub-optimal? ◮ More generally, why do we believe certain regions are optimal while certain

  • thers are not?

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 8 / 17

slide-20
SLIDE 20

Suboptimality of superposition coding

Superposition coding is sub-optimal for the setting (N-Yazdanpanah 2017)

◮ Constructed a channel whose 2-letter superposition-coding region was larger than

the 1-letter one Note: The same (counter)-example showed that Korner’s region is a proper subset of H(X; Y ; Z|S). Remarks

◮ It took us three years to get counterexamples

The optimization problems involved are non-convex In small dimensions counter-examples lie in a set of very small "size" (random sampling does not work)

◮ Question: Why did we believe that superposition coding was sub-optimal? ◮ More generally, why do we believe certain regions are optimal while certain

  • thers are not?

An unpublished conjecture: Local-tensorization implies global tensorization

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 8 / 17

slide-21
SLIDE 21

Suboptimality of superposition coding

Superposition coding is sub-optimal for the setting (N-Yazdanpanah 2017)

◮ Constructed a channel whose 2-letter superposition-coding region was larger than

the 1-letter one Note: The same (counter)-example showed that Korner’s region is a proper subset of H(X; Y ; Z|S). This talk: Focus on the (counter)-example

◮ Bounds on the capacity region ◮ Shed light to properties of good codes for this channel

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 8 / 17

slide-22
SLIDE 22

Multilevel product broadcast erasure channel (ISIT ’17)

Xa → Ya : BEC(ea), Xb → Yb : BEC(eb) Xa → ˆ Ya : BEC(ˆ ea), Xb → ˆ Yb : BEC(ˆ eb) Xa → Za : BEC(fa), Xb → Zb : BEC(fb) ˆ ea ≥ fa ≥ ea & eb ≥ fb ≥ ˆ eb CZ = (1 − fa) + (1 − fb)

Xa Xb Ya ˆ Yb Za Zb ˆ Ya Yb ωa1 ωb1 ωa2 ωb2 ωa3 ωb3 ωa1 ωb1 ωa2 ωb2 ωa3 ωb3

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 9 / 17

slide-23
SLIDE 23

Multilevel product broadcast erasure channel (ISIT ’17)

Xa → Ya : BEC(ea), Xb → Yb : BEC(eb) Xa → ˆ Ya : BEC(ˆ ea), Xb → ˆ Yb : BEC(ˆ eb) Xa → Za : BEC(fa), Xb → Zb : BEC(fb) ˆ ea ≥ fa ≥ ea & eb ≥ fb ≥ ˆ eb CZ = (1 − fa) + (1 − fb)

Xa Xb Ya ˆ Yb Za Zb ˆ Ya Yb ωa1 ωb1 ωa2 ωb2 ωa3 ωb3 ωa1 ωb1 ωa2 ωb2 ωa3 ωb3

Theorem For ea = 1/2 ˆ ea = 1 fa = 17/22 eb = 1/2 ˆ eb = 0 fb = 9/34 1-letter SC : R0 + R1 ≤ 1 and 11 10R0 + R1 ≤ 18 17 = 11 10CZ 2-letter SC : R0 + R1 ≤ 1 and 484 435R0 + R1 ≤ 528 493 = 484 435CZ

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 9 / 17

slide-24
SLIDE 24

Plot

1

180 187

R1 R0 2-letter & 1-letter SC regions

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 10 / 17

slide-25
SLIDE 25

Plot

44 119 7 17 10 17 75 119

R1 R0 2-letter & 1-letter SC regions

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 10 / 17

slide-26
SLIDE 26

1-letter SC

The distribution that achieves the corner-point.

◮ Let U be a ternary random variable

     P(U = 0) = 13

34

(Xa, Xb)|{U = 0} = (0, 0) P(U = 1) = 7

34

(Xa, Xb)|{U = 1} = (M, 0) P(U = 2) = 14

34

(Xa, Xb)|{U = 2} = (M, M) where M is an unbiased binary random variable

◮ Let Q be the random variable that symmetrizes the distribution of X ◮ Let ˜

U = (U, Q) and substitute ( ˜ U, X) into SC R0 ≤ I( ˜ U; Z) = 10 17 R0 + R1 ≤ I( ˜ U; Z) + I(X; Y | ˜ U) = 1 R0 + R1 ≤ I(X; Y ) = 1 R0 + R1 ≤ I( ˜ U; Z) + I(X; ˆ Y | ˜ U) = 1 R0 + R1 ≤ I(X; ˆ Y ) = 1

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 11 / 17

slide-27
SLIDE 27

2-letter SC

The distribution that achieves the corner-point. M1 & M2 two independent unbiased binary r.v.

     P(U = 0) =

20 119

(Xa1, Xb1, Xa2, Xb2)|{U = 0} = (0, 0, 0, 0) P(U = 1) =

11 119

(Xa1, Xb1, Xa2, Xb2)|{U = 1} = (M1, 0, M2, 0) P(U = 2) =

88 119

(Xa1, Xb1, Xa2, Xb2)|{U = 2} = (M1, M1, M1, 0)

◮ Let Q be the random variable that symmetrizes the distribution of X ◮ Let ˜

U = (U, Q) and substitute ( ˜ U, X) into SC ⇒ (R0, R1) = ( 75

119, 44 119)

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 12 / 17

slide-28
SLIDE 28

2-letter SC

The distribution that achieves the corner-point. M1 & M2 two independent unbiased binary r.v.

     P(U = 0) =

20 119

(Xa1, Xb1, Xa2, Xb2)|{U = 0} = (0, 0, 0, 0) P(U = 1) =

11 119

(Xa1, Xb1, Xa2, Xb2)|{U = 1} = (M1, 0, M2, 0) P(U = 2) =

88 119

(Xa1, Xb1, Xa2, Xb2)|{U = 2} = (M1, M1, M1, 0)

◮ Let Q be the random variable that symmetrizes the distribution of X ◮ Let ˜

U = (U, Q) and substitute ( ˜ U, X) into SC ⇒ (R0, R1) = ( 75

119, 44 119)

Observation

◮ A linear code achieves the 2-letter region

A natural question Let M = (M1, ..., Mm) be mutually independent unbiased bits. Let Xn

a = AM and Xn b = BM, where A, B are n × m matrices.

What is the rate region achieved by such a linear coding scheme. (variables are m,A, B).

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 12 / 17

slide-29
SLIDE 29

Outer bound

Routine Idea: Intersection of the two capacity regions (ignoring one of the users) Theorem For ea = 1/2 ˆ ea = 1 fa = 17/22 eb = 1/2 ˆ eb = 0 fb = 9/34 ignoring ˆ Y : R0 + R1 ≤ 1 and 11 5 R0 + R1 ≤ 11 5 CZ ignoring Y : R0 + R1 ≤ 1 and 34 25R0 + R1 ≤ 34 25CZ

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 13 / 17

slide-30
SLIDE 30

Plot

14 99 7 27 44 119 7 17 10 17 75 119 20 27 85 99

R1 R0 2-letter & 1-letter SC

  • Trad. Outer Bound

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 14 / 17

slide-31
SLIDE 31

Plot

14 99 7 27 44 119 7 17 10 17 75 119 20 27 85 99

R1 R0 2-letter & 1-letter SC

  • Trad. Outer Bound

New Outer Bound

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 14 / 17

slide-32
SLIDE 32

Idea for new outer-bound

From limiting n-letter inner bound that goes to capacity:

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 15 / 17

slide-33
SLIDE 33

Idea for new outer-bound

From limiting n-letter inner bound that goes to capacity: Theorem (Concentration of mutual information over memoryless product erasure channel) Consider a product erasure channel, Wa(ya|xa) ⊗ Wb(yb|xb), mapping Xa, Xb to Ya, Yb with erasure probabilities ǫa, ǫb, respectively. Then I(Xn

a , Xn b ; Y n a , Y n b ) = H(⌊n(1 − ǫa)⌋, ⌊n(1 − ǫb)⌋) + O

  • n log n
  • ,

where Hn(k, l) = 1 n

k

n

l

  • S,T⊆[n]:|S|=k,|T|=l

H(XaS, XbT ).

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 15 / 17

slide-34
SLIDE 34

Idea for new outer-bound

From limiting n-letter inner bound that goes to capacity: Theorem (Concentration of mutual information over memoryless product erasure channel) Consider a product erasure channel, Wa(ya|xa) ⊗ Wb(yb|xb), mapping Xa, Xb to Ya, Yb with erasure probabilities ǫa, ǫb, respectively. Then I(Xn

a , Xn b ; Y n a , Y n b ) = H(⌊n(1 − ǫa)⌋, ⌊n(1 − ǫb)⌋) + O

  • n log n
  • ,

where Hn(k, l) = 1 n

k

n

l

  • S,T⊆[n]:|S|=k,|T|=l

H(XaS, XbT ). Using (essentially) sub-modularity of entropy, we can establish that lim sup

n

max

p(xn

a,xn b )

1 n 85 160H(n 2 , n 2 ) + 75 160H(0, n) − 187 160H(5n 22 , 25n 34 )

  • ≤ 0.

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 15 / 17

slide-35
SLIDE 35

Outer bound continued

Theorem (Outer bound) Any achievable rate pair (R0, R1) must satisfy the constraints.

R0 + R1 ≤ 1 and 187 160R0 + R1 ≤ 18 16.

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 16 / 17

slide-36
SLIDE 36

Outer bound continued

Theorem (Outer bound) Any achievable rate pair (R0, R1) must satisfy the constraints.

R0 + R1 ≤ 1 and 187 160R0 + R1 ≤ 18 16.

Achievability If there is a non-trivial collection (Xn

a , Xn b ) such that

Hn(n 2 , n 2 ) = Hn(n 2 , 25n 34 ) + o(n), 5 11Hn(n 2 , 25n 34 ) + 6 11Hn(0, 25n 34 ) = H(5n 22 , 25n 34 ) + o(n), 8 17Hn(0, n) + 9 17Hn(0, n 2 ) = Hn(0, 25n 34 ) + o(n), 17 25Hn(0, 25n 34 ) = Hn(0, n 2 ) + o(n),

then there are non-trivial points of the outer bound that are achievable.

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 16 / 17

slide-37
SLIDE 37

Outer bound continued

Theorem (Outer bound) Any achievable rate pair (R0, R1) must satisfy the constraints.

R0 + R1 ≤ 1 and 187 160R0 + R1 ≤ 18 16.

Achievability If there is a non-trivial collection (Xn

a , Xn b ) such that

Hn(n 2 , n 2 ) = Hn(n 2 , 25n 34 ) + o(n), 5 11Hn(n 2 , 25n 34 ) + 6 11Hn(0, 25n 34 ) = H(5n 22 , 25n 34 ) + o(n), 8 17Hn(0, n) + 9 17Hn(0, n 2 ) = Hn(0, 25n 34 ) + o(n), 17 25Hn(0, 25n 34 ) = Hn(0, n 2 ) + o(n),

then there are non-trivial points of the outer bound that are achievable. Suggests: "MDS-like" (linear increase followed by flat region) code-construction for Xn

a , Xn b

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 16 / 17

slide-38
SLIDE 38

Conclusion

Observations

◮ Sub-optimality of superposition coding region ◮ Sub-optimality of Korner’s image-size characterization ◮ Linear code achieves 2-letter inner bound ◮ A new (explicit) outer bound from limiting n-letter inner bound ◮ Outer bound yields insights into structure of good codes

Chandra Nair and Mehdi Yazadanpanah GIC 23 Mar, 2018 17 / 17