Universal Polarization for Processes with Memory Boaz Shuval and - - PowerPoint PPT Presentation

universal polarization for processes with memory
SMART_READER_LITE
LIVE PREVIEW

Universal Polarization for Processes with Memory Boaz Shuval and - - PowerPoint PPT Presentation

Universal Polarization for Processes with Memory Boaz Shuval and Ido Tal Andrew and Erna Viterbi Department of Electrical Engineering Technion Israel Institute of Technology Haifa, 32000, Israel July 2019 1/16 Setting Communication


slide-1
SLIDE 1

Universal Polarization for Processes with Memory

Boaz Shuval and Ido Tal

Andrew and Erna Viterbi Department of Electrical Engineering Technion — Israel Institute of Technology Haifa, 32000, Israel

July 2019

1/16

slide-2
SLIDE 2

Setting

Communication with uncertainty:

Encoder: Knows channel belongs to a set of channels Decoder: Knows channel statistics (e.g., via estimation)

Memory:

In channels In input distribution

Universal code:

Vanishing error probability over set Best rate (infimal information rate over set)

Goal:

Universal Code based on Polarization

2/16

slide-3
SLIDE 3

Why?

Polar codes have many good properties

rate-optimal (even under memory!) vanishing error probability low complexity encoding/decoding/construction

But...

Polar codes must be tailored to the channel at hand

Sometimes, the channel isn’t known apriori to encoder

Example: Frequency Selective Fading ⇒ ISI

  • Yn = h0Xn+

m

  • i=1

hiXn−i +noise

3/16

slide-4
SLIDE 4

Polar Codes: lightning reminder

Channel XN

1

YN

1

Goal: Decode XN

1 from YN 1

4/16

slide-5
SLIDE 5

Polar Codes: lightning reminder

Channel XN

1

YN

1

FN

1 = fArıkan(XN 1)

Goal: Decode XN

1 from YN 1

Transform fArıkan is one-to-one and onto

recursively defined

Decoding XN

1 ⇐

⇒ Decoding FN

1

4/16

slide-6
SLIDE 6

Polar Codes: lightning reminder

Channel XN

1

YN

1

FN

1 = fArıkan(XN 1)

Gi = (Fi−1

1

, YN

1 )

Successive-Cancellation decoding:

Compute Gi from decoded Fi−1

1

Decode Fi from Gi

Polarization: fix β < 1/2

Low-Entropy set: LN = {i | H(Fi|Gi) < 2−Nβ} High-Entropy set: HN = {i | H(Fi|Gi) > 1 − 2−Nβ} For N large, |LN| + |HN| ≈ N

Coding scheme (simplified):

i ∈ LN ⇒ Transmit data i ∈ HN ⇒ Reveal to decoder

4/16

slide-7
SLIDE 7

Polar Codes: lightning reminder

Channel XN

1

YN

1

FN

1 = fArıkan(XN 1)

Gi = (Fi−1

1

, YN

1 )

Successive-Cancellation decoding:

Compute Gi from decoded Fi−1

1

Decode Fi from Gi

Polarization: fix β < 1/2

Low-Entropy set: LN = {i | H(Fi|Gi) < 2−Nβ} High-Entropy set: HN = {i | H(Fi|Gi) > 1 − 2−Nβ} For N large, |LN| + |HN| ≈ N

Coding scheme (simplified):

i ∈ LN ⇒ Transmit data i ∈ HN ⇒ Reveal to decoder

Not Universal! LN, HN channel-dependent

4/16

slide-8
SLIDE 8

Previous Work on Universal Polarization

All for the memoryless case Works with memoryless settings similar to ours:

Hassani & Urbanke 2014 S ¸as ¸o˘ glu& Wang 2016 (conference version: 2014)

5/16

slide-9
SLIDE 9

Previous Work on Universal Polarization

All for the memoryless case Works with memoryless settings similar to ours:

Hassani & Urbanke 2014 S ¸as ¸o˘ glu& Wang 2016 (conference version: 2014)

5/16

slide-10
SLIDE 10

Our Construction

Simplified generalization of S ¸as ¸o˘ glu-Wang construction Memory at channel and/or input Two stages: “slow” and “fast”

6/16

slide-11
SLIDE 11

Our Construction

Simplified generalization of S ¸as ¸o˘ glu-Wang construction Memory at channel and/or input Two stages: “slow” and “fast” XN

1

FN

1

f H L

f one-to-one and onto, recursively defined (η, L, H)-monopolarization: For any η > 0, there exist N and index sets L, H such that either H(Fi|Gi) < η for all i ∈ L

  • r

H(Fi|Gi) > 1 − η for all i ∈ H Universal: L, H process independent Slow

6/16

slide-12
SLIDE 12

Our Construction

Simplified generalization of S ¸as ¸o˘ glu-Wang construction Memory at channel and/or input Two stages: “slow” and “fast”

f H L f H L f H L f H L

ˆ N copies N i ∈ L ⇒ H(Fi|Gi) < η

6/16

slide-13
SLIDE 13

Our Construction

Simplified generalization of S ¸as ¸o˘ glu-Wang construction Memory at channel and/or input Two stages: “slow” and “fast”

f H L f H L f H L f H L

ˆ N copies N fArıkan ˆ N

6/16

slide-14
SLIDE 14

Our Construction

Simplified generalization of S ¸as ¸o˘ glu-Wang construction Memory at channel and/or input Two stages: “slow” and “fast”

f H L f H L f H L f H L

ˆ N copies N fArıkan ˆ N fArıkan fArıkan |L| copies

6/16

slide-15
SLIDE 15

Our Construction

Simplified generalization of S ¸as ¸o˘ glu-Wang construction Memory at channel and/or input Two stages: “slow” and “fast”

f H L f H L f H L f H L

ˆ N copies N fArıkan ˆ N fArıkan fArıkan |L| copies Pe ≤ |L| · 2−ˆ

Rate ≈ |L| N

6/16

slide-16
SLIDE 16

Our Construction

Simplified generalization of S ¸as ¸o˘ glu-Wang construction Memory at channel and/or input Two stages: “slow” and “fast”

f H L f H L f H L f H L

ˆ N copies N fArıkan ˆ N fArıkan fArıkan |L| copies Pe ≤ |L| · 2−ˆ

Rate ≈ |L| N

Our focus

6/16

slide-17
SLIDE 17

A framework for memory

Stationary process: (Si, Xi, Yi)N

i=1

Finite number of states: Si ∈ S, where |S| < ∞ Hidden state: Si is unknown to encoder and decoder

Markov property: P(si, xi, yi|{sj, xj, yj}j<i) = P(si, xi, yi|si−1) FAIM state sequence: Finite-state, aperiodic, irreducible Markov chain (Xi, Yi)N

i=1 FAIM-derived process

FAIM ⇒ mixing: if M − N large enough, (XN

−∞, YN −∞) and (X∞ M , Y∞ M ) almost independent

7/16

slide-18
SLIDE 18

Forgetfulness

Required for proof of monopolarization FAIM process (Si, Xi, Yi) is forgetful if for any ǫ > 0 there exists natural λ such that if k ≥ λ, I(S1; Sk|Xk

1, Yk 1) ≤ ǫ

I(S1; Sk|Yk

1) ≤ ǫ

Neither inequality implies the other FAIM does not imply forgetfulness We have a sufficient condition for forgetfulness

Under it, ǫ decreases exponentially with λ

8/16

slide-19
SLIDE 19

FAIM Does Not Imply Forgetfulness

1 2 3 4 a b Yj =

  • a,

Sj ∈ {1, 2} b, Sj ∈ {3, 4} I(S1; Sk|Yk

1) → 0

9/16

slide-20
SLIDE 20

Why Forgetfulness?

(Si, Xi, Yi) forgetful if for any ǫ > 0 exists λ such that k ≥ λ = ⇒

  • I(S1; Sk|Xk

1, Yk 1) ≤ ǫ

I(S1; Sk|Yk

1) ≤ ǫ

Can show: for any k + 1 ≤ i ≤ N − k 0 ≤ H(Xi|Xi−1

i−k , Yi+k i−k) − H(Xi|Xi−1 1

, YN

1 ) ≤ 2ǫ

Takeaway point

Only a “window” surrounding i really matters

10/16

slide-21
SLIDE 21

Slow Stage is Monopolarizing

FAIM-derived: (Xi, Yi) derived from (Si, Xi, Yi) such that P(si, xi, yi|{sj, xj, yj}j<i) = P(si, xi, yi|si−1) with Si finite-state, aperiodic, irreducible, Markov Forgetful: for any ǫ > 0 there exists λ such that if k ≥ λ, I(S1; Sk|Xk

1, Yk 1) ≤ ǫ

I(S1; Sk|Yk

1) ≤ ǫ

Main Result (simplified)

If process (Xi, Yi) is FAIM-derived and forgetful, the slow stage is monopolarizing, with universal L, H (unrelated to process)

11/16

slide-22
SLIDE 22

Slow Stage

Presented for the case |L| = |H| Transforms XNn

1 ֌ YNn 1 f

= ⇒ FNn

1 ֌ GNn 1

Recursively defined

Parameters L0, M0 Level 0 length: N0 = 2L0 + M0 Level n length: Nn = 2Nn−1

Index types at level n:

First Ln indices: lateral Middle Mn indices: medial Last Ln indices: lateral

Level-n block

lateral lateral medial

F1 ֌ G1 FLn ֌ GLn FLn+1 ֌ GLn+1 FLn+Mn ֌ GLn+Mn FLn+Mn+1 ֌ GLn+Mn+1 FNn ֌ GNn

Ln Ln Mn

transmitted received decode Fi from Gi

12/16

slide-23
SLIDE 23

Slow Stage — Lateral Recursion

lateral lateral

U ֌ Q

lateral lateral

V ֌ R

lateral lateral

F ֌ G

Ln Mn Ln Ln Mn Ln

Ln+1 = 2Ln + 1 Mn+1 = 2(Mn − 1) Ln+1 = 2Ln + 1

Level-n block Level-n block Level-(n + 1) block

Lateral indices always remain lateral Two medial indices become lateral

13/16

slide-24
SLIDE 24

Slow Stage — Medial Recursion

Two type of medial indices:

H L

Alternating: H, L, H, L, . . . Two medial become lateral: ULn+1, VLn+Mn Join H from one block with L from

  • ther

lateral lateral

U ֌ Q

lateral lateral

V ֌ R Level-n block Level-n block

ULn+1 VLn+Mn

H L H L H L H L H L H L H L H L

14/16

slide-25
SLIDE 25

Slow Stage — Medial Recursion

Two type of medial indices:

H L

Alternating: H, L, H, L, . . . Two medial become lateral: ULn+1, VLn+Mn Join H from one block with L from

  • ther

lateral lateral

U ֌ Q

lateral lateral

V ֌ R Level-n block Level-n block

ULn+2 VLn+1

H L H L H L H L H L H L H L H L + F2Ln+2 F2Ln+3

medial (n + 1)

14/16

slide-26
SLIDE 26

Slow Stage — Medial Recursion

Two type of medial indices:

H L

Alternating: H, L, H, L, . . . Two medial become lateral: ULn+1, VLn+Mn Join H from one block with L from

  • ther

lateral lateral

U ֌ Q

lateral lateral

V ֌ R Level-n block Level-n block

ULn+2 VLn+1 ULn+3 VLn+2

H L H L H L H L H L H L H L H L + + F2Ln+2 F2Ln+3 F2Ln+4 F2Ln+5

medial (n + 1)

14/16

slide-27
SLIDE 27

Slow Stage — Medial Recursion

Two type of medial indices:

H L

Alternating: H, L, H, L, . . . Two medial become lateral: ULn+1, VLn+Mn Join H from one block with L from

  • ther

lateral lateral

U ֌ Q

lateral lateral

V ֌ R Level-n block Level-n block

ULn+2 VLn+1 ULn+3 VLn+2 ULn+Mn−1 VLn+Mn−2

H L H L H L H L H L H L H L H L + + + F2Ln+2 F2Ln+3 F2Ln+4 F2Ln+5 F2Ln+2Mn−4 F2Ln+2Mn−3

medial (n + 1)

14/16

slide-28
SLIDE 28

Slow Stage — Medial Recursion

Two type of medial indices:

H L

Alternating: H, L, H, L, . . . Two medial become lateral: ULn+1, VLn+Mn Join H from one block with L from

  • ther

lateral lateral

U ֌ Q

lateral lateral

V ֌ R Level-n block Level-n block

ULn+2 VLn+1 ULn+3 VLn+2 ULn+Mn−1 VLn+Mn−2 ULn+Mn VLn+Mn−1

H L H L H L H L H L H L H L H L + + + + F2Ln+2 F2Ln+3 F2Ln+4 F2Ln+5 F2Ln+2Mn−4 F2Ln+2Mn−3 F2Ln+2Mn−2 F2Ln+2Mn−1

medial (n + 1)

14/16

slide-29
SLIDE 29

Slow Stage is Monopolarizing

XNn

1

FNn

1

f H L Slow stage:

Main Result

If process (Xi, Yi) is FAIM-derived and forgetful, for every η > 0, there exist L0, M0, nth such that the slow stage of level at least nth is (η,L,H)-monopolarizing H⋆(X|Y) ≤ 1/2 ⇒ H(Fi|Gi) < η for all i ∈ L H⋆(X|Y) ≥ 1/2 ⇒ H(Fi|Gi) > 1 − η for all i ∈ H

15/16

slide-30
SLIDE 30

Slow Stage is Monopolarizing

XNn

1

FNn

1

f H L Slow stage:

Main Result

If process (Xi, Yi) is FAIM-derived and forgetful, for every η > 0, there exist L0, M0, nth such that the slow stage of level at least nth is (η,L,H)-monopolarizing H⋆(X|Y) ≤ 1/2 ⇒ H(Fi|Gi) < η for all i ∈ L H⋆(X|Y) ≥ 1/2 ⇒ H(Fi|Gi) > 1 − η for all i ∈ H Universal: sets L, H process independent

15/16

slide-31
SLIDE 31

Elements of Proof

Parameters L0, M0 related to memory:

L0 large if forgetfulness slow M0 large if mixing slow

Step 1:

Replace slow stage with a modification Replace process with a block-independent process Establish monopolarization

Step 2:

Choose suitable L0, M0 Show negligible difference between step 1 replacements and actual process, slow stage Implies main result

16/16