on the construction of polar codes for channels with
play

On the Construction of Polar Codes for Channels with Moderate Input - PowerPoint PPT Presentation

On the Construction of Polar Codes for Channels with Moderate Input Alphabet Sizes Ido Tal 1 / 19 Problem: Construction of polar (LDPC) codes, for a channel with moderate input alphabet size q . Say, q 16. Punchline: Provably hard


  1. On the Construction of Polar Codes for Channels with Moderate Input Alphabet Sizes Ido Tal 1 / 19

  2. Problem: Construction of polar (LDPC) codes, for a channel with moderate input alphabet size q . Say, q ≥ 16. Punchline: Provably hard ∗†‡§ . ∗ For a specific channel † under a certain construction model ‡ deterministically § some more assumptions 2 / 19

  3. Given: ◮ Underlying channel W : X → Y und ◮ |X| = q ◮ Uniform input distribution is capacity achieving ◮ Codeword length n = 2 m Goal: ◮ Assuming uniform input, calculate misdecoding probability of synthesized channels W ( m ) : X → Y i , 0 ≤ i < n i ◮ Unfreeze channels with very low probability of misdecoding 3 / 19

  4. P U ( X ) � uniform distribution on input alphabet X Algorithm: Naive solution input : Underlying channel W , index i = � b 1 , b 2 , . . . , b m � 2 output : P e ( W ( m ) , P U ( X ) ) i W ← W for j = 1 , 2 , . . . , m do if b j = 0 then W ← W − else W ← W + return P e (W , P U ( X ) ) Problem: Y i grows exponentially with n . 4 / 19

  5. P U ( X ) � uniform distribution on input alphabet X Algorithm: Degrading solution input : Underlying channel W , index i = � b 1 , b 2 , . . . , b m � 2 , bound on output alphabet size L output : Upper bound on P e ( W ( m ) , P U ( X ) ) i Q ← degrading merge ( W , L , P U ( X ) ) for j = 1 , 2 , . . . , m do if b j = 0 then W ← Q − else W ← Q + Q ← degrading merge (W , L , P U ( X ) ) return P e (Q , P U ( X ) ) Question: How good of an approximation to W is degrading merge (W , L , P U ( X ) )? 5 / 19

  6. Notation: ◮ W : X → Y — generic memoryless channel ◮ q = |X| — input alphabet size ◮ P X — input distribution ◮ Q : X → Y ′ — degraded version of W ◮ L — bound on new output alphabet size, |Y ′ | ≤ L ◮ X — input to W or Q ◮ Y — output of W ◮ Y ′ — output of Q Goal: degrading merge ( W , L , P X ) must find Q : X → Y ′ such that ◮ Q degraded with respect to W ◮ |Y ′ | ≤ L ◮ ∆ = I ( X ; Y ) − I ( X ; Y ′ ) is “small” 6 / 19

  7. An implementation of degrading merge ( W , L , P X ) exists [TalSharovVardy] for which �� 1 � 1 / q � ∆ = I ( X ; Y ) − I ( X ; Y ′ ) ≤ O L Apropos: similar behaviour in upgraded case [PeregTal] Totally useless (at least in theory), for moderate q : L ≈ 10 32 q = 16 , ∆ ≤ 0 . 01 = ⇒ Good luck. . . 7 / 19

  8. An inherent difficulty? What can be said about DC ( q , L ) � sup min ( I ( W ) − I ( Q )) . Q : Q ≺ W , W , P X | out ( Q ) |≤ L We already know that �� 1 � 1 / q � DC ( q , L ) ≤ O L Need: a lower bound on DC ( q , L ) 8 / 19

  9. Cut to the end DC ( q , L ) � sup min ( I ( W ) − I ( Q )) Q : Q ≺ W , W , P X | out ( Q ) |≤ L We will shortly prove that 2 �� 1 � � q − 1 DC ≥ O L Above attained for ◮ Uniform input distribution P X = P U ( X ) ◮ Sequence W 1 , W 2 , . . . of “progressively hard channels” ◮ The capacity achieving input distribution of each W M is the uniform distribution P U ( X ) 9 / 19

  10. Consequences: Try and build a polar code for W M . . . Algorithm: Degrading solution input : Underlying channel W , index i = � b 1 , b 2 , . . . , b m � 2 , bound on output alphabet size L output : Upper bound on P e ( W ( m ) , P U ( X ) ) i Q ← degrading merge ( W , L , P U ( X ) ) for j = 1 , 2 , . . . , m do if b j = 0 then W ← Q − else W ← Q + Q ← degrading merge (W , L , P U ( X ) ) return P e (Q , P U ( X ) ) 10 / 19

  11. Consequences: Try and build a polar code for W M . . . ◮ Would like number of good channels to be ≈ n · I ( W M ) ◮ However, number of good channels is upper bounded by � � n · I degrading merge ( W M , L , P U ( X ) ) 2 � �� 1 �� � q − 1 ≥ n · I ( W M ) − O L For q = 16, in order to lose at most 0 . 01, need L ≈ 10 15 11 / 19

  12. LDPC: Same problem when trying to design an LDPC code for W M ◮ Pick a code ensamble with rate close to I ( W M ) ◮ Use density evolution to asses code: 1. Initialize ◮ Assume all-zero codeword ◮ Quantize output letters: letters with close posteriors are grouped together 2. Main loop ◮ Already hopeless at this point: main loop is with respect to quantized channel, which has mutual information below design rate 12 / 19

  13. The channel W M : For an integer M ≥ 1, define W M : X → Y M as follows: ◮ Input alphabet is X = { 1 , 2 , . . . , q } ◮ Output alphabet is q � � � Y M = � j 1 , j 2 , . . . , j q � : j 1 , j 2 , . . . , j q ≥ 0 , j x = M , x =1 where j x are non-negative integers summing to M ◮ Channel transition probabilities: q · j x W ( � j 1 , j 2 , . . . , j q �| x ) = � M + q − 1 � M q − 1 ◮ Input distribution unifrom = ⇒ all output letters equally likely 13 / 19

  14. The channel W M : ◮ Posterior probabilities P ( X = x | Y = � j 1 , j 2 , . . . , j q � ) = j x M ◮ Shorthand: output letter is labelled by posterior probabilities vector � j 1 , j 2 , . . . , j q � � ( j 1 / M , j 2 / M , . . . , j q / M ) 14 / 19

  15. Optimal degrading: Claim [KurkoskiYagi]: ◮ Let W : X → Y , P X , and L be given. ◮ Let Q : X → Z be an optimal degrading of W to a channel Q with |Z| ≤ L . ◮ That is, I ( X , Y ) − I ( X , Y ′ ) is minimized. ◮ Then, Q is gotten from W by defining a partition ( A i ) L i =1 of Y and mapping with probability 1 all symbols in A i to a single symbol z i ∈ Z Let ( A i ) L i =1 be such a partition with respect to W M 15 / 19

  16. L 2 squared bound: Lemma: For A = A i as above, let ∆( A ) be the drop in mutual information incurred by merging all the letters in A i into a single letter. Then, ∆( A ) ≥ ˜ ∆( A ) , where 1 1 ˜ � p � 2 � � p − ¯ ¯ ∆( A ) = 2 , p = | A | p . � M + q − 1 � 2 q − 1 p ∈ A p ∈ A 16 / 19

  17. Bounding in terms of | A | : Lemma: L L L q +1 � � ˜ � q − 1 + o (1) , ∆( A i ) ≥ ∆( A i ) ≥ const ( q ) · | A i | i =1 i =1 i =1 where the o (1) is a function of M alone and goes to 0 as M → ∞ Observation: Up to the o (1), expression is convex in | A i | . Thus, sum is lower bounded by setting | A i | = |Y M | / L . 17 / 19

  18. Theorem: 2 2 q − 1 � 1 � � 1 � q − 1 q − 1 DC ( q , L ) ≥ 2( q + 1) · · , σ q − 1 · ( q − 1)! L where σ q − 1 is the constant for which the volume of a sphere in R q − 1 of radius r is σ q − 1 r q − 1 18 / 19

  19. Backup ◮ Just how representative is W M ? ◮ What can be done? ◮ Channels W M “converges” to ◮ W ∞ : X → X × [0 , 1] q ◮ Given an input x , the channel picks ϕ 1 , ϕ 2 , . . . , ϕ q , non-negative reals summing to 1. All possible choices are equally likely, Dirichlet(1,1,. . . ,1) ◮ Then, the input x is transformed into x + i (with a modulo operation where appropriate) with probability ϕ i ◮ The transformed symbol along with the vector ( ϕ 1 , ϕ 2 , . . . , ϕ q ) are the output of the channel 19 / 19

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend