SLIDE 1 Polar Codes for the Deletion Channel: Weak and Strong Polarization
Ido Tal1 Henry D. Pfister2 Arman Fazeli3 Alexander Vardy3
1Technion 2Duke 3UCSD
SLIDE 2 1 / 21
Big picture first
A polar coding scheme for the deletion channel where the:
◮ Deletion channel has constant deletion probability δ ◮ Fix a hidden-Markov input distribution1. ◮ Code rate converges to information rate ◮ Error probability decays like 2−Λγ, where γ < 1
3 and Λ is the
codeword length ◮ Decoding complexity is at most O(Λ1+3γ) ◮ Achieves hidden-Markov capacity!
1i.e., a function of an aperiodic, irreducible, finite-state Markov chain
SLIDE 3 1 / 21
Big picture first
A polar coding scheme for the deletion channel where the:
◮ Deletion channel has constant deletion probability δ ◮ Fix a hidden-Markov input distribution1. ◮ Code rate converges to information rate ◮ Error probability decays like 2−Λγ, where γ < 1
3 and Λ is the
codeword length ◮ Decoding complexity is at most O(Λ1+3γ) ◮ Achieves hidden-Markov capacity! Equals true capacity?
1i.e., a function of an aperiodic, irreducible, finite-state Markov chain
SLIDE 4 1 / 21
Big picture first
A polar coding scheme for the deletion channel where the:
◮ Deletion channel has constant deletion probability δ ◮ Fix a hidden-Markov input distribution1. ◮ Code rate converges to information rate ◮ Error probability decays like 2−Λγ, where γ < 1
3 and Λ is the
codeword length ◮ Decoding complexity is at most O(Λ1+3γ) ◮ Achieves hidden-Markov capacity! Equals true capacity? ◮ Key ideas:
◮ Polarization operations defined for trellises ◮ Polar codes modified to have guard bands of ‘0’ symbols
1i.e., a function of an aperiodic, irreducible, finite-state Markov chain
SLIDE 5
2 / 21
A brief history of the binary deletion channel
◮ Early Work: Levenshtein [Lev66] and Dobrushin [Dob67] ◮ LDPC Codes + Turbo Equalization: Davey-MacKay [DM01] ◮ Coding and Capacity Bounds by Mitzenmacher [Mit09] and many more: [FD10], [MTL12], [CK15], [RD15], [Che19] ◮ Polar codes: [TTVM17], [TFVL17], [TFV18] ◮ Our Contributions:
◮ Proof of weak polarization for constant deletion rate ◮ Strong polarization for constant deletion rate with guard bands ◮ Our trellis perspective also establishes weak polarization for channels with insertions, deletions, and substitutions
SLIDE 6 3 / 21
Hidden-Markov input process
Example: (1, ∞) Run-Length Constraint
1
1−α 1 α
◮ Input process is (Xj), j ∈ Z ◮ Marginalization of (Sj, Xj), j ∈ Z ◮ State (Sj), j ∈ Z, is Markov, stationary, irreducible, aperiodic ◮ For all j, it holds that PSj,Xj|Sj−1
−∞,X j−1 −∞ = PSj,Xj|Sj−1
SLIDE 7
4 / 21
Code rate
The code rate of our scheme approaches I(X; Y ) = lim
N→∞
1 N H(X) − lim
N→∞
1 N H(X|Y) , ◮ X = (X1, . . . , XN) is hidden-Markov input ◮ Y is the deletion channel output
SLIDE 8
5 / 21
Theorem (Strong polarization)
Fix a regular hidden-Markov input process. For any fixed γ ∈ (0, 1/3), the rate of our coding scheme approaches the mutual-information rate between the input process and the deletion channel output. For large enough blocklength Λ, the probability of error is at most 2−Λγ.
SLIDE 9
6 / 21
Uniform input process
◮ It is known that a memoryless input distribution is suboptimal ◮ To keep this talk simple, we will however assume that the input process is uniform, and thus memoryless ◮ That is, the Xi are i.i.d. and Ber(1/2)
SLIDE 10 7 / 21
The polar transform
◮ Let x = (x1, . . . , xN) ∈ {0, 1}N be a vector of length N = 2n ◮ Define
◮ minus transform: x[0] (x1 ⊕ x2, x3 ⊕ x4, . . . , xN−1 ⊕ xN) ◮ plus transform: x[1] ( x2, x4, . . . , xN) ◮ Both are vectors of length N/2
◮ Define x[b1,b2,...,bλ] recursively: z = x[b1,b2,...,bλ−1] , x[b1,b2,...,bλ] = z[bλ] ◮ The polar transform of x is u = (u1, u2, . . . , uN), where for i = 1 +
n
bj2n−j we have ui = x[b1,b2,...,bn]
SLIDE 11
8 / 21
Polarization of trellises
◮ The decoder sees the received sequence y ◮ Ultimately, we want an efficient method of calculating P(Ui = ˆ ui|Ui−1 = ˆ ui−1, Y = y) ◮ Towards this end, let us first show an efficient method of calculating the joint probability P(X = x, Y = y)
◮ Generalizes the SC trellis decoder of Wang et. al. [WLH14], and the polar decoder for deletions by Tian et. al. [TFVL17]
SLIDE 12
9 / 21
Deletion channel trellis
y1 =0 y2 =1 y3 =1 x1 x2 x3 x4 xj yi
◮ Example: N = 4 inputs with length-3 output 011 ◮ Edge labels: blue xj = 0 and red xj = 1 ◮ Direction: diagonal = no deletion and horizontal = deletion
SLIDE 13
9 / 21
Deletion channel trellis
y1 =0 y2 =1 y3 =1 x1 x2 x3 x4 xj yi
◮ Example: N = 4 inputs with length-3 output 011 ◮ Edge labels: blue xj = 0 and red xj = 1 ◮ Direction: diagonal = no deletion and horizontal = deletion
SLIDE 14
10 / 21
Deletion channel trellis and the minus operation
x1 x2 x3 x4
δ/2 δ/2 δ/2 δ/2 δ/2 δ/2 δ/2 δ/2 δ / 2 δ / 2 δ / 2 δ / 2 δ / 2 δ / 2
x1 ⊕ x2 x3 ⊕ x4
◮ Half as many sections representing twice the channel uses
SLIDE 15 10 / 21
Deletion channel trellis and the minus operation
x1 x2 x3 x4
δ/2 δ/2 δ/2 δ/2 δ/2 δ/2 δ/2 δ/2 δ / 2 δ / 2 δ / 2 δ / 2 δ / 2 δ / 2
x1 ⊕ x2 x3 ⊕ x4
δ δ / 2 δδ/2 δ
2
/ 4
◮ Half as many sections representing twice the channel uses
◮ Edge weight is product of edge weights along length-2 paths ◮ Edge label (i.e., color) is the xor of labels along length-2 paths
SLIDE 16 10 / 21
Deletion channel trellis and the minus operation
x1 x2 x3 x4
δ/2 δ/2 δ/2 δ/2 δ/2 δ/2 δ/2 δ/2 δ / 2 δ / 2 δ / 2 δ / 2 δ / 2 δ / 2
x1 ⊕ x2 x3 ⊕ x4
δ δ / 2 δδ/2 δ
2
/ 4 δδ/2 δδ/2 δ
2
/ 4
◮ Half as many sections representing twice the channel uses
◮ Edge weight is product of edge weights along length-2 paths ◮ Edge label (i.e., color) is the xor of labels along length-2 paths
SLIDE 17 11 / 21
Weak polarization
Theorem
For any ǫ > 0, lim
N→∞
1 N
1
, Y) ∈ [ǫ, 1 − ǫ]
The proof follows along similar lines as the seminal proof: ◮ Define a tree process ◮ Show that the process is a submartingale ◮ Show that the submartingale can only converge to 0 or 1 All the above follow easily, once we notice the following ◮ Let X ⊙ X′ be two concatenated inputs to the channel ◮ Denote the corresponding output Y ⊙ Y′ ◮ Then, H(A|B, Y ⊙ Y′) ≥ H(A|B, Y, Y′)
SLIDE 18
12 / 21
Strong polarization
◮ Fix N = 2n, n0 = ⌊γ · n⌋ and n1 = ⌈(1 − γ) · n⌉ ◮ Define N0 = 2n0 and N1 = 2n1 ◮ Let X1, X2, . . . , XN1 by i.i.d. blocks of length N0 ◮ Suppose the channel input is X1 ⊙ X2 ⊙ · · · ⊙ XN1 ◮ Decoder sees Y1 ⊙ Y2 ⊙ · · · ⊙ YN1 ◮ If only we had a genie to “punctuate” the output to Y1, Y2, . . . , YN1, proving strong polarization would be easy. . .
SLIDE 19
13 / 21
A “good enough” genie
◮ We would like this:
SLIDE 20
13 / 21
A “good enough” genie
◮ We would like this: ◮ We will settle for this:
SLIDE 21
13 / 21
A “good enough” genie
◮ We would like this: ◮ We will settle for this: ◮ No head. . .
SLIDE 22
13 / 21
A “good enough” genie
◮ We would like this: ◮ We will settle for this: ◮ No head. . . ◮ No tail. . .
SLIDE 23
14 / 21
A “good enough” genie
◮ Decoder sees Y1 ⊙ Y2 ⊙ · · · ⊙ YN1 ◮ Decoder wants a genie to punctuate the above into Y1, Y2, . . . , YN1 ◮ Our “good enough” genie will give the decoder Y⋆
1, Y⋆ 2, . . . , Y⋆ N1
where Y⋆
i is Yi, with leading and trailing ‘0’ symbols removed
◮ Asymptotically, we have sacrificed nothing because I(X; Y) = I(X; Y⋆)
SLIDE 24
15 / 21
Building our genie
◮ Guard bands added at the encoder ◮ Denote x = xI ⊙ xII ∈ X 2n, where X = {0, 1} and xI = x2n−1
1
∈ X 2n−1 , xII = x2n
2n−1+1 ∈ X 2n−1
◮ That is, instead of transmitting x, we transmit, g(x), where g(x) x if n ≤ n0 g(xI) ⊙
ℓn
00 . . . 0 ⊙g(xII) if n > n0, ℓn 2⌊(1−ǫ)(n−1)⌋ ◮ ǫ is a ‘small’ constant
SLIDE 25
16 / 21
The genie in action
X XI XII G GI G△ GII Y YI Y△ YII Z ZI Z△ ZII ◮ Z is Y with leading and trailing ‘0’ symbols removed ◮ Guard band Z△ removed by splitting Z in half, and then removing leading and trailing 0 symbols from each half ◮ Genie successful if the middle of Z falls in the guard band
SLIDE 26
17 / 21
Conclusions
◮ Strong polarization for the deletion channel with constant deletion probability δ ◮ Error rate 2−Λγ comes from balancing strong polarization and guard-band failure ◮ If capacity of deletion channel achievable by hidden-Markov inputs, then we can achieve capacity!
SLIDE 27 18 / 21
References I
[Che19] Mahdi Cheraghchi. Capacity upper bounds for deletion-type channels. Journal of the ACM (JACM), 66(2):9, 2019. [CK15] Jason Castiglione and Aleksandar Kavcic. Trellis based lower bounds on capacities of channels with synchronization errors. In Information Theory Workshop, pages 24–28, Jeju, South Korea,
[DM01] Matthew C Davey and David JC MacKay. Reliable communication over channels with insertions, deletions, and substitutions. IEEE Transactions on Information Theory, 47(2):687–698, 2001. [Dob67] Roland L’vovich Dobrushin. Shannon’s theorems for channels with synchronization errors. Problemy Peredachi Informatsii, 3(4):18–36, 1967.
SLIDE 28 19 / 21
References II
[FD10] Dario Fertonani and Tolga M Duman. Novel bounds on the capacity of the binary deletion channel. IEEE Transactions on Information Theory, 56(6):2753–2765, 2010. [Lev66]
Binary codes capable of correcting deletions, insertions and reversals. Soviet Physics - Doklady, 10(8):707–710, February 1966. [Mit09] Michael Mitzenmacher. A survey of results for deletion channels and related synchronization channels. Probability Surveys, 6:1–33, 2009. [MTL12] Hugues Mercier, Vahid Tarokh, and Fabrice Labeau. Bounds on the capacity of discrete memoryless channels corrupted by synchronization and substitution errors. IEEE Transactions on Information Theory, 58(7):4306–4330, 2012.
SLIDE 29
20 / 21
References III
[RD15] Mojtaba Rahmati and Tolga M Duman. Upper bounds on the capacity of deletion channels using channel fragmentation. IEEE Transactions on Information Theory, 61(1):146–156, 2015. [TFV18] Kuangda Tian, Arman Fazeli, and Alexander Vardy. Polar coding for deletion channels: Theory and implementation. In IEEE International Symposium on Information Theory, pages 1869–1873, 2018. [TFVL17] Kuangda Tian, Arman Fazeli, Alexander Vardy, and Rongke Liu. Polar codes for channels with deletions. In 55th Annual Allerton Conference on Communication, Control, and Computing, pages 572–579, 2017. [TTVM17] E. K. Thomas, V. Y. F. Tan, A. Vardy, and M. Motani. Polar coding for the binary erasure channel with deletions. IEEE Communications Letters, 21(4):710–713, April 2017.
SLIDE 30
21 / 21
References IV
[WLH14] Runxin Wang, Rongke Liu, and Yi Hou. Joint successive cancellation decoding of polar codes over intersymbol interference channels. arXiv preprint arXiv:1404.3001, 2014.