in the finite memory length regime
play

in the Finite Memory Length Regime Joint work of Yu-Chih Huang , - PowerPoint PPT Presentation

2020 IEEE International Symposium on Information Theory Error Rate Analysis for Random Linear Streaming Codes in the Finite Memory Length Regime Joint work of Yu-Chih Huang , National Chiao Tung University, Shih-Chun Lin , National Taiwan


  1. 2020 IEEE International Symposium on Information Theory Error Rate Analysis for Random Linear Streaming Codes in the Finite Memory Length Regime Joint work of Yu-Chih Huang , National Chiao Tung University, Shih-Chun Lin , National Taiwan University of Science and Technology, I-Hsiang Wang , National Taiwan University, and Chih-Chun Wang , Pin-Wen Su , Purdue University 6/7/2020 Sponsored by NSF CCF-1422997, CCF-1618475 and CCF-1816013, and by MOST Taiwan 107-2628-E-011-003-MY3.

  2. Outline  Motivation and Related Work  System Model  Main Contributions  Information-Debt Under Finite Memory  Exact Error Rate Analysis  A Provably-Tight Closed-Form Error Rate Approximation  Numerical Verification and Conclusion

  3. 5G Communication Systems  IMT-2020  Enhanced mobile broadband (eMBB)  Massive machine type communications (mMTC)  Ultra-reliable and low latency communications (URLLC)  Low Latency Requirements  End-to-end delay ≤ 1 ms  Streaming codes may be a possible solution. The figure is copied from International Telecommunication Union, “Setting the Scene for 5G: Opportunities and Challenges.” (2018) . https://www.itu.int/dms_pub/itu-d/opb/pref/D-PREF-BB.5G_01-2018-PDF-E.pdf

  4. Streaming Codes  Very small queueing delay : Good for URLLC & mMTC  Streaming Codes:  Align the convolutional code structure with the actual transmission/encoding schedule.  There are other definitions of Streaming Codes, but we use this basic definition.  What we are interested: Error Rate of Streaming Codes with a given finite memory length.

  5. Existing Results for Streaming Codes  Adversarial Channel Model: Optimal streaming code rate and code construction given a deterministic set of possible channel error patterns  Burst Channel Model [Martinian and Sundberg 2004] [Khisti and Singh 2009]  Burs t and Arbitrary Erasure Channel Model [Fong et al. 2019] [Krishnan et al. 2018] [Badr et al. 2017]  Variable-Size Arrivals [Rudow and Rashmi 2018]  Stochastic Channel Model: Error exponent analysis and finite-memor y code construction [Draper and Khisti 2011] Draper and Khisti 2011 Our Work Deadline Constraint (Δ) Finite Infinite Memory Length (𝛽) 𝛽 ≥ Δ Arbitrary Error Probability Error Exponent Analysis Exact Error Rate Analysis

  6. Error Exponent Analysis for Convolutional-based Codes  Continue from previous slide Draper and Khisti 2011 Our Work Deadline Constraint (Δ) Finite Infinite Memory Length (𝛽) 𝛽 ≥ Δ Arbitrary Error Probability Error Exponent Analysis Exact Error Rate Analysis  Exponentially tight is not tight enough. Draper et al. 2011; Viterbi 1967 Our Work Asymptotic analysis, Exponential in addition to exact error rate analysis

  7. Outline  Motivation and Related Work  System Model  Main Contributions  Information-Debt Under Finite Memory  Exact Error Rate Analysis  A Provably-Tight Closed-Form Error Rate Approximation  Numerical Verification and Conclusion

  8. Slotted Coding System  𝛽 = 2 : In every time slot 𝑢 ≥ 1 ,  Encoder:  Receives 𝐿 packets:  𝑡 𝑙 (𝑢) in GF(2 𝑟 ) .  Stores 𝛽𝐿 packets in the previous 𝛽 slots  𝛽 is the memory length .  Encodes (𝛽 + 1)𝐿 packets and outputs 𝑂 coded packets:  Linear encoder : Define 𝐇 𝑢 as the 𝑂 -by- (min 𝛽 + 1, 𝑢 𝐿) generator matrix, we have  Random linear streaming codes (RLSCs): each entry of 𝐇 𝑢 is chosen uniformly randomly from GF(2 𝑟 ) , excluding 0.  Cumulative generator matrix:

  9. Comparison to [Martinian 2004] Coding System  Our work: Finite memory  Martinian setting: Infinite memory 𝛽 = 2 finite 𝛽 𝛽 = ∞

  10. Slotted Coding System  𝛽 = 2 : In every time slot 𝑢 ≥ 1 ,  Packet Erasure Channel:  Only a random subset of 𝑂 coded packets, denoted by , will arrive at the decoder perfectly.  is i.i.d. across 𝑢 . Define  is the probability of receiving 𝑗 packets successfully.

  11. Slotted Coding System  𝛽 = 2 : In every time slot 𝑢 ≥ 1 ,  Packet Erasure Channel:  Only a random subset of 𝑂 coded packets, denoted by , will arrive at the decoder perfectly.  is i.i.d. across 𝑢 . Define  is the probability of receiving 𝑗 packets successfully.  Received Signal:  The received packets:  Denote 𝐈 𝑢 the projection of 𝐇 𝑢 onto the random set  Cumulative receiver matrix:

  12. Slotted Coding System Random matrix depends on channel realization Definition 1. The vector 𝐭(𝑢) is decodable by time 𝑢 + Δ if all 𝑡 𝑙 𝑢 ∶ 𝑙 ∈ 1, 𝐿 are decodable by time 𝑢 + Δ . 𝑢+Δ , we  With optimal decoder on the received 𝐳 1 aim to solve the following:  Objective: Given any finite 𝑂 , 𝐿 , 𝛽 and 𝑄 𝑗 , Slot Error Rate Average Error Rate Infinite Deadline

  13. Technical Assumptions  Less-than-Capacity ( LC ) condition: Assume  Each slot: 𝑂 pkts 𝐿 pkts 𝐷 𝑢 pkts Encoder Channel  Generalized MDS Condition:  𝐇 (𝑢) : as full rank as possible (details in the paper)  MDS holds when 𝑟 → ∞  See Schwartz-Zippel Theorem  in [Ho et al. 2006, Theorems 3 and 4]  Avoid corner cases in the analysis

  14. Outline  Motivation and Related Work  System Model  Main Contributions  Information-Debt Under Finite Memory  Exact Error Rate Analysis  A Provably-Tight Closed-Form Error Rate Approximation  Numerical Verification and Conclusion

  15. Information-Debt Under Infinite Memory  Mutual information debt under infinite memory 𝛽 = ∞ [Martinian 2004] Definition. Initialize . For any , we iteratively comput e Debt is Nonnegative

  16. Information-Debt Under Infinite Memory  Mutual information debt under infinite memory 𝛽 = ∞ [Martinian 2004] 3 1 Definition. Initialize . For any , 3 2 3 we iteratively comput e 7 Observation: wherever 𝐽 𝑒 (𝑢) hits 0, we can decode 𝐭(𝑢) backwards. 𝐿 < 𝐷 𝑢 : decrease 𝐿 > 𝐷 𝑢 : increase

  17. Information-Debt Under Infinite Memory  Mutual information debt under infinite memory 𝛽 = ∞ [Martinian 2004] Definition. Initialize . For any , we iteratively comput e Observation: wherever 𝐽 𝑒 (𝑢) hits 0, we can decode 𝐭(𝑢) backwards. Q: What if 𝐈 (𝑢+Δ) is NOT full triangular? E.g. a 4-by-4 matrix which is not full rank

  18. Information-Debt Under Finite Memory  New information debt definition under finite memory 𝛽 < ∞ Definition 2. Define a constant and initialize _ . For any , we iteratively comput e  Absolute “ceiling” ∵ Memory length 𝛽  Bankruptcy ∴ Maximum allowable debt one can carry forward is at most 𝛽𝐿 Bankruptcy 𝛽𝐿 Maximum Allowable Debt

  19. Decodability Events Define , , and Proposition 1. For any fixed 𝑗 0 ≥ 0 , a) No 𝜐 𝑘 ∈ 𝑢 𝑗 0 , 𝑢 𝑗 0 +1 , then 𝐭(𝑢) is decodable by time 𝑢 𝑗 0 +1 for all 𝑢 ∈ (𝑢 𝑗 0 , 𝑢 𝑗 0 +1 ] . b) Exists 𝜐 𝑘 ∈ 𝑢 𝑗 0 , 𝑢 𝑗 0 +1 , define 𝜐 𝑘 ∗ the one with the as the 𝑗 -th time that 𝐽 𝑒 𝑢 hits 0 and , largest 𝑘 . Then 𝐭(𝑢) is decodable by time 𝑢 𝑗 0 +1 for all respectively. 𝑢 ∈ (𝜐 𝑘 ∗ − 𝛽 , 𝑢 𝑗 0 +1 ] . 𝐿 = 1, 𝛽 = 3 𝛽 − 1 o o o o o o o o o o o o o Decodable Decodable

  20. Error Events Define , , and Proposition 2. None of 𝐭 𝑢 : 𝑢 ∈ (𝑢 𝑗 0 , 𝜐 𝑘 ∗ − 𝛽] is decodable by time 𝜐 𝑘 ∗ − 𝛽 + Δ , regardless how large we set the deadline Δ . as the 𝑗 -th time that 𝐽 𝑒 𝑢 hits 0 and , respectively. 𝐿 = 1, 𝛽 = 3 𝛽 − 1 o o o o o o o x o o o x x x x x o o o x x Decodable Error Decodable

  21. Intuition Behind Define , , and  Enough linear equations  Start decoding from 𝐭(𝑢 𝑗 0 +1 ) , 𝐭(𝑢 𝑗 0 +1 − 1) , ⋯ , in a backward fashion as the 𝑗 -th time that 𝐽 𝑒 𝑢 hits 0 and , respectively. 𝐿 = 1, 𝛽 = 3 𝛽 − 1 o o o o o o o x o o o x x x x x o o o x x Decodable Error Decodable

  22. Intuition Behind Define , , and  Coupling between 𝐭 𝑢 : 𝑢 ≤ 𝜐 𝑘 ∗ − 𝛽 and 𝐭 𝑢 : 𝑢 > 𝜐 𝑘 ∗ − 𝛽 is severed once 𝐽 𝑒 (𝑢) hits (bankruptcy) as the 𝑗 -th time that 𝐽 𝑒 𝑢 hits 0 and , respectively. 𝐿 = 1, 𝛽 = 3 𝛽 − 1 o o o o o o o x o o o x x x x x o o o x x Decodable Error Decodable

  23. Exact Error Rate Analysis  𝐷 𝑢 is i.i.d. ⟹ 𝐽 𝑒 𝑢 is a renewal Markov process.  Information debt:  The state space: .  With 𝐽 𝑒 (𝑢) being renewal Markov process , the long term average error rate can be computed by Lemma 2. Assuming the LC and MDS conditions, we have  Not Stopping Time  More involved analysis  𝐽 𝑒 𝑢 from state-0 to state-0  Stopping Time for any fixed 𝑗 0 , where is the indicator function.

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend