today proof of converse coding theorem
play

Today Proof of Converse Coding Theorem Intuition: For message m , - PDF document

Today Proof of Converse Coding Theorem Intuition: For message m , let S m { 0 , 1 } n More on Shannons theory be the set of received words that decode to Proof of converse. m. ( S m = D 1 ( m ) ). Few words on


  1. Today Proof of Converse Coding Theorem • Intuition: For message m , let S m ⊆ { 0 , 1 } n • More on Shannon’s theory be the set of received words that decode to − Proof of converse. m. ( S m = D − 1 ( m ) ). − Few words on generality. − Contrast with Hamming theory. • Average size of D ( m ) = 2 n − k . • Back to error-correcting codes: Goals. • Volume of disc of radius pn around E ( m ) is 2 H ( p ) n . • Tools: • Intuition: If volume ≫ 2 n − k can’t have this − Probability theory: − Algebra: Finite fields, Linear spaces. ball decoding to m — but we need to! • Formalize? � Madhu Sudan, Fall 2004: Essential Coding Theory: MIT 6.895 c 1 � Madhu Sudan, Fall 2004: Essential Coding Theory: MIT 6.895 c 2 Proof of Converse Coding Theorem (contd.) Prob. [correct decoding ] � η ∈ { 0 , 1 } n � = Pr[ m sent , η error and m ∈{ 0 , 1 } k 2 − k · 2 � � � ≤ Pr[ η error] + m η ∈ B ( p ′ n,n ) η �∈ B ( p ′ n,n ) exp( − n ) + 2 − k − H ( p ′ ) n · � ≤ I m,η m,η Let I m,η be the indicator variable that is 1 iff exp( − n ) + 2 − k − H ( p ′ ) n · 2 n D (( E ( m ) + η )) = m . = ≤ exp( − n ) Let p ′ < p be such that R > 1 − H ( p ′ ) . � Madhu Sudan, Fall 2004: Essential Coding Theory: MIT 6.895 c 3 � Madhu Sudan, Fall 2004: Essential Coding Theory: MIT 6.895 c 4

  2. Generalizations of Shannon’s theorem Some of the main contributions • Channels more general • Rigorous Definition of elusive concepts: Information, Randomness. − Input symbols Σ , Output symbols Γ , where both may be infinite • Mathematical tools: Entropy, Mutual (reals/complexes). information, Relative entropy. − Channel given by its probability transition matrix P = P σ,γ . • Theorems: Coding theorem, converse. − Channel need not be independent - could be Markovian (remembers finite amount • Emphasis on the “feasible” as opposed to of state in determining next error bit). “done”. • In almost all cases: random coding + mld works. • Always non-constructive. � Madhu Sudan, Fall 2004: Essential Coding Theory: MIT 6.895 c 5 � Madhu Sudan, Fall 2004: Essential Coding Theory: MIT 6.895 c 6 Contrast between Hamming and • Most important difference: modelling of Shannon error — adversarial vs. probabilistic. Accounts for the huge difference in our ability to analyze one while having gaps in the other. • Works intertwined in time. • Nevertheless good to build Hamming like • Hamming’s work focusses on distance, and codes, even when trying to solve the image of E . Shannon problem. • Shannon’s work focusses on probabilities only (no mention of distance) and E, D but not properties of image of E . • Hamming’s results more constructive, definitions less so. • Shannon’s results not constructive, though definitions beg constructivitty. � Madhu Sudan, Fall 2004: Essential Coding Theory: MIT 6.895 c 7 � Madhu Sudan, Fall 2004: Essential Coding Theory: MIT 6.895 c 8

  3. Our focus • Will combine with analysis of encoding complexity and decoding complexity. • Codes, and associated encoding and decoding functions. • Distance is not the only measure, but we will say what we can about it. • Code parameters: n, k, d, q ; • typical goal: given three optimize fourth. • Coarser goal: consider only R = k/n , δ = d/n and q and given two, optimize the third. • In particular, can we get R, δ > 0 for constant q ? � Madhu Sudan, Fall 2004: Essential Coding Theory: MIT 6.895 c 9 � Madhu Sudan, Fall 2004: Essential Coding Theory: MIT 6.895 c 10 Tools Finite fields and linear error-correcting codes • Probability tools: • Field: algebraic structure with addition, − Linearity of expections, Union bound. multiplication, both commutative and − Expectation of product of independent associative with inverses, and multiplication] r.v.s distributive over addition. − Tail inqualities: Markov, Chebychev, Chernoff. • Finite field: Number of elements finite. Well known fact: field exists iff size is a • Algebra prime power. See lecture notes on algebra − Finite fields. for further details. Denote field of size q by − Vector spaces over finite fields. F q . • Elementary combinatorics and algorithmics. • Vector spaces: V defined over a field F . Addition of vectors, multiplication of vector with “scalar” (i.e., field element) is defined, � Madhu Sudan, Fall 2004: Essential Coding Theory: MIT 6.895 c 11 � Madhu Sudan, Fall 2004: Essential Coding Theory: MIT 6.895 c 12

  4. Why study this category? and finally an inner product (product of two vectors yielding a scalar is defined). • If alphabet is a field, then ambient space • Linear codes are the most common. Σ n becomes a vector space F n q . • Seem to be as strong as general ones. • If a code forms a vector space within F n q then it is a linear code. Denoted [ n, k, d ] q • Have succinct specification, efficient code. encoding and efficient error-detecting algorithms. Why? (Generator matrix and Parity check matrix.) • Linear algebra provides other useful tools: Duals of codes provide interesting constructions. • Dual of linear code is code generated by transpose of parity check matrix. � Madhu Sudan, Fall 2004: Essential Coding Theory: MIT 6.895 c 13 � Madhu Sudan, Fall 2004: Essential Coding Theory: MIT 6.895 c 14 Example: Dual of Hamming codes the identity matrix. Such matrices are called Hadamard matrices, and hence the code is called a Hadamard code.) • Message m = � m 1 , . . . , m ℓ � . • Moral of the story: Duals of good codes • Encoding given by �� m , x �� x ∈ F ℓ 2 − 0 . end up being good. No proven reason. • Fact: (will prove later): m � = 0 implies Pr x [ �� m , x � = 0] = 1 2 • Implies dual of [2 ℓ − 1 , 2 ℓ − ℓ − 1 , 3] 2 Hamming code is a [2 ℓ − 1 , ℓ, 2 ℓ − 1 ] code. • Often called the simplex code or the Hadamard code. (If we add a coordinate that is zero to all coordinates, and write 0 s as − 1 s, then the matrix whose rows are all the codewords form a +1 / − 1 matrix whose product with its transpose is a multiple of � Madhu Sudan, Fall 2004: Essential Coding Theory: MIT 6.895 c 15 � Madhu Sudan, Fall 2004: Essential Coding Theory: MIT 6.895 c 16

  5. Next few lectures • Towards asymptotically good codes: − Some good codes that are not asymptotically good. − Some compositions that lead to good codes. � Madhu Sudan, Fall 2004: Essential Coding Theory: MIT 6.895 c 17

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend