Whats New and Exciting in Algebraic and Combinatorial Coding - - PowerPoint PPT Presentation

what s new and exciting in algebraic and combinatorial
SMART_READER_LITE
LIVE PREVIEW

Whats New and Exciting in Algebraic and Combinatorial Coding - - PowerPoint PPT Presentation

Whats New and Exciting in Algebraic and Combinatorial Coding Theory? Alexander Vardy University of California San Diego vardy@kilimanjaro.ucsd.edu Notice: Persons attempting to find anything useful in this talk will be pro- secuted; persons


slide-1
SLIDE 1

What’s New and Exciting in Algebraic and Combinatorial Coding Theory?

Alexander Vardy

University of California San Diego

vardy@kilimanjaro.ucsd.edu

Notice: Persons attempting to find anything useful in this talk will be pro-

secuted; persons attempting to find anything original in it will be banished; persons attempting to draw conclusions from it will be shot.

— with apologies to Samuel L. Clemens

slide-2
SLIDE 2

The story of coding theory

Claude E. Shannon (1916–2001) A mathematical theory of commu- nication, Bell Systems Tech. Journal, 27, pp. 623–656, October 1948.

slide-3
SLIDE 3

The story of coding theory

Claude E. Shannon (1916–2001) A mathematical theory of commu- nication, Bell Systems Tech. Journal, 27, pp. 623–656, October 1948.

Shannon’s promise: For every channel, there exist error-correcting codes of rate up to the channel ca- pacity that achieve prob- ability of error as small as we please!

slide-4
SLIDE 4

The story of coding theory

Claude E. Shannon (1916–2001) A mathematical theory of commu- nication, Bell Systems Tech. Journal, 27, pp. 623–656, October 1948.

Shannon’s promise: For every channel, there exist error-correcting codes of rate up to the channel ca- pacity that achieve prob- ability of error as small as we please! Shannon’s puzzle How could we find such codes? How could we ef- ficiently decode them?

slide-5
SLIDE 5

The story of coding theory

Claude E. Shannon (1916–2001) A mathematical theory of commu- nication, Bell Systems Tech. Journal, 27, pp. 623–656, October 1948.

Shannon’s promise: For every channel, there exist error-correcting codes of rate up to the channel ca- pacity that achieve prob- ability of error as small as we please!

Algebraic and combinatorial coding theory 1948 —

Shannon’s puzzle How could we find such codes? How could we ef- ficiently decode them?

slide-6
SLIDE 6

The story of coding theory

Claude E. Shannon (1916–2001) A mathematical theory of commu- nication, Bell Systems Tech. Journal, 27, pp. 623–656, October 1948.

Shannon’s promise: For every channel, there exist error-correcting codes of rate up to the channel ca- pacity that achieve prob- ability of error as small as we please!

Algebraic and combinatorial coding theory 1948 —

Shannon’s puzzle How could we find such codes? How could we ef- ficiently decode them?

New, probabilistic coding theory 1994 —

slide-7
SLIDE 7

Magic solution to Shannon’s puzzle

Turbo codes and LDPC codes: codes defined on graphs with iterative decoding algorithms!

Random enough to come very close to capacity, but constructive enough to be iteratively decodable in poly- nomial (in fact, linear!) time.

During the last ten years:

Enormous amount of research — thousands of papers! Usually probabilistic rather than algebraic or combinatorial tools

✇Probabilistic coding theory

Powerful design methods available: density evolution, EXIT charts

slide-8
SLIDE 8

Progress toward the Shannon limit

The original turbo codes: about 0.7 dB from capacity

  • C. Berrou, A. Glavieux, and P. Thitimajshima, Near Shannon limit error-correcting

coding and decoding: Turbo codes, IEEE Int. Communications Conference, 1993.

slide-9
SLIDE 9

Progress toward the Shannon limit

The original turbo codes: about 0.7 dB from capacity

  • C. Berrou, A. Glavieux, and P. Thitimajshima, Near Shannon limit error-correcting

coding and decoding: Turbo codes, IEEE Int. Communications Conference, 1993.

Irregular LDPC codes: about 0.1 dB from capacity

T.J. Richardson and R. Urbanke, The capacity of low-density parity-check codes, IEEE Transactions on Information Theory, February 2001.

slide-10
SLIDE 10

Progress toward the Shannon limit

The original turbo codes: about 0.7 dB from capacity

  • C. Berrou, A. Glavieux, and P. Thitimajshima, Near Shannon limit error-correcting

coding and decoding: Turbo codes, IEEE Int. Communications Conference, 1993.

Irregular LDPC codes: about 0.1 dB from capacity

T.J. Richardson and R. Urbanke, The capacity of low-density parity-check codes, IEEE Transactions on Information Theory, February 2001.

How about 0.01 dB from capacity?

  • J. Boutros, G. Caire, E. Viterbo, H. Sawaya, and S. Vialle, Turbo code at 0.03 dB

from capacity limit, IEEE Symp. Inform. Theory, July 2002.

slide-11
SLIDE 11

Progress toward the Shannon limit

The original turbo codes: about 0.7 dB from capacity

  • C. Berrou, A. Glavieux, and P. Thitimajshima, Near Shannon limit error-correcting

coding and decoding: Turbo codes, IEEE Int. Communications Conference, 1993.

Irregular LDPC codes: about 0.1 dB from capacity

T.J. Richardson and R. Urbanke, The capacity of low-density parity-check codes, IEEE Transactions on Information Theory, February 2001.

How about 0.01 dB from capacity? And 0.001 dB?

  • J. Boutros, G. Caire, E. Viterbo, H. Sawaya, and S. Vialle, Turbo code at 0.03 dB

from capacity limit, IEEE Symp. Inform. Theory, July 2002. S-Y. Chung, G.D. Forney, Jr., T.J. Richardson, and R. Urbanke, On the design of low-density parity-check codes within 0.0045 dB of the Shannon limit, IEEE Communications Letters, February 2001.

slide-12
SLIDE 12

Progress toward the Shannon limit

The original turbo codes: about 0.7 dB from capacity

  • C. Berrou, A. Glavieux, and P. Thitimajshima, Near Shannon limit error-correcting

coding and decoding: Turbo codes, IEEE Int. Communications Conference, 1993.

Irregular LDPC codes: about 0.1 dB from capacity

T.J. Richardson and R. Urbanke, The capacity of low-density parity-check codes, IEEE Transactions on Information Theory, February 2001.

How about 0.01 dB from capacity? And 0.001 dB?

  • J. Boutros, G. Caire, E. Viterbo, H. Sawaya, and S. Vialle, Turbo code at 0.03 dB

from capacity limit, IEEE Symp. Inform. Theory, July 2002. S-Y. Chung, G.D. Forney, Jr., T.J. Richardson, and R. Urbanke, On the design of low-density parity-check codes within 0.0045 dB of the Shannon limit, IEEE Communications Letters, February 2001.

Conclusion: For all practical purposes, Shannon’s puzzle has

been now solved and Shannon’s promise has been achieved!

slide-13
SLIDE 13

Is coding theory dead?

slide-14
SLIDE 14

Is coding theory dead?

Last summer, I gave gave a lecture to non- specialists on the remarkable advances to- ward the Shannon limit made possible by the turbo/LDPC code revolution. At the end of the lecture, a member of the audi- ence asked, in effect, if this success meant that coding was dead. Then, I rather clum- sily ducked the question, but now I wish I had answered as follows:

Robert J. McEliece

slide-15
SLIDE 15

Is coding theory dead?

Last summer, I gave gave a lecture to non- specialists on the remarkable advances to- ward the Shannon limit made possible by the turbo/LDPC code revolution. At the end of the lecture, a member of the audi- ence asked, in effect, if this success meant that coding was dead. Then, I rather clum- sily ducked the question, but now I wish I had answered as follows:

Robert J. McEliece

As research topics, turbo codes and LDPC codes may be senior citizens by now, but other parts of coding are still quite youthful. Reed-Solomon codes, forever young, are a prime example.

slide-16
SLIDE 16

Outline: A few beautiful vistas

What is old and exciting in algebraic and combinatorial coding theory?

Asymptotic coding theory: the GV bound Perfect codes and the Delsarte conjecture Singleton bound and the MDS conjecture

Recent advances in algebraic list deco- ding of Reed-Solomon codes Applications of error-correcting codes in theoretical computer science Coding theory and networks: the new theory of network coding

slide-17
SLIDE 17

Outline: A few beautiful vistas

What is old and exciting in algebraic and combinatorial coding theory?

Asymptotic coding theory: the GV bound Perfect codes and the Delsarte conjecture Singleton bound and the MDS conjecture

Recent advances in algebraic list deco- ding of Reed-Solomon codes Applications of error-correcting codes in theoretical computer science Coding theory and networks: the new theory of network coding

slide-18
SLIDE 18

Outline: A few beautiful vistas

What is old and exciting in algebraic and combinatorial coding theory?

Asymptotic coding theory: the GV bound Perfect codes and the Delsarte conjecture Singleton bound and the MDS conjecture

Recent advances in algebraic list deco- ding of Reed-Solomon codes Applications of error-correcting codes in theoretical computer science Coding theory and networks: the new theory of network coding

slide-19
SLIDE 19

Outline: A few beautiful vistas

What is old and exciting in algebraic and combinatorial coding theory?

Asymptotic coding theory: the GV bound Perfect codes and the Delsarte conjecture Singleton bound and the MDS conjecture

Recent advances in algebraic list deco- ding of Reed-Solomon codes Applications of error-correcting codes in theoretical computer science Coding theory and networks: the new theory of network coding And a lot more that we shall skip...

slide-20
SLIDE 20

What’s Old and Exciting in Algebraic and Combinatorial Coding Theory?

Tale of Three Bounds and Three Conjectures

slide-21
SLIDE 21

The Hamming paradigm

Hamming distance: is a metric in Fn

q defined by

d(x, y) def

= # of positions where x and y differ

Minimum distance of a code:

d def

= min

x,y∈C d(x, y)

slide-22
SLIDE 22

The Hamming paradigm

Hamming distance: is a metric in Fn

q defined by

d(x, y) def

= # of positions where x and y differ

Minimum distance of a code:

d def

= min

x,y∈C d(x, y)

A code with minimum distance d can be re- garded as a packing of spheres of radius

t = ⌊(d−1)/2⌋

in the Hamming space Fn

q . Such a code cor-

rects any t adversarial errors. Thus coding theory may be thought of as a science of packing spheres densely and efficiently in metric spaces.

slide-23
SLIDE 23

What are the best codes?

Aq(n, d)

def

= the largest # of vectors of length n

  • ver an alphabet with q letters so that

any two of them are distance d apart

Vq(n, d)

def

= volume of the Hamming sphere

  • f radius d in the space of n-tuples
  • ver an alphabet with q letters
slide-24
SLIDE 24

What are the best codes?

Aq(n, d)

def

= the largest # of vectors of length n

  • ver an alphabet with q letters so that

any two of them are distance d apart

Vq(n, d)

def

= volume of the Hamming sphere

  • f radius d in the space of n-tuples
  • ver an alphabet with q letters

Theorem (Gilbert-Varshamov bound) Aq(n, d) qn Vq(n, d−1)

=

qn

d−1

i=0

n i

  • (q − 1)i

E.N. Gilbert, A comparison of signaling alphabets, Bell Systems Technical Journal, October 1952.

slide-25
SLIDE 25

Proof of the GV bound

Greedy construction algorithm. Take an arbitrary vector from the space, adjoin it to the code being constructed, and remove from the space the Hamming sphere or radius d − 1 around it. Repeat.

d 1 d 1 d 1

· · ·

d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1

· · ·

d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1

qn qn qn If after M steps there is nothing left, then the spheres of radius d − 1 about the M codewords cover the space, so M Vq(n, d−1) qn.

slide-26
SLIDE 26

Proof of the GV bound

Greedy construction algorithm. Take an arbitrary vector from the space, adjoin it to the code being constructed, and remove from the space the Hamming sphere or radius d − 1 around it. Repeat.

d 1 d 1 d 1

· · ·

d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1

· · ·

d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1 d 1

qn qn qn If after M steps there is nothing left, then the spheres of radius d − 1 about the M codewords cover the space, so M Vq(n, d−1) qn.

Open problem: Can we do better asymptotically?

slide-27
SLIDE 27

Asymptotic improvements of the GV bound

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

δ

R

Xing / Elkies Tsfasman−Vladuts−Zink Gilbert−Varshamov

Improving on the Gilbert-Var- shamov bound asymptotically is a notoriously difficult task!

M.A. Tsfasman, S.G. Vlˇ adu¸ t, and T. Zink, Modular curves, Shimura curves, and Goppa codes better than the Varshamov-Gilbert bound, Math. Nachrichten, 104, (1982), 13–28.

  • C. Xing, Nonlinear codes from algebraic curves improving the Tsfasman-Vlˇ

adu¸ t-Zink bound, IEEE Transactions on Information Theory, 49, (2003), 1653–1657.

  • N. Elkies, Still better codes from modular curves, preprint arXiv:math.NT/0308046, 2003.
slide-28
SLIDE 28

Asymptotic improvements of the GV bound

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

δ

R

Xing / Elkies Tsfasman−Vladuts−Zink Gilbert−Varshamov

Improving on the Gilbert-Var- shamov bound asymptotically is a notoriously difficult task!

For q 45, no asymptotic improvements of the GV bound are yet known...

M.A. Tsfasman, S.G. Vlˇ adu¸ t, and T. Zink, Modular curves, Shimura curves, and Goppa codes better than the Varshamov-Gilbert bound, Math. Nachrichten, 104, (1982), 13–28.

  • C. Xing, Nonlinear codes from algebraic curves improving the Tsfasman-Vlˇ

adu¸ t-Zink bound, IEEE Transactions on Information Theory, 49, (2003), 1653–1657.

  • N. Elkies, Still better codes from modular curves, preprint arXiv:math.NT/0308046, 2003.
slide-29
SLIDE 29

The GV bound for binary codes

For binary codes, the Gilbert-Varshamov bound takes the simple form:

A2(n, d) fGV(n, d) def

=

2n V2(n, d−1)

Since log2V2(n, d) H(d/n), this implies that for all n and all d n/2, there exist binary codes of rate R2(n, d) 1 − H(d/n).

A well-known conjecture The Gilbert-Varshamov bound is asympto- tically exact for binary codes.

V.D. Goppa, Bounds for codes, Doklady Academii Nauk, 1993, reports that he has a “wonderful proof” of this conjecture.

slide-30
SLIDE 30

The GV bound for binary codes

For binary codes, the Gilbert-Varshamov bound takes the simple form:

A2(n, d) fGV(n, d) def

=

2n V2(n, d−1)

Since log2V2(n, d) H(d/n), this implies that for all n and all d n/2, there exist binary codes of rate R2(n, d) 1 − H(d/n).

A well-known conjecture The Gilbert-Varshamov bound is asympto- tically exact for binary codes.

V.D. Goppa, Bounds for codes, Doklady Academii Nauk, 1993, reports that he has a “wonderful proof” of this conjecture.

Open problem: prove or disprove this conjecture!

slide-31
SLIDE 31

Recent improvement of GV bound

Theorem (Asymptotic improvement of Gilbert-Varshamov bound) Given positive integers n and d, with d n, let e(n, d) denote the fol- lowing quantity: e(n, d) def = 1 6

d

w=1 d

i=1 min{w,i}

j=⌈ w+i−d

2

⌉ n w w j n − w i − j

  • − 1

6

d

w=1

n w

  • Then:

A2(n, d) 2n V(n, d−1)

  • fGV(n, d)

· log2V(n, d−1) − log2

  • e(n, d−1)

10

  • improvement over the GV bound

by a factor linear in n

  • T. Jiang and A. Vardy, Asymptotic improvement of the Gilbert-Varshamov

bound on the size of binary codes, IEEE Trans. Inform. Theory, July 2004.

slide-32
SLIDE 32

Proof of the new bound

Definition (Gilbert graph) The Gilbert graph GG is defined as follows: V(GG) = all binary vectors of length n, and {x, y} ∈ E(GG) ⇐ ⇒ d(x, y) d − 1

slide-33
SLIDE 33

Proof of the new bound

Definition (Gilbert graph) The Gilbert graph GG is defined as follows: V(GG) = all binary vectors of length n, and {x, y} ∈ E(GG) ⇐ ⇒ d(x, y) d − 1 Then A2(n, d) is simply the independence number α(GG) of this graph.

slide-34
SLIDE 34

Proof of the new bound

Definition (Gilbert graph) The Gilbert graph GG is defined as follows: V(GG) = all binary vectors of length n, and {x, y} ∈ E(GG) ⇐ ⇒ d(x, y) d − 1 Then A2(n, d) is simply the independence number α(GG) of this graph. Theorem (Generalization of Ajtai, Komlós, and Szemerédi bound) For any ∆-regular graph with with at most T triangles, we have α(G) |V(G)| 10∆

  • log2 ∆ − 1

/

2 log2

  • T

|V(G)|

slide-35
SLIDE 35

Proof of the new bound

Definition (Gilbert graph) The Gilbert graph GG is defined as follows: V(GG) = all binary vectors of length n, and {x, y} ∈ E(GG) ⇐ ⇒ d(x, y) d − 1 Then A2(n, d) is simply the independence number α(GG) of this graph. Theorem (Generalization of Ajtai, Komlós, and Szemerédi bound) For any ∆-regular graph with with at most T triangles, we have α(G) |V(G)| 10∆

  • log2 ∆ − 1

/

2 log2

  • T

|V(G)| It remains to count the number of triangles in the Gilbert graph GG. This number is precisely the e(n, d) in the previous theorem.

slide-36
SLIDE 36

New bound versus old conjecture

How does this relate to the famous conjecture that the Gilbert- Varshamov bound is asymptotically exact for binary codes?

Conjecture A. The Gilbert-Varshamov bound on the size of

binary codes, namely A2(n, d), is asymptotically exact. That is lim

n→∞

A2(n, d) fGV(n, d) = const Our result implies that this is certainly false. The limit does not

  • exist. In fact, we prove that:

log2 A2(n, d) log2 fGV(n, d) + log n + const

Conjecture B. The Gilbert-Varshamov bound on the rate of

binary codes R2(n,d) = log2A2(n,d)/n is asymptotically exact.

This could still be true!

slide-37
SLIDE 37

New bound versus old conjecture

How does this relate to the famous conjecture that the Gilbert- Varshamov bound is asymptotically exact for binary codes?

Conjecture A. The Gilbert-Varshamov bound on the size of

binary codes, namely A2(n, d), is asymptotically exact. That is lim

n→∞

A2(n, d) fGV(n, d) = const Our result implies that this is certainly false. The limit does not

  • exist. In fact, we prove that:

log2 A2(n, d) log2 fGV(n, d) + log n + const

Conjecture B. The Gilbert-Varshamov bound on the rate of

binary codes R2(n,d) = log2A2(n,d)/n is asymptotically exact.

This could still be true!

slide-38
SLIDE 38

New bound versus old conjecture

How does this relate to the famous conjecture that the Gilbert- Varshamov bound is asymptotically exact for binary codes?

Conjecture A. The Gilbert-Varshamov bound on the size of

binary codes, namely A2(n, d), is asymptotically exact. That is lim

n→∞

A2(n, d) fGV(n, d) = const Our result implies that this is certainly false. The limit does not

  • exist. In fact, we prove that:

log2 A2(n, d) log2 fGV(n, d) + log n + const

Conjecture B. The Gilbert-Varshamov bound on the rate of

binary codes R2(n,d) = log2A2(n,d)/n is asymptotically exact.

This could still be true!

slide-39
SLIDE 39

Hamming bound and perfect codes

2n V2(n, 2e)

  • GV bound

A2(n, 2e+1)

2n V2(n, e)

  • Hamming

bound Definition (Perfect codes) Codes that attain the Hamming bound with equality are called perfect.

What perfect binary codes are out there?

Trivial codes: the whole space, any single codeword, the (n, 1, n) repetition code for all odd n Nontrivial codes: the (n, n − m, 3) Hamming codes for n = 2m− 1 and nonlinear perfect codes with the same parameters, the unique (23, 12, 7) binary Golay code

slide-40
SLIDE 40

Hamming bound and perfect codes

Definition (Perfect codes) Codes that attain the Hamming bound with equality are called perfect.

What perfect binary codes are out there?

Trivial codes: the whole space, any single codeword, the (n, 1, n) repetition code for all odd n Nontrivial codes: the (n, n − m, 3) Hamming codes for n = 2m− 1 and nonlinear perfect codes with the same parameters, the unique (23, 12, 7) binary Golay code Theorem (Complete characterization of perfect binary codes)

There are no more perfect binary codes!

— Van Lint, Tietäväinen, Zinoviev, and others, 1974

slide-41
SLIDE 41

Perfect codes in other metric spaces?

Instead of the Hamming space Fn

2 consider the Johnson space J(n, w) of

all binary vectors of length n and constant weight w. The sphere-pack- ing bound in this space is given by: A2(n, 4e+2, w) |J(n, w)| V2(n, e, w) = n w

  • e

i=0

w i n−w i

  • Definition (Perfect codes in the Johnson scheme)

Codes in J(n, w) that attain this bound with equality are called perfect.

What perfect codes are out there?

Trivial codes: the whole space, any single codeword, any pair of disjoint codewords for n = 2w, with w odd. Nontrivial codes: Are there any?

slide-42
SLIDE 42

The Delsarte Conjecture

After having recalled that there are “very few” perfect codes in the Hamming schemes, one must say that there is not a single one known in the Johnson schemes. It is temp- ting to risk the conjecture that such codes do not exist.

Philippe Delsarte, An algebraic approach to association schemes and coding theory, Philips Journal Research, October 1973.

Open problem: prove or disprove this conjecture!

slide-43
SLIDE 43

The Delsarte Conjecture

After having recalled that there are “very few” perfect codes in the Hamming schemes, one must say that there is not a single one known in the Johnson schemes. It is temp- ting to risk the conjecture that such codes do not exist.

Philippe Delsarte, An algebraic approach to association schemes and coding theory, Philips Journal Research, October 1973.

Open problem: prove or disprove this conjecture!

Biggs (1973), Bannai (1977), Hammond (1982) Roos (1983), Martin (1992), Etzion (1996) Ahlswede, Aydinian, and Khachatrian (2001) Etzion (2001), Etzion and Schwartz (2004) Shimabukuro (2005)

slide-44
SLIDE 44

Recent result on the Delsarte conjecture

Theorem

If there exists a prime p such that e ≡ −1 (mod p2), for example if e = 3, 7, 8, 11, 15, 17, 19 . . ., then there can be only finitely many non- trivial e-perfect codes in J(n, w). In particular, there are no nontrivial 3-perfect, 7-perfect, or 8-perfect codes in J(n, w) for all n and w.

  • T. Etzion and M. Schwartz, Perfect constant-weight codes,

IEEE Trans. Information Theory, September 2004.

  • Proof. Follows by showing that if there is an e-perfect code in J(n, w),

then the polynomial: Pn,w(X) def =

e

i=0

−1 i X + 1 i e−i

j=0

w−i j n − w + i − X−1 i + j

  • must have integer zeros ϕ in the range e <ϕ < w. Etzion and Schwartz

conjecture that Pn,w(X) does not have any integer zeros if e > 2.

slide-45
SLIDE 45

Singleton bound and MDS codes

Aq(n, d) qn−d+1

List of all the codewords

· · ·

} d − 1

all these columns are distinct

  • n − d + 1

Definition (MDS codes) Codes that attain the Singleton bound with equality are called max- imum distance separable or simply MDS.

slide-46
SLIDE 46

Singleton bound and MDS codes

Aq(n, d) qn−d+1

List of all the codewords

· · ·

} d − 1

all these columns are distinct

  • n − d + 1

Definition (MDS codes) Codes that attain the Singleton bound with equality are called max- imum distance separable or simply MDS.

What kind of MDS codes are out there?

A cyclic code of prime length p over GF(q) is MDS for almost all q A random code over GF(q) is MDS with probability → 1 as q → ∞ Reed-Solomon codes and generalized Reed-Solomon codes

slide-47
SLIDE 47

Some properties of MDS codes

MDS codes have many beautiful and useful properties:

Any k positions form an inform- ation set (linearly independent) Any d positions support a code- word of minimum weight The weight distribution of MDS codes is completely determined Trellis structure of MDS codes is also completely determined

But what about their length?

The MDS conjecture If C is an (n, k, d) MDS code with 1 < k < n − 1, then n q + 1 with two exceptions: the (q + 2, 3, q) code and its dual, if q = 2m.

slide-48
SLIDE 48

Some properties of MDS codes

MDS codes have many beautiful and useful properties:

Any k positions form an inform- ation set (linearly independent) Any d positions support a code- word of minimum weight The weight distribution of MDS codes is completely determined Trellis structure of MDS codes is also completely determined

But what about their length?

The MDS conjecture If C is an (n, k, d) MDS code with 1 < k < n − 1, then n q + 1 with two exceptions: the (q + 2, 3, q) code and its dual, if q = 2m.

slide-49
SLIDE 49

Some properties of MDS codes

MDS codes have many beautiful and useful properties:

Any k positions form an inform- ation set (linearly independent) Any d positions support a code- word of minimum weight The weight distribution of MDS codes is completely determined Trellis structure of MDS codes is also completely determined

But what about their length?

The MDS conjecture If C is an (n, k, d) MDS code with 1 < k < n − 1, then n q + 1 with two exceptions: the (q + 2, 3, q) code and its dual, if q = 2m.

slide-50
SLIDE 50

The MDS Conjecture

The MDS conjecture If C is an (n, k, d) MDS code with 1 < k < n − 1, then n q + 1 with two exceptions: the (q + 2, 3, q) code and its dual, if q = 2m. Equivalent conjecture: linear algebra Let V be a vector space over F

q and let S be a set of vectors of V

such that any k of them form a basis for V. Then |S| q + 1. Equivalent conjecture: matrix theory Let M be a k × m matrix over F

q such that every square subma-

trix of M is nonsingular. Then m + k q + 1. Equivalent conjecture: orthogonal arrays Let OA be an orthogonal array with q levels, strength k, and in-

slide-51
SLIDE 51

The MDS Conjecture

The MDS conjecture If C is an (n, k, d) MDS code with 1 < k < n − 1, then n q + 1 with two exceptions: the (q + 2, 3, q) code and its dual, if q = 2m. Equivalent conjecture: linear algebra Let V be a vector space over F

q and let S be a set of vectors of V

such that any k of them form a basis for V. Then |S| q + 1. Equivalent conjecture: matrix theory Let M be a k × m matrix over F

q such that every square subma-

trix of M is nonsingular. Then m + k q + 1. Equivalent conjecture: orthogonal arrays Let OA be an orthogonal array with q levels, strength k, and in-

slide-52
SLIDE 52

The MDS Conjecture

The MDS conjecture If C is an (n, k, d) MDS code with 1 < k < n − 1, then n q + 1 with two exceptions: the (q + 2, 3, q) code and its dual, if q = 2m. Equivalent conjecture: linear algebra Let V be a vector space over F

q and let S be a set of vectors of V

such that any k of them form a basis for V. Then |S| q + 1. Equivalent conjecture: matrix theory Let M be a k × m matrix over F

q such that every square subma-

trix of M is nonsingular. Then m + k q + 1. Equivalent conjecture: orthogonal arrays Let OA be an orthogonal array with q levels, strength k, and in-

slide-53
SLIDE 53

The MDS Conjecture

The MDS conjecture If C is an (n, k, d) MDS code with 1 < k < n − 1, then n q + 1 with two exceptions: the (q + 2, 3, q) code and its dual, if q = 2m. Equivalent conjecture: matrix theory Let M be a k × m matrix over F

q such that every square subma-

trix of M is nonsingular. Then m + k q + 1. Equivalent conjecture: orthogonal arrays Let OA be an orthogonal array with q levels, strength k, and in- dex one. Then the number of constraints in OA is at most q + 1. Equivalent conjecture: projective geometry Let A be an arc in the projective geometry PG(k−1, q). Then the

slide-54
SLIDE 54

The MDS Conjecture

The MDS conjecture If C is an (n, k, d) MDS code with 1 < k < n − 1, then n q + 1 with two exceptions: the (q + 2, 3, q) code and its dual, if q = 2m. Equivalent conjecture: orthogonal arrays Let OA be an orthogonal array with q levels, strength k, and in- dex one. Then the number of constraints in OA is at most q + 1. Equivalent conjecture: projective geometry Let A be an arc in the projective geometry PG(k−1, q). Then the number of points of A is at most q + 1.

slide-55
SLIDE 55

The MDS Conjecture

The MDS conjecture If C is an (n, k, d) MDS code with 1 < k < n − 1, then n q + 1 with two exceptions: the (q + 2, 3, q) code and its dual, if q = 2m. Equivalent conjecture: projective geometry Let A be an arc in the projective geometry PG(k−1, q). Then the number of points of A is at most q + 1. All this is known to be true if q 27, if k 5, if √q > 4k−9 and q is odd, and in certain other cases.

Segre, Singleton, Casse, Hirchfeld, Roth, many others, 1955 —

slide-56
SLIDE 56

The MDS Conjecture

The MDS conjecture If C is an (n, k, d) MDS code with 1 < k < n − 1, then n q + 1 with two exceptions: the (q + 2, 3, q) code and its dual, if q = 2m. Equivalent conjecture: projective geometry Let A be an arc in the projective geometry PG(k−1, q). Then the number of points of A is at most q + 1. All this is known to be true if q 27, if k 5, if √q > 4k−9 and q is odd, and in certain other cases.

Segre, Singleton, Casse, Hirchfeld, Roth, many others, 1955 —

Open problem: prove or disprove any of these conjectures!

slide-57
SLIDE 57
slide-58
SLIDE 58

Information Theory and Applications (ITA)

University of California San Diego

slide-59
SLIDE 59

Information Theory and Applications (ITA) Center at UCSD will offer prizes for the solution of impor- tant problems in information theory:

Information Theory and Applications (ITA)

University of California San Diego

slide-60
SLIDE 60

Information Theory and Applications (ITA) Center at UCSD will offer prizes for the solution of impor- tant problems in information theory: The asymptotic GV conjecture $1,000 The Delsarte conjecture $1,000 The MDS conjecture $1,000

details soon at http://ita.ucsd.edu Information Theory and Applications (ITA)

University of California San Diego

slide-61
SLIDE 61

Information Theory and Applications (ITA) Center at UCSD will offer prizes for the solution of impor- tant problems in information theory: The asymptotic GV conjecture $1,000 The Delsarte conjecture $1,000 The MDS conjecture $1,000

details soon at http://ita.ucsd.edu

Do these problems have potential applications in practice?

Information Theory and Applications (ITA)

University of California San Diego

slide-62
SLIDE 62

Information Theory and Applications (ITA) Center at UCSD will offer prizes for the solution of impor- tant problems in information theory: The asymptotic GV conjecture $1,000 The Delsarte conjecture $1,000 The MDS conjecture $1,000

details soon at http://ita.ucsd.edu

Do these problems have potential applications in practice?

There is much pleasure to be gained from useless knowledge!

— Bertrand Russel, 1912

Information Theory and Applications (ITA)

University of California San Diego

slide-63
SLIDE 63

Recent Advances in Algebraic List-Decoding of Reed-Solomon Codes

slide-64
SLIDE 64

The best algebraic codes

Millions of error-correcting codes are decoded every minute, with efficient algorithms implemented in custom VLSI circuits. At least 75% of these VLSI circuits decode Reed-Solomon codes.

I.S. Reed and G. Solomon, Polynomial codes over certain finite fields, Journal Society Indust. Appl. Math. 8, pp. 300-304, June 1960.

slide-65
SLIDE 65

Construction of Reed-Solomon codes

We describe the code via its encoder mapping E : Fk

q → Fn q . Fix in-

tegers k n q and n distinct elements x1, x2, . . . xn ∈F

  • q. Then

u0, u1, . . . , uk−1 k information symbols

fu(X) = u0 + u1X + · · · + uk−1Xk−1

c1 = fu(x1), c2 = fu(x2), · · · , cn = fu(xn)

(c1, c2, . . . , cn)

n codeword symbols Thus Reed-Solomon codes are linear. They have rate R = k/n and distance d = n − k + 1, which is the best possible (MDS).

slide-66
SLIDE 66

Algebraic decoding of Reed-Solomon codes

Every codeword of a Reed-Solomon code Cq(n, k) consists of some

n values of a polynomial f (X) of degree < k. This polynomial can

be uniquely recovered by interpolation from any k of its values. Thus a Reed-Solomon code Cq(n, k) can correct up to n−k erasures

  • r, equivalently, up to (n − k)/2 = (d − 1)/2 errors.

The Berlekamp-Massey algorithm is a very efficient way of doing

  • this. It has applications outside of coding theory as well.
slide-67
SLIDE 67

Algebraic decoding of Reed-Solomon codes

Every codeword of a Reed-Solomon code Cq(n, k) consists of some

n values of a polynomial f (X) of degree < k. This polynomial can

be uniquely recovered by interpolation from any k of its values. Thus a Reed-Solomon code Cq(n, k) can correct up to n−k erasures

  • r, equivalently, up to (n − k)/2 = (d − 1)/2 errors.

The Berlekamp-Massey algorithm is a very efficient way of doing

  • this. It has applications outside of coding theory as well.
slide-68
SLIDE 68

Algebraic decoding of Reed-Solomon codes

Every codeword of a Reed-Solomon code Cq(n, k) consists of some

n values of a polynomial f (X) of degree < k. This polynomial can

be uniquely recovered by interpolation from any k of its values. Thus a Reed-Solomon code Cq(n, k) can correct up to n−k erasures

  • r, equivalently, up to (n − k)/2 = (d − 1)/2 errors.

Error−locator polynomial

b b b b b b

n n−1 4 3 2 1

The Berlekamp-Massey algorithm is a very efficient way of doing

  • this. It has applications outside of coding theory as well.
slide-69
SLIDE 69

Algebraic decoding of Reed-Solomon codes

Every codeword of a Reed-Solomon code Cq(n, k) consists of some

n values of a polynomial f (X) of degree < k. This polynomial can

be uniquely recovered by interpolation from any k of its values. Thus a Reed-Solomon code Cq(n, k) can correct up to n−k erasures

  • r, equivalently, up to (n − k)/2 = (d − 1)/2 errors.

Error−locator polynomial

b b b b b b

n n−1 4 3 2 1

The Berlekamp-Massey algorithm is a very efficient way of doing

  • this. It has applications outside of coding theory as well.

Clearly, this is the best possible.

slide-70
SLIDE 70

Algebraic decoding of Reed-Solomon codes

Every codeword of a Reed-Solomon code Cq(n, k) consists of some

n values of a polynomial f (X) of degree < k. This polynomial can

be uniquely recovered by interpolation from any k of its values. Thus a Reed-Solomon code Cq(n, k) can correct up to n−k erasures

  • r, equivalently, up to (n − k)/2 = (d − 1)/2 errors.

Error−locator polynomial

b b b b b b

n n−1 4 3 2 1

The Berlekamp-Massey algorithm is a very efficient way of doing

  • this. It has applications outside of coding theory as well.

Clearly, this is the best possible. Or is it?

slide-71
SLIDE 71

Correcting more errors than thought possible

The 2002 Nevanlinna Prize went to M. Sudan with the citation “...in the theory of error-correcting codes, Sudan showed that certain coding methods could correct many more errors than was previously thought possible.”

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

Fraction of errors corrected Rate

Sudan Guruswami−Sudan Berlekamp−Massey

  • M. Sudan, Decoding of Reed-Solomon codes beyond the error

correction bound, Journal of Complexity, 1997.

  • V. Guruswami and M. Sudan, Improved decoding of Reed-Solomon and

algebraic-geometric codes, IEEE Trans. Information Theory, 1999.

slide-72
SLIDE 72

How does it work: the principle

Every codeword of the Reed-Solomon code Cq(n, k) corresponds to a polynomial. The unknown trans- mitted codeword can be represented by the algebraic curve Y − f (X) of degree at most k − 1.

Bézout’s Theorem

Two algebraic curves of degrees d and δ intersect in δd points, and cannot meet in more than δd points unless the equations defining them have a common factor.

  • E. Bézout, Théorie générale des équations algébriques, Paris, 1779.
slide-73
SLIDE 73

How does it work: the principle

Every codeword of the Reed-Solomon code Cq(n, k) corresponds to a polynomial. The unknown trans- mitted codeword can be represented by the algebraic curve Y − f (X) of degree at most k − 1.

Bézout’s Theorem

Two algebraic curves of degrees d and δ intersect in δd points, and cannot meet in more than δd points unless the equations defining them have a common factor.

  • E. Bézout, Théorie générale des équations algébriques, Paris, 1779.

Application of Bézout’s Theorem for decoding

If we could construct Q(X, Y) ∈ F

q[X, Y] which defines a curve of de-

gree δ that intersects Y − f (X) in more than (k−1)δ points (including points at ∞), then Y − f (X) can be recovered as a factor of Q(X, Y)!

slide-74
SLIDE 74

A couple of quotations

The real mathematics of the real mathematicians, the mathematics of Fermat and Euler and Gauss and Abel and Riemann and Bézout, is almost wholly useless.

— G.H. Hardy, A Mathematician’s Apology, 1941

This is right where I started being so favorably im- pressed: the KV algorithm is fully 2 dB better than what I’m using, and the advantage holds up over a wide range of SNRs and error rates. The use of your Reed-Solomon decoder in this program has been a spectacular success. Many dozens (perhaps hundreds?) of Earth-Moon-Earth contacts are be- ing made with it every day now, all over the world.

— Joseph H. Taylor, Nobel Laureate, 2004

slide-75
SLIDE 75

Reed-Solomon decoding: toy example

Suppose k = 2, so that the Reed-Solomon codewords are lines f (X) = aX + b. Given 14 interpolation points, we want to com- pute all lines passing through at least 5 of these points.

slide-76
SLIDE 76

Reed-Solomon decoding: toy example

Suppose k = 2, so that the Reed-Solomon codewords are lines f (X) = aX + b. Given 14 interpolation points, we want to com- pute all lines passing through at least 5 of these points. Compute a polynomial Q(X, Y) of degree < 5 such that Q(αi, βi) = 0 for all the 14 points:

Q(X, Y) = Y4 − X4 − Y2 + X2

slide-77
SLIDE 77

Reed-Solomon decoding: toy example

Suppose k = 2, so that the Reed-Solomon codewords are lines f (X) = aX + b. Given 14 interpolation points, we want to com- pute all lines passing through at least 5 of these points. Compute a polynomial Q(X, Y) of degree < 5 such that Q(αi, βi) = 0 for all the 14 points:

Q(X, Y) = Y4 − X4 − Y2 + X2

Let’s plot all the zeros of Q(X, Y). All the relevant lines now emerge!

slide-78
SLIDE 78

Reed-Solomon decoding: toy example

Suppose k = 2, so that the Reed-Solomon codewords are lines f (X) = aX + b. Given 14 interpolation points, we want to com- pute all lines passing through at least 5 of these points. Compute a polynomial Q(X, Y) of degree < 5 such that Q(αi, βi) = 0 for all the 14 points:

Q(X, Y) = Y4 − X4 − Y2 + X2

Let’s plot all the zeros of Q(X, Y). All the relevant lines now emerge! Formally, Q(X, Y) factors as:

(Y + X)(Y − X)(X2 + Y2 − 1)

Bézout’s Theorem says it must be so, since deg Q × deg f = 4 is strictly less than the number of intersection points, which is 5.

slide-79
SLIDE 79

Key decoding problems

Channel output ——

Multiplicity assignment

Assign interpolation weights M = [mi,j]

Polynomial interpolation

Interpolate through M = [mi,j] to compute Q(X, Y)

Partial factorization

Given Q(X, Y), find factors

  • Y− f (X)

|Q(X, Y)

slide-80
SLIDE 80

Key decoding problems

Channel output ——

Multiplicity assignment

Assign interpolation weights M = [mi,j]

Polynomial interpolation

Interpolate through M = [mi,j] to compute Q(X, Y)

Partial factorization

Given Q(X, Y), find factors

  • Y− f (X)

|Q(X, Y)

Decoder output = F(multiplicity assignment)       

Determines deco- der performance

                        

Determines deco- der complexity

slide-81
SLIDE 81

Key decoding problems

Channel output ——

Multiplicity assignment

Assign interpolation weights M = [mi,j]

Polynomial interpolation

Interpolate through M = [mi,j] to compute Q(X, Y)

Partial factorization

Given Q(X, Y), find factors

  • Y− f (X)

|Q(X, Y)

Decoder output = F(multiplicity assignment)       

Determines deco- der performance

                        

Determines deco- der complexity

slide-82
SLIDE 82

Algebraic soft-decision decoding

15 15.5 16 16.5 17 17.5 18 10

−6

10

−5

10

−4

10

−3

10

−2

10

−1

10 SNR [dB] Frame Error Rate Berlekamp−Welch Guruswami−Sudan (m = ∞) GMD Soft−decoding (L ≤ 32) Soft−decoding (L = ∞)

Soft−decision

1.5 dB

Hard−decision

A soft-decision decoder makes use of probabilis- tic information available at the output of almost every channel.

  • R. Koetter and A. Vardy, Algebraic soft-decision decoding of Reed-Solomon

codes, IEEE Transactions on Information Theory, 49, November 2003.

slide-83
SLIDE 83

Algebraic soft-decision decoding

15 15.5 16 16.5 17 17.5 18 10

−6

10

−5

10

−4

10

−3

10

−2

10

−1

10 SNR [dB] Frame Error Rate Berlekamp−Welch Guruswami−Sudan (m = ∞) GMD Soft−decoding (L ≤ 32) Soft−decoding (L = ∞)

Soft−decision

1.5 dB

Hard−decision

A soft-decision decoder makes use of probabilis- tic information available at the output of almost every channel. To decode, we’ll have to convert channel probab- ilities into algebraic in- terpolation conditions.

  • R. Koetter and A. Vardy, Algebraic soft-decision decoding of Reed-Solomon

codes, IEEE Transactions on Information Theory, 49, November 2003.

slide-84
SLIDE 84

Proportional multiplicity assignment

The # of intersections between Y − f (X) and Q(X, Y) is a random vari- able SM whose distribution depends on the channel observations and the multiplicity assignment. Which assignment maximizes its mean?

slide-85
SLIDE 85

Proportional multiplicity assignment

The # of intersections between Y − f (X) and Q(X, Y) is a random vari- able SM whose distribution depends on the channel observations and the multiplicity assignment. Which assignment maximizes its mean? Channel reliabilities Interpolation multiplicities R =    p1,1 · · · p1,n . . . ... . . . pq,1 · · · pq,n    ⇒ M =    m1,1 · · · m1,n . . . ... . . . mq,1 · · · mq,n    = ⌊λR⌋ with mi,j = ⌊λpi,j⌋

slide-86
SLIDE 86

Proportional multiplicity assignment

The # of intersections between Y − f (X) and Q(X, Y) is a random vari- able SM whose distribution depends on the channel observations and the multiplicity assignment. Which assignment maximizes its mean? Channel reliabilities Interpolation multiplicities R =    p1,1 · · · p1,n . . . ... . . . pq,1 · · · pq,n    ⇒ M =    m1,1 · · · m1,n . . . ... . . . mq,1 · · · mq,n    = ⌊λR⌋ with mi,j = ⌊λpi,j⌋ S

Score random variable

M

Probability density

assignment multiplicity Proportional

∆(Μ)

Theorem

The probability of decoding failure can be expressed as Pr{SM ∆(M)}. The pro- portional multiplicity assignment maxim- izes the mean of SM for a given ∆(M).

slide-87
SLIDE 87

Algebraic soft-decision decoder in real life

5.2 5.4 5.6 5.8 6 6.2 6.4 6.6 6.8 7 7.2 10 −10 10 −9 10 −8 10 −7 10 −6 10 −5 10 −4 10 −3 10 −2 10 −1 10 SNR [dB] Codeword error rate Berlekamp−Welch Sudan−Guruswami Koetter−Vardy Gaussian

Guruswami−Sudan SNR[dB] Codeword Error Rate Gaussian Approximation Proportional Assignment

Performance of the (468, 420, 49) Reed-Solomon code on a BPSK modulated AWGN channel

slide-88
SLIDE 88

Algebraic soft-decision decoder in real life

5.2 5.4 5.6 5.8 6 6.2 6.4 6.6 6.8 7 7.2 10 −10 10 −9 10 −8 10 −7 10 −6 10 −5 10 −4 10 −3 10 −2 10 −1 10 SNR [dB] Codeword error rate Berlekamp−Welch Sudan−Guruswami Koetter−Vardy Gaussian

Guruswami−Sudan SNR[dB] Codeword Error Rate Gaussian Approximation Proportional Assignment

Performance of the (468, 420, 49) Reed-Solomon code on a BPSK modulated AWGN channel

( ) ⋅ α ( ) ⋅ α ( ) ⋅ α

+

y ~

+

r [ ] a s x − ~

x ~

( ) , s q

( ) ,r s q

( ) 1 , s q

( ) 1 , s q

( ) 1 ,r s q

( ) 1 1 , s q

( ) r s q 0 ,

( ) r r s q ,

( ) r s q 1 ,

( ) b a d , ( ) b a d , 1 ( ) b a r d ,

⊕ ⊕

MACE0 HME [ ] b − 1 [ ] b r −

X

t [ ] b t −

+

D ( )( ) X b 0 ( )( ) X br ( )( ) X b 1 ( )( ) X br 1 ( )( ) X b 0 ( )( ) X br D D D

X

MACE1 MACEr 1 2 2

+

2 D D

D ( ) y s b a ~ , , , , c ( ) y s b a ~ , 1 , , , c ( ) y r s b a ~ , , , , c

The VLSI architecture is designed for data throughput of over 3.0 Gbps, at hardware cost of 3-4 times that of a conventional Berlekamp-Massey

  • decoder. Work in progress: VHDL description and ASIC design.
  • J. Ma, A. Vardy and Z. Wang, Efficient fast interpolation architecture for soft-

decision decoding of RS codes, IEEE Symp. Circuits and Systems, May 2006.

slide-89
SLIDE 89

Beyond the Guruswami-Sudan radius?

Since 1999, the Guruswami-Sudan decoding radius τGS = 1 − √ R was the best known. As of a few months ago, we can do much better!

Key idea: multivariate interpolation decoding

univariate interpolation

✇ ✇

interpolation decoding in three

  • r more dimensions

Key idea: new family of Reed-Solomon-like codes

1

Given information symbols u0, u1, . . . , uk−1, form the correspon- ding polynomial f (X) = u0 + u1X + · · · + uk−1Xk−1.

2

Compute g(X) :=

  • f (X)

a mod e(X), where e(X) is a fixed irre- ducible polynomial of degree k and a is an integer parameter.

3

Transmit the evaluation of f (X) +αg(X), where {1,α} is a basis for Fn

q2 over F q, namely

  • f (x1) +αg(x1), . . . , f (xn) +αg(xn)
  • .
slide-90
SLIDE 90

Beyond the Guruswami-Sudan radius?

Since 1999, the Guruswami-Sudan decoding radius τGS = 1 − √ R was the best known. As of a few months ago, we can do much better!

Key idea: multivariate interpolation decoding

univariate interpolation

✇ ✇

interpolation decoding in three

  • r more dimensions

Key idea: new family of Reed-Solomon-like codes

1

Given information symbols u0, u1, . . . , uk−1, form the correspon- ding polynomial f (X) = u0 + u1X + · · · + uk−1Xk−1.

2

Compute g(X) :=

  • f (X)

a mod e(X), where e(X) is a fixed irre- ducible polynomial of degree k and a is an integer parameter.

3

Transmit the evaluation of f (X) +αg(X), where {1,α} is a basis for Fn

q2 over F q, namely

  • f (x1) +αg(x1), . . . , f (xn) +αg(xn)
  • .
slide-91
SLIDE 91

Geometric interpretation of decoding

Given a received vector (y1 +αz1, y2 +αz2, . . . , yn +αzn), we interpo- late through the points (x1, y1, z1), (x2, y2, z2), . . . , (xn, yn, zn) to obtain a trivariate interpolation polynomial Q(X, Y, Z).

Interpolation Polynomial Transmitted Received word codeword

Interpolation polynomial Q

  • X, f (X), g(X)

≡ 0

slide-92
SLIDE 92

Geometric interpretation of decoding

Given a received vector (y1 +αz1, y2 +αz2, . . . , yn +αzn), we interpo- late through the points (x1, y1, z1), (x2, y2, z2), . . . , (xn, yn, zn) to obtain a trivariate interpolation polynomial Q(X, Y, Z).

Interpolation Polynomial Transmitted Received word codeword

Interpolation polynomial Q

  • X, f (X), g(X)

≡ 0

+

Transmitted codeword Encoder Polynomial

Encoder polynomial

  • f (X)

a mod e(X) ≡ g(X)

slide-93
SLIDE 93

Geometric interpretation of decoding

Given a received vector (y1 +αz1, y2 +αz2, . . . , yn +αzn), we interpo- late through the points (x1, y1, z1), (x2, y2, z2), . . . , (xn, yn, zn) to obtain a trivariate interpolation polynomial Q(X, Y, Z).

Encoder Polynomial Transmitted codeword Received word Interpolation Polynomial

Recovery of information

slide-94
SLIDE 94

Decoding radius of the new scheme

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

Fraction of errors corrected Rate

Sudan Guruswami−Sudan Berlekamp−Massey

Today: 1−R

1

  • F. Parvaresh and A. Vardy, Correct-

ing errors beyond the Guruswami- Sudan radius in polynomial time, IEEE Symp. Foundations of Computer Science (FOCS), October 2005.

2

  • V. Guruswami and A. Rudra, Expli-

cit capacity-achieving list-decodable codes, ACM Symposium on Theory of Computing (STOC), May 2006.

The main problem of list-decoding solved: We can construct the best possible codes in polynomial time and decode them in polynomial time!

Remaining problems: the alphabet-size is very large, the list-size in-

creases (polynomially) with the length of the code.

Conjecture: these methods achieve the capacity of the q-ary symmetric

channel, with better complexity than anything known.

slide-95
SLIDE 95

Decoding radius of the new scheme

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

Fraction of errors corrected Rate

Sudan Guruswami−Sudan Berlekamp−Massey

Today: 1−R

1

  • F. Parvaresh and A. Vardy, Correct-

ing errors beyond the Guruswami- Sudan radius in polynomial time, IEEE Symp. Foundations of Computer Science (FOCS), October 2005.

2

  • V. Guruswami and A. Rudra, Expli-

cit capacity-achieving list-decodable codes, ACM Symposium on Theory of Computing (STOC), May 2006.

The main problem of list-decoding solved: We can construct the best possible codes in polynomial time and decode them in polynomial time!

Remaining problems: the alphabet-size is very large, the list-size in-

creases (polynomially) with the length of the code.

Conjecture: these methods achieve the capacity of the q-ary symmetric

channel, with better complexity than anything known.

slide-96
SLIDE 96

Applications of Coding Theory in Theoretical Computer Science

slide-97
SLIDE 97

Codes most useful in computer science

Locally decodable codes

Codes with sub-linear time error-correction algorithms

Applications: private information retrieval, hardness amplification,

hard-core predicates, generation of pseudo-random bits

Locally testable codes

Codes with sub-linear time error-detection algorithms

Applications: at the core of probabilistically-checkable proofs theory

slide-98
SLIDE 98

Codes most useful in computer science

Locally decodable codes

Codes with sub-linear time error-correction algorithms

Applications: private information retrieval, hardness amplification,

hard-core predicates, generation of pseudo-random bits

Locally testable codes

Codes with sub-linear time error-detection algorithms

Applications: at the core of probabilistically-checkable proofs theory

How is decoding in sub-linear time possible?

Not enough time to read the input or write the output.

slide-99
SLIDE 99

Codes most useful in computer science

Locally decodable codes

Codes with sub-linear time error-correction algorithms

Applications: private information retrieval, hardness amplification,

hard-core predicates, generation of pseudo-random bits

Locally testable codes

Codes with sub-linear time error-detection algorithms

Applications: at the core of probabilistically-checkable proofs theory

How is decoding in sub-linear time possible?

Not enough time to read the input or write the output. We do the impossible for breakfast! — anonymous theoretical computer scientist, 2005

slide-100
SLIDE 100

Hardness amplification problem

How hard is it to compute a Boolean function f : {0, 1}n→{0, 1}?

Worst-case hardness: no polynomial-time algorithm can com-

pute f correctly on all possible inputs.

Example: NP-complete problems are believed to be worst-case hard.

Average-case hardness: no polynomial-time algorithm can

compute f correctly on a small fraction δ of the inputs.

Note: the smallest possible value is δ = 0.5 + ε, since random guess computes any f correctly on half the inputs in polynomial time. Average-case hard functions (δ-hard for some δ < 1) are needed in cryp- tography, pseudo-random generators, and many other applications!

Convert a worst-case hard function f into a δ-hard function g

slide-101
SLIDE 101

Hardness amplification problem

How hard is it to compute a Boolean function f : {0, 1}n→{0, 1}?

Worst-case hardness: no polynomial-time algorithm can com-

pute f correctly on all possible inputs.

Example: NP-complete problems are believed to be worst-case hard.

Average-case hardness: no polynomial-time algorithm can

compute f correctly on a small fraction δ of the inputs.

Note: the smallest possible value is δ = 0.5 + ε, since random guess computes any f correctly on half the inputs in polynomial time. Average-case hard functions (δ-hard for some δ < 1) are needed in cryp- tography, pseudo-random generators, and many other applications!

Convert a worst-case hard function f into a δ-hard function g

slide-102
SLIDE 102

Hardness amplification problem

How hard is it to compute a Boolean function f : {0, 1}n→{0, 1}?

Worst-case hardness: no polynomial-time algorithm can com-

pute f correctly on all possible inputs.

Example: NP-complete problems are believed to be worst-case hard.

Average-case hardness: no polynomial-time algorithm can

compute f correctly on a small fraction δ of the inputs.

Note: the smallest possible value is δ = 0.5 + ε, since random guess computes any f correctly on half the inputs in polynomial time. Average-case hard functions (δ-hard for some δ < 1) are needed in cryp- tography, pseudo-random generators, and many other applications!

Convert a worst-case hard function f into a δ-hard function g

slide-103
SLIDE 103

Hardness amplification using codes

0110111010001011001010010 0110111010001011001010010 00101110100110101000101100101001 01001101001100100100100000010001

f : {0, 1}n → {0, 1} Truth table of f Truth table of f [worst-case hard ] Truth table of g [δ-hard ] g: {0, 1}m → {0, 1} m = 1.1n Encoder for an (N, K, D) code C Decoder for the (N, K, D) code C K = 2n and N = 2m Polynomial-time algorithm that computes g correctly

  • n δ-fraction of inputs

Polynomial-time algorithm A that computes f correctly on all inputs

slide-104
SLIDE 104

Properties of the error-correcting code

Parameters of the code: Length is N = 2m and dimension is K = 2n.

If m = 1.1n, say, then the rate of C is given by K/N = 2−(m−n) = 2−0.1n. The rate of such codes is exponentially low!

Error-correction radius: For δ = 0.5 + ε, we need to recover from

0.5 −ε fraction of errors. This is possible only with list decoding.

Decoding complexity: Let A′ denote the hypothetical algorithm that

computes g correctly on δ-fraction of inputs. Then the computation of f (x) upon input x ∈ {0, 1}n proceeds as follows: x f (x)

A′

This computation is polynomial-time only if the decoder for C runs in time: polynomial in n = sub-linear in the code length N = 2m > 2n

slide-105
SLIDE 105

Properties of the error-correcting code

Parameters of the code: Length is N = 2m and dimension is K = 2n.

If m = 1.1n, say, then the rate of C is given by K/N = 2−(m−n) = 2−0.1n. The rate of such codes is exponentially low!

Error-correction radius: For δ = 0.5 + ε, we need to recover from

0.5 −ε fraction of errors. This is possible only with list decoding.

Decoding complexity: Let A′ denote the hypothetical algorithm that

computes g correctly on δ-fraction of inputs. Then the computation of f (x) upon input x ∈ {0, 1}n proceeds as follows: x f (x)

A′

This computation is polynomial-time only if the decoder for C runs in time: polynomial in n = sub-linear in the code length N = 2m > 2n

slide-106
SLIDE 106

Properties of the error-correcting code

Parameters of the code: Length is N = 2m and dimension is K = 2n.

If m = 1.1n, say, then the rate of C is given by K/N = 2−(m−n) = 2−0.1n. The rate of such codes is exponentially low!

Error-correction radius: For δ = 0.5 + ε, we need to recover from

0.5 −ε fraction of errors. This is possible only with list decoding.

Decoding complexity: Let A′ denote the hypothetical algorithm that

computes g correctly on δ-fraction of inputs. Then the computation of f (x) upon input x ∈ {0, 1}n proceeds as follows: x

Encoder Algorithm Decoder

f (x)

A′

This computation is polynomial-time only if the decoder for C runs in time: polynomial in n = sub-linear in the code length N = 2m > 2n

slide-107
SLIDE 107

Properties of the error-correcting code

Parameters of the code: Length is N = 2m and dimension is K = 2n.

If m = 1.1n, say, then the rate of C is given by K/N = 2−(m−n) = 2−0.1n. The rate of such codes is exponentially low!

Error-correction radius: For δ = 0.5 + ε, we need to recover from

0.5 −ε fraction of errors. This is possible only with list decoding.

Decoding complexity: Let A′ denote the hypothetical algorithm that

computes g correctly on δ-fraction of inputs. Then the computation of f (x) upon input x ∈ {0, 1}n proceeds as follows: x

Encoder Algorithm Decoder

f (x)

A′

This computation is polynomial-time only if the decoder for C runs in time: polynomial in n = sub-linear in the code length N = 2m > 2n

Observation: the decoder needs to produce only one bit!

slide-108
SLIDE 108

Locally decodable codes: definition

Definition (locally decodable codes)

A binary linear code C of length n and dimension k is (q, δ, p)-locally decodable if there is a decoder D for C with the following properties:

1

On input y ∈ {0, 1}n and i ∈ {1, 2, . . . , k}, the decoder reads q bits of y uniformly at random and produces a single bit D(y; i).

2

Let c be the encoding of a message u. Then for all inputs y that agree with c on at least δn positions, we have Pr D(y; i) = ui p This holds for all messages u ∈ {0, 1}k and indices i = 1, 2, . . . , k.

Open problem: given the parameters q, δ, and p,

what is the best possible tradeoff between n and k?

slide-109
SLIDE 109

Example: first-order RM code

Theorem

The (2k, k, 2k−1) binary first-order Reed-Muller code is (2, δ, 2δ−1)-loc- ally decodable, for all δ > 0.5.

Encoding: Given u ∈ Fk

2 , let fu(X) = u1X1 + u2X2 + · · · + ukXk. Then

u = (u1, u2, . . . , uk) → c = evaluation of fu(X) at all x ∈ Fk

2

Decoding: Given y = c + errors and i, choose x = (x1, x2, . . . , xk) uni-

formly at random, and read two entries of y corresponding to x and to x + 1i, where 1i = (0 · · · 010 · · · 0) with the 1 at the i-th position, namely a = fu(x) + error and b = fu(x + 1i) + error If there are no errors at the two positions, then the sum a + b produces:

  • u1x1 + u2x2 + · · · + ukxk

+

  • u1x1 + · · · + ui(xi + 1) + · · · + ukxk

= ui

Probabilistic analysis: Given that d(y, c) (1 − δ)n, it is very easy

to see that Pr{a in error} = Pr{b in error} 1 − δ, and we are done!

slide-110
SLIDE 110

Example: first-order RM code

Theorem

The (2k, k, 2k−1) binary first-order Reed-Muller code is (2, δ, 2δ−1)-loc- ally decodable, for all δ > 0.5.

Encoding: Given u ∈ Fk

2 , let fu(X) = u1X1 + u2X2 + · · · + ukXk. Then

u = (u1, u2, . . . , uk) → c = evaluation of fu(X) at all x ∈ Fk

2

Decoding: Given y = c + errors and i, choose x = (x1, x2, . . . , xk) uni-

formly at random, and read two entries of y corresponding to x and to x + 1i, where 1i = (0 · · · 010 · · · 0) with the 1 at the i-th position, namely a = fu(x) + error and b = fu(x + 1i) + error If there are no errors at the two positions, then the sum a + b produces:

  • u1x1 + u2x2 + · · · + ukxk

+

  • u1x1 + · · · + ui(xi + 1) + · · · + ukxk

= ui

Probabilistic analysis: Given that d(y, c) (1 − δ)n, it is very easy

to see that Pr{a in error} = Pr{b in error} 1 − δ, and we are done!

slide-111
SLIDE 111

Example: first-order RM code

Theorem

The (2k, k, 2k−1) binary first-order Reed-Muller code is (2, δ, 2δ−1)-loc- ally decodable, for all δ > 0.5.

Encoding: Given u ∈ Fk

2 , let fu(X) = u1X1 + u2X2 + · · · + ukXk. Then

u = (u1, u2, . . . , uk) → c = evaluation of fu(X) at all x ∈ Fk

2

Decoding: Given y = c + errors and i, choose x = (x1, x2, . . . , xk) uni-

formly at random, and read two entries of y corresponding to x and to x + 1i, where 1i = (0 · · · 010 · · · 0) with the 1 at the i-th position, namely a = fu(x) + error and b = fu(x + 1i) + error If there are no errors at the two positions, then the sum a + b produces:

  • u1x1 + u2x2 + · · · + ukxk

+

  • u1x1 + · · · + ui(xi + 1) + · · · + ukxk

= ui

Probabilistic analysis: Given that d(y, c) (1 − δ)n, it is very easy

to see that Pr{a in error} = Pr{b in error} 1 − δ, and we are done!

slide-112
SLIDE 112

Example: first-order RM code

Theorem

The (2k, k, 2k−1) binary first-order Reed-Muller code is (2, δ, 2δ−1)-loc- ally decodable, for all δ > 0.5.

Encoding: Given u ∈ Fk

2 , let fu(X) = u1X1 + u2X2 + · · · + ukXk. Then

u = (u1, u2, . . . , uk) → c = evaluation of fu(X) at all x ∈ Fk

2

Decoding: Given y = c + errors and i, choose x = (x1, x2, . . . , xk) uni-

formly at random, and read two entries of y corresponding to x and to x + 1i, where 1i = (0 · · · 010 · · · 0) with the 1 at the i-th position, namely a = fu(x) + error and b = fu(x + 1i) + error If there are no errors at the two positions, then the sum a + b produces:

  • u1x1 + u2x2 + · · · + ukxk

+

  • u1x1 + · · · + ui(xi + 1) + · · · + ukxk

= ui

Probabilistic analysis: Given that d(y, c) (1 − δ)n, it is very easy

to see that Pr{a in error} = Pr{b in error} 1 − δ, and we are done!

slide-113
SLIDE 113

Best-known locally decodable binary codes

# of queries Constructions Lower bounds q = 2 queries n = 2k n 2Ω(k) q = 3 queries n = 2

√ k

n = Ω(k2) . . . . . . . . . q = const queries n = 2 k

log log q q log q

n = Ω

  • k

q−1 q−2

  • Open problem

There are enormous gaps between known bounds and constructions!

Open problem

It is known that it is not possible to have n = O(k), namely codes of constant rate. Is it possible to have n = O(kc) for a constant c? That is, locally decodable codes with rate going to zero only polynomially fast?

slide-114
SLIDE 114

Conclusions

slide-115
SLIDE 115

Conclusions

slide-116
SLIDE 116

Conclusions

Past: For the past 20 years, coding theo-

ry has been an amazingly vibrant field. Working in this field was great fun.

slide-117
SLIDE 117

Conclusions

Past: For the past 20 years, coding theo-

ry has been an amazingly vibrant field. Working in this field was great fun.

Present: It has never been more fun than today!

slide-118
SLIDE 118

Conclusions

Past: For the past 20 years, coding theo-

ry has been an amazingly vibrant field. Working in this field was great fun.

Present: It has never been more fun than today!

slide-119
SLIDE 119

Conclusions

today!

slide-120
SLIDE 120

Conclusions

Past: For the past 20 years, coding theo-

ry has been an amazingly vibrant field. Working in this field was great fun.

Present: It has never been more fun than today! Future: I hope that it will stay this

way for the next twenty years.

slide-121
SLIDE 121

Conclusions

Past: For the past 20 years, coding theo-

ry has been an amazingly vibrant field. Working in this field was great fun.

Present: It has never been more fun than today! Future: I hope that it will stay this

way for the next twenty years.

This is up to you!