Bounds on the Rate of 2-D Bit-Stuffing Encoders Ido Tal Ron M. - - PowerPoint PPT Presentation

bounds on the rate of 2 d bit stuffing encoders
SMART_READER_LITE
LIVE PREVIEW

Bounds on the Rate of 2-D Bit-Stuffing Encoders Ido Tal Ron M. - - PowerPoint PPT Presentation

Introduction Three facts, and an assumption The bounds Quasi-Stationarity Bounds on the Rate of 2-D Bit-Stuffing Encoders Ido Tal Ron M. Roth Computer Science Department Technion, Haifa 32000, Israel Introduction Three facts, and an


slide-1
SLIDE 1

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Bounds on the Rate of 2-D Bit-Stuffing Encoders

Ido Tal Ron M. Roth

Computer Science Department Technion, Haifa 32000, Israel

slide-2
SLIDE 2

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

2-D constraints

The square constraint – our running example A binary M × N array satisfies the square constraint iff no two ‘1’ symbols are adjacent on a row, column, or diagonal. Example:

1 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0

If a bold-face 0 is changed to 1, then the square constraint does not hold.

Notation for the general case Let S be a constraint over an alphabet Σ. Denote by S ∩ ΣM×N all the M × N arrays satisfying the constraint.

slide-3
SLIDE 3

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Bit stuffing encoders

Encoder Definition E = (Ψ, µ, δ = (δM,N)M,N>0) . δM,N and ∂M,N ∂M,N = ∂M,N(E) is the border index set of the array we wish to encode into. ¯ ∂M,N is the complementary set.

slide-4
SLIDE 4

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Bit stuffing encoders

Encoder Definition E = (Ψ, µ, δ = (δM,N)M,N>0) . δM,N and ∂M,N ∂M,N = ∂M,N(E) is the border index set of the array we wish to encode into. ¯ ∂M,N is the complementary set. δM,N is a probability distribution on all valid borders, δM,N : S[∂M,N] → [0, 1] .

1 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1

slide-5
SLIDE 5

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Encoder

Ψ and µ Let σα,β be the index shifting operator: σα,β(U) = {(i + α, j + β) : (i, j) ∈ U} . Encoding into ¯ ∂M,N is done in raster fashion. When encoding to position (i, j) ∈ ¯ ∂M,N, we only look at positions σi,j(Ψ): the neighborhood of (i, j). The probability distribution of entry (i, j) is given by the function µ(·|·).

slide-6
SLIDE 6

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Encoder

Ψ and µ Let σα,β be the index shifting operator: σα,β(U) = {(i + α, j + β) : (i, j) ∈ U} . Encoding into ¯ ∂M,N is done in raster fashion. When encoding to position (i, j) ∈ ¯ ∂M,N, we only look at positions σi,j(Ψ): the neighborhood of (i, j). The probability distribution of entry (i, j) is given by the function µ(·|·).

  • 1 0 0 0 0 1 0 0

0 0 0 0 0 0 0 1 µ

  • = 0.258132
slide-7
SLIDE 7

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Encoder

Ψ and µ Let σα,β be the index shifting operator: σα,β(U) = {(i + α, j + β) : (i, j) ∈ U} . Encoding into ¯ ∂M,N is done in raster fashion. When encoding to position (i, j) ∈ ¯ ∂M,N, we only look at positions σi,j(Ψ): the neighborhood of (i, j). The probability distribution of entry (i, j) is given by the function µ(·|·).

1 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 µ

  • = 0.258132
slide-8
SLIDE 8

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Encoder

Ψ and µ Let σα,β be the index shifting operator: σα,β(U) = {(i + α, j + β) : (i, j) ∈ U} . Encoding into ¯ ∂M,N is done in raster fashion. When encoding to position (i, j) ∈ ¯ ∂M,N, we only look at positions σi,j(Ψ): the neighborhood of (i, j). The probability distribution of entry (i, j) is given by the function µ(·|·).

  • 1 0 0 0 0 1 0 0

0 0 0 0 0 0 0 1 µ

  • = 0.258132
slide-9
SLIDE 9

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Encoder

Ψ and µ Let σα,β be the index shifting operator: σα,β(U) = {(i + α, j + β) : (i, j) ∈ U} . Encoding into ¯ ∂M,N is done in raster fashion. When encoding to position (i, j) ∈ ¯ ∂M,N, we only look at positions σi,j(Ψ): the neighborhood of (i, j). The probability distribution of entry (i, j) is given by the function µ(·|·).

1 1 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 µ

  • = 0.258132
slide-10
SLIDE 10

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Encoder

Ψ and µ Let σα,β be the index shifting operator: σα,β(U) = {(i + α, j + β) : (i, j) ∈ U} . Encoding into ¯ ∂M,N is done in raster fashion. When encoding to position (i, j) ∈ ¯ ∂M,N, we only look at positions σi,j(Ψ): the neighborhood of (i, j). The probability distribution of entry (i, j) is given by the function µ(·|·).

  • 1 0 0 0 0 1 0 0

0 0 0 0 0 0 0 1 0 1 µ

  • 1

1

  • = 1
slide-11
SLIDE 11

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Encoder

Encoder? Q: So, why is this an “encoder”? A: The “coins” are in fact (invertible) probability transformers, the input of which is the information we wish to encode.

slide-12
SLIDE 12

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Encoder

Encoder? Q: So, why is this an “encoder”? A: The “coins” are in fact (invertible) probability transformers, the input of which is the information we wish to encode. Encoder rate Let A = A(E, M, N) be the random variable corresponding to the array we produce. The rate of our encoder is R(E) lim inf

M,N→∞

H(A[¯ ∂M,N]|A[∂M,N]) M · N . Problem: How does one calculate the rate. . .

slide-13
SLIDE 13

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

First fact: Locality of conditional entropy

Let Ti,j be all the indices preceding (i, j) in the raster scan. R(E) = lim inf

M,N→∞

  • (i,j)∈¯

∂M,N

H(ai,j|A[∂M,N] ∪ A[Ti,j]) M · N = lim inf

M,N→∞

  • (i,j)∈¯

∂M,N

H(ai,j|A[σi,j(Ψ)]) M · N

slide-14
SLIDE 14

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Second fact: If we know the border’s distribution, then we know the whole distribution

Consider a (relatively small) patch Λ with border Γ. If we know the probability distribution of A[Γ], then we know the probability distribution of A[Λ].

slide-15
SLIDE 15

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Third fact: Stationarity inside the patch

Let Γ′ be Γ, without the last column. We will prove later that w.l.o.g., the probability distributions of A[Γ′] is equal to the probability distribution of A[σ0,1(Γ′)].

slide-16
SLIDE 16

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Third fact: Stationarity inside the patch

Let Γ′ be Γ, without the last column. We will prove later that w.l.o.g., the probability distributions of A[Γ′] is equal to the probability distribution of A[σ0,1(Γ′)].

slide-17
SLIDE 17

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Third fact (assumption): Stationarity inside the patch

Let Γ′′ be Γ, without the last row. We will prove later that w.l.o.g., the probability distributions of A[Γ′′] is equal to the probability distribution of A[σ1,−1(Γ′′)].

slide-18
SLIDE 18

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Third fact (assumption): Stationarity inside the patch

Let Γ′′ be Γ, without the last row. We will prove later that w.l.o.g., the probability distributions of A[Γ′′] is equal to the probability distribution of A[σ1,−1(Γ′′)].

slide-19
SLIDE 19

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

The bound

Recall that R(E) = lim infM,N→∞

  • (i,j)∈¯

∂M,N H(ai,j|A[σi,j(Ψ)]) M·N

. Consider all patch border probabilities which result in a stationary patch. For each such probability, look at H(ai,j|σi,j(Ψ)). The smallest (largest) value is an lower (upper) bound on the rate of our encoder. The above minimization (maximization) problem is a linear

  • program. It gets more accurate, but harder, as we enlarge

the patch.

slide-20
SLIDE 20

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Some numerical results

Constraint Coins lp∗

min

lp∗

max

[Halevy+:04] (2, ∞)-RLL 1 0.440722 0.444679 0.4267 (3, ∞)-RLL 1 0.349086 0.386584 0.3402 n.i.b. 2 0.91773 0.919395 0.91276 (1, ∞)-RLL 3 0.587776 0.587785 —

slide-21
SLIDE 21

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Some more numerical results

Constraint Coins lp∗

min

lp∗

max

Others (2, ∞)-RLL 5 0.444202 0.444997 0.4423 (3, ∞)-RLL 2 0.359735 0.368964 0.3641 (0, 2)-RLL 66 0.815497 0.816821 0.7736 18 0.815013 0.816176 9 0.810738 0.819660 n.i.b. 56 0.922640 0.923748 0.9156

slide-22
SLIDE 22

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

A(k)

Define [Halevy+:04] a new random variable, A(k): Out of the k2 contiguous (M − k + 1) × (N − k + 1) sub-arrays of A , pick one uniformly at random, and call it A(k).

slide-23
SLIDE 23

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

A(k)

Define [Halevy+:04] a new random variable, A(k): Out of the k2 contiguous (M − k + 1) × (N − k + 1) sub-arrays of A , pick one uniformly at random, and call it A(k).

0 1 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 1 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 1 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 1 0 0 0 0 0 0

slide-24
SLIDE 24

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

A(k)

Define [Halevy+:04] a new random variable, A(k): Out of the k2 contiguous (M − k + 1) × (N − k + 1) sub-arrays of A , pick one uniformly at random, and call it A(k). k = 5:

0 1 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 1 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 1 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 1 0 0 0 0 0 0

slide-25
SLIDE 25

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

A(k)

Define [Halevy+:04] a new random variable, A(k): Out of the k2 contiguous (M − k + 1) × (N − k + 1) sub-arrays of A , pick one uniformly at random, and call it A(k). k = 5:

0 1 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 1 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 1 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 1 0 0 0 0 0 0

slide-26
SLIDE 26

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

A(k)

Define [Halevy+:04] a new random variable, A(k): Out of the k2 contiguous (M − k + 1) × (N − k + 1) sub-arrays of A , pick one uniformly at random, and call it A(k). k = 5:

0 1 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 1 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 1 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 1 0 0 0 0 0 0

slide-27
SLIDE 27

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

A(k)

Define [Halevy+:04] a new random variable, A(k): Out of the k2 contiguous (M − k + 1) × (N − k + 1) sub-arrays of A , pick one uniformly at random, and call it A(k). k = 5:

0 1 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 1 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 1 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 1 0 0 0 0 0 0

slide-28
SLIDE 28

Introduction Three facts, and an assumption The bounds Quasi-Stationarity

Merits of A(k)

Since A(k) is an “averaging out” of A, we have that as k grows, A(k) becomes more and more “locally stationary”. We can define an encoder E(k) = (Ψ, µ, δ(k)) with A(E(k)) having the same probability distribution as A(k). The rate of the encoders are the same, R(E) = R(E(k)). So, essentially, we can assume w.l.o.g. that the patch in A is stationary.

0 1 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 1 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 1 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 1 0 0 0 0 0 0