Two added structures in sparse recovery: nonnegativity and - - PowerPoint PPT Presentation

two added structures in sparse recovery nonnegativity and
SMART_READER_LITE
LIVE PREVIEW

Two added structures in sparse recovery: nonnegativity and - - PowerPoint PPT Presentation

Two added structures in sparse recovery: nonnegativity and disjointedness Simon Foucart University of Georgia Semester Program on High-Dimensional Approximation ICERM 7 October 2014 Part I: Nonnegative Sparse Recovery (joint work with


slide-1
SLIDE 1

Two added structures in sparse recovery: nonnegativity and disjointedness Simon Foucart

University of Georgia Semester Program on “High-Dimensional Approximation” ICERM 7 October 2014

slide-2
SLIDE 2

Part I: Nonnegative Sparse Recovery (joint work with D. Koslicki)

slide-3
SLIDE 3

Motivation from Metagenomics

slide-4
SLIDE 4

Motivation from Metagenomics

◮ x ∈ RN (N = 273, 727): concentrations of known bacteria in

a given environmental sample.

slide-5
SLIDE 5

Motivation from Metagenomics

◮ x ∈ RN (N = 273, 727): concentrations of known bacteria in

a given environmental sample. Sparsity assumption is realistic.

slide-6
SLIDE 6

Motivation from Metagenomics

◮ x ∈ RN (N = 273, 727): concentrations of known bacteria in

a given environmental sample. Sparsity assumption is realistic. Note also that x ≥ 0 and

j xj = 1.

slide-7
SLIDE 7

Motivation from Metagenomics

◮ x ∈ RN (N = 273, 727): concentrations of known bacteria in

a given environmental sample. Sparsity assumption is realistic. Note also that x ≥ 0 and

j xj = 1. ◮ y ∈ Rm (m = 46 = 4, 096): frequencies of length-6 subwords

(in 16S rRNA gene reads or in whole-genome shotgun reads)

slide-8
SLIDE 8

Motivation from Metagenomics

◮ x ∈ RN (N = 273, 727): concentrations of known bacteria in

a given environmental sample. Sparsity assumption is realistic. Note also that x ≥ 0 and

j xj = 1. ◮ y ∈ Rm (m = 46 = 4, 096): frequencies of length-6 subwords

(in 16S rRNA gene reads or in whole-genome shotgun reads)

◮ A ∈ Rm×N: frequencies of length-6 subwords in all known

(i.e., sequenced) bacteria.

slide-9
SLIDE 9

Motivation from Metagenomics

◮ x ∈ RN (N = 273, 727): concentrations of known bacteria in

a given environmental sample. Sparsity assumption is realistic. Note also that x ≥ 0 and

j xj = 1. ◮ y ∈ Rm (m = 46 = 4, 096): frequencies of length-6 subwords

(in 16S rRNA gene reads or in whole-genome shotgun reads)

◮ A ∈ Rm×N: frequencies of length-6 subwords in all known

(i.e., sequenced) bacteria. It is a frequency matrix, that is,

slide-10
SLIDE 10

Motivation from Metagenomics

◮ x ∈ RN (N = 273, 727): concentrations of known bacteria in

a given environmental sample. Sparsity assumption is realistic. Note also that x ≥ 0 and

j xj = 1. ◮ y ∈ Rm (m = 46 = 4, 096): frequencies of length-6 subwords

(in 16S rRNA gene reads or in whole-genome shotgun reads)

◮ A ∈ Rm×N: frequencies of length-6 subwords in all known

(i.e., sequenced) bacteria. It is a frequency matrix, that is, Ai,j ≥ 0 and m

i=1Ai,j = 1.

slide-11
SLIDE 11

Motivation from Metagenomics

◮ x ∈ RN (N = 273, 727): concentrations of known bacteria in

a given environmental sample. Sparsity assumption is realistic. Note also that x ≥ 0 and

j xj = 1. ◮ y ∈ Rm (m = 46 = 4, 096): frequencies of length-6 subwords

(in 16S rRNA gene reads or in whole-genome shotgun reads)

◮ A ∈ Rm×N: frequencies of length-6 subwords in all known

(i.e., sequenced) bacteria. It is a frequency matrix, that is, Ai,j ≥ 0 and m

i=1Ai,j = 1. ◮ Quikr improves on traditional read-by-read methods,

especially in terms of speed.

slide-12
SLIDE 12

Motivation from Metagenomics

◮ x ∈ RN (N = 273, 727): concentrations of known bacteria in

a given environmental sample. Sparsity assumption is realistic. Note also that x ≥ 0 and

j xj = 1. ◮ y ∈ Rm (m = 46 = 4, 096): frequencies of length-6 subwords

(in 16S rRNA gene reads or in whole-genome shotgun reads)

◮ A ∈ Rm×N: frequencies of length-6 subwords in all known

(i.e., sequenced) bacteria. It is a frequency matrix, that is, Ai,j ≥ 0 and m

i=1Ai,j = 1. ◮ Quikr improves on traditional read-by-read methods,

especially in terms of speed.

◮ Codes available at

sourceforge.net/projects/quikr/ sourceforge.net/projects/wgsquikr/

slide-13
SLIDE 13

Exact Measurements

slide-14
SLIDE 14

Exact Measurements

Let x ∈ RN be a nonnegative vector with support S.

slide-15
SLIDE 15

Exact Measurements

Let x ∈ RN be a nonnegative vector with support S.

◮ x is the unique minimizer of z1 s.to Az = y iff

(BP) for all v ∈ ker A \ {0},

  • j∈S vj
  • <

ℓ∈S |vℓ|.

slide-16
SLIDE 16

Exact Measurements

Let x ∈ RN be a nonnegative vector with support S.

◮ x is the unique minimizer of z1 s.to Az = y iff

(BP) for all v ∈ ker A \ {0},

  • j∈S vj
  • <

ℓ∈S |vℓ|. ◮ x is the unique minimizer of z1 s.to Az = y and z ≥ 0 iff

(NNBP) for all v ∈ ker A \ {0}, vS ≥ 0 ⇒ N

i=1 vi > 0.

slide-17
SLIDE 17

Exact Measurements

Let x ∈ RN be a nonnegative vector with support S.

◮ x is the unique minimizer of z1 s.to Az = y iff

(BP) for all v ∈ ker A \ {0},

  • j∈S vj
  • <

ℓ∈S |vℓ|. ◮ x is the unique minimizer of z1 s.to Az = y and z ≥ 0 iff

(NNBP) for all v ∈ ker A \ {0}, vS ≥ 0 ⇒ N

i=1 vi > 0. ◮ x is the unique z ≥ 0 s.to Az = y iff

(F) for all v ∈ ker A \ {0}, vS ≥ 0 is impossible.

slide-18
SLIDE 18

Exact Measurements

Let x ∈ RN be a nonnegative vector with support S.

◮ x is the unique minimizer of z1 s.to Az = y iff

(BP) for all v ∈ ker A \ {0},

  • j∈S vj
  • <

ℓ∈S |vℓ|. ◮ x is the unique minimizer of z1 s.to Az = y and z ≥ 0 iff

(NNBP) for all v ∈ ker A \ {0}, vS ≥ 0 ⇒ N

i=1 vi > 0. ◮ x is the unique z ≥ 0 s.to Az = y iff

(F) for all v ∈ ker A \ {0}, vS ≥ 0 is impossible. In general, (F)⇒(NNBP) and (BP)⇒(NNBP).

slide-19
SLIDE 19

Exact Measurements

Let x ∈ RN be a nonnegative vector with support S.

◮ x is the unique minimizer of z1 s.to Az = y iff

(BP) for all v ∈ ker A \ {0},

  • j∈S vj
  • <

ℓ∈S |vℓ|. ◮ x is the unique minimizer of z1 s.to Az = y and z ≥ 0 iff

(NNBP) for all v ∈ ker A \ {0}, vS ≥ 0 ⇒ N

i=1 vi > 0. ◮ x is the unique z ≥ 0 s.to Az = y iff

(F) for all v ∈ ker A \ {0}, vS ≥ 0 is impossible. In general, (F)⇒(NNBP) and (BP)⇒(NNBP). If 1 ∈ im(A⊤) (e.g. if A is a frequency matrix), then (NNBP)⇒(F)⇒(BP).

slide-20
SLIDE 20

Exact Measurements

Let x ∈ RN be a nonnegative vector with support S.

◮ x is the unique minimizer of z1 s.to Az = y iff

(BP) for all v ∈ ker A \ {0},

  • j∈S vj
  • <

ℓ∈S |vℓ|. ◮ x is the unique minimizer of z1 s.to Az = y and z ≥ 0 iff

(NNBP) for all v ∈ ker A \ {0}, vS ≥ 0 ⇒ N

i=1 vi > 0. ◮ x is the unique z ≥ 0 s.to Az = y iff

(F) for all v ∈ ker A \ {0}, vS ≥ 0 is impossible. In general, (F)⇒(NNBP) and (BP)⇒(NNBP). If 1 ∈ im(A⊤) (e.g. if A is a frequency matrix), then (NNBP)⇒(F)⇒(BP). Morale: ℓ1-minimization not suited for nonnegative sparse recovery.

slide-21
SLIDE 21

Nonnegative Least Squares

◮ To solve the feasibility problem, one may consider

minimize

z∈RN

y − Az2

2

subject to z ≥ 0.

slide-22
SLIDE 22

Nonnegative Least Squares

◮ To solve the feasibility problem, one may consider

minimize

z∈RN

y − Az2

2

subject to z ≥ 0.

◮ MATLAB’s lsqnonneg implements [Lawson–Hanson 74].

slide-23
SLIDE 23

Nonnegative Least Squares

◮ To solve the feasibility problem, one may consider

minimize

z∈RN

y − Az2

2

subject to z ≥ 0.

◮ MATLAB’s lsqnonneg implements [Lawson–Hanson 74]. ◮ This algorithm iterates the scheme

Sn+1 = Sn ∪

  • jn+1 = argmax j
  • A∗(y − Axn)
  • j
  • ,

xn+1 = argmin

  • y − Az2, supp(z) ⊆ Sn+1

,

slide-24
SLIDE 24

Nonnegative Least Squares

◮ To solve the feasibility problem, one may consider

minimize

z∈RN

y − Az2

2

subject to z ≥ 0.

◮ MATLAB’s lsqnonneg implements [Lawson–Hanson 74]. ◮ This algorithm iterates the scheme

Sn+1 = Sn ∪

  • jn+1 = argmax j
  • A∗(y − Axn)
  • j
  • ,

xn+1 = argmin

  • y − Az2, supp(z) ⊆ Sn+1

, and inner loop to make sure that xn+1 ≥ 0.

slide-25
SLIDE 25

Nonnegative Least Squares

◮ To solve the feasibility problem, one may consider

minimize

z∈RN

y − Az2

2

subject to z ≥ 0.

◮ MATLAB’s lsqnonneg implements [Lawson–Hanson 74]. ◮ This algorithm iterates the scheme

Sn+1 = Sn ∪

  • jn+1 = argmax j
  • A∗(y − Axn)
  • j
  • ,

xn+1 = argmin

  • y − Az2, supp(z) ⊆ Sn+1

, and inner loop to make sure that xn+1 ≥ 0.

◮ Connection with OMP explains suitability for sparse recovery.

slide-26
SLIDE 26

Inaccurate Measurements

slide-27
SLIDE 27

Inaccurate Measurements

◮ When y = Ax + e with e = 0, a classical strategy consists in

solving the ℓ1-regularization minimize

z∈RN

z1 + νy − Az2

2

subject to z ≥ 0.

slide-28
SLIDE 28

Inaccurate Measurements

◮ When y = Ax + e with e = 0, a classical strategy consists in

solving the ℓ1-regularization minimize

z∈RN

z1 + νy − Az2

2

subject to z ≥ 0.

◮ We prefer the ℓ1-squared regularization

minimize

z∈RN

z2

1 + λ2y − Az2 2

subject to z ≥ 0,

slide-29
SLIDE 29

Inaccurate Measurements

◮ When y = Ax + e with e = 0, a classical strategy consists in

solving the ℓ1-regularization minimize

z∈RN

z1 + νy − Az2

2

subject to z ≥ 0.

◮ We prefer the ℓ1-squared regularization

minimize

z∈RN

z2

1 + λ2y − Az2 2

subject to z ≥ 0, because it is recast as the Nonnegative Least Squares problem minimize

z∈RN

  • y −

Az2

2

subject to z ≥ 0, where A = 1 · · · 1 λA

  • and

y = λy

  • .
slide-30
SLIDE 30

Inaccurate Measurements

◮ When y = Ax + e with e = 0, a classical strategy consists in

solving the ℓ1-regularization minimize

z∈RN

z1 + νy − Az2

2

subject to z ≥ 0.

◮ We prefer the ℓ1-squared regularization

minimize

z∈RN

z2

1 + λ2y − Az2 2

subject to z ≥ 0, because it is recast as the Nonnegative Least Squares problem minimize

z∈RN

  • y −

Az2

2

subject to z ≥ 0, where A = 1 · · · 1 λA

  • and

y = λy

  • .

◮ For frequency matrices, as λ → ∞, the minimizer xλ tends to

the minimizer of z1 subject to Az = y and z ≥ 0.

slide-31
SLIDE 31

Extension: Sparse Recovery via NNLS

slide-32
SLIDE 32

Extension: Sparse Recovery via NNLS

◮ Decompose vectors z ∈ RN as z = z+ − z− with z+, z− ∈ RN +.

slide-33
SLIDE 33

Extension: Sparse Recovery via NNLS

◮ Decompose vectors z ∈ RN as z = z+ − z− with z+, z− ∈ RN +. ◮ The ℓ1-squared regularization

(REG) minimize z2

1 + λ2y − Az2 2

is recast as the Nonnegative Least Squares problem minimize y − A z2

2

subject to z ≥ 0, where y = λy

  • ,

A = 1 · · · 1 1 · · · 1 λA −λA

  • ,

z = z+ z−

  • .
slide-34
SLIDE 34

Extension: Sparse Recovery via NNLS

◮ Decompose vectors z ∈ RN as z = z+ − z− with z+, z− ∈ RN +. ◮ The ℓ1-squared regularization

(REG) minimize z2

1 + λ2y − Az2 2

is recast as the Nonnegative Least Squares problem minimize y − A z2

2

subject to z ≥ 0, where y = λy

  • ,

A = 1 · · · 1 1 · · · 1 λA −λA

  • ,

z = z+ z−

  • .

◮ For Gaussian matrices (RNSP and QP hold), the solutions xλ

  • f (REG) with y = Ax + e obey, for all x ∈ RN and e ∈ Rm,

x − xλ1 ≤ C σs(x)1 + D√s e2 + E s λ2 x1.

slide-35
SLIDE 35

Part II: Disjointed Sparse Recovery (joint work with M. Minner and T. Needham)

slide-36
SLIDE 36

Motivation from Radar

slide-37
SLIDE 37

Motivation from Radar

◮ x ∈ RN: positions of airplanes relative to a discretized grid.

slide-38
SLIDE 38

Motivation from Radar

◮ x ∈ RN: positions of airplanes relative to a discretized grid. ◮ Few airplanes that are not too close to one another:

sparsity and disjointedness.

slide-39
SLIDE 39

Motivation from Radar

◮ x ∈ RN: positions of airplanes relative to a discretized grid. ◮ Few airplanes that are not too close to one another:

sparsity and disjointedness.

◮ Disjointedness is also relevant to model neural spike trains.

[Hedge–Duarte–Cevher 09]

slide-40
SLIDE 40

Motivation from Radar

◮ x ∈ RN: positions of airplanes relative to a discretized grid. ◮ Few airplanes that are not too close to one another:

sparsity and disjointedness.

◮ Disjointedness is also relevant to model neural spike trains.

[Hedge–Duarte–Cevher 09]

◮ We say that x ∈ RN is s-sparse and d-disjointed if

slide-41
SLIDE 41

Motivation from Radar

◮ x ∈ RN: positions of airplanes relative to a discretized grid. ◮ Few airplanes that are not too close to one another:

sparsity and disjointedness.

◮ Disjointedness is also relevant to model neural spike trains.

[Hedge–Duarte–Cevher 09]

◮ We say that x ∈ RN is s-sparse and d-disjointed if

◮ x has no more than s nonzero entries,

slide-42
SLIDE 42

Motivation from Radar

◮ x ∈ RN: positions of airplanes relative to a discretized grid. ◮ Few airplanes that are not too close to one another:

sparsity and disjointedness.

◮ Disjointedness is also relevant to model neural spike trains.

[Hedge–Duarte–Cevher 09]

◮ We say that x ∈ RN is s-sparse and d-disjointed if

◮ x has no more than s nonzero entries, ◮ there are ≥ d zero entries between two nonzero entries.

slide-43
SLIDE 43

Resolution of the Fundamental Question

slide-44
SLIDE 44

Resolution of the Fundamental Question

◮ The minimal number of linear measurements for the recovery

  • f all s-sparse vectors is

mspa ≍ s ln

  • e N

s

  • .
slide-45
SLIDE 45

Resolution of the Fundamental Question

◮ The minimal number of linear measurements for the recovery

  • f all s-sparse vectors is

mspa ≍ s ln

  • e N

s

  • .

◮ The minimal number of linear measurements for the recovery

  • f all d-disjointed vectors is [Cand`

es–Fernandez-Granda 14] mdis ≍ N d .

slide-46
SLIDE 46

Resolution of the Fundamental Question

◮ The minimal number of linear measurements for the recovery

  • f all s-sparse vectors is

mspa ≍ s ln

  • e N

s

  • .

◮ The minimal number of linear measurements for the recovery

  • f all d-disjointed vectors is [Cand`

es–Fernandez-Granda 14] mdis ≍ N d .

◮ What is the minimal number of linear measurements needed

for the recovery of all s-sparse d-disjointed vectors?

slide-47
SLIDE 47

Resolution of the Fundamental Question

◮ The minimal number of linear measurements for the recovery

  • f all s-sparse vectors is

mspa ≍ s ln

  • e N

s

  • .

◮ The minimal number of linear measurements for the recovery

  • f all d-disjointed vectors is [Cand`

es–Fernandez-Granda 14] mdis ≍ N d .

◮ What is the minimal number of linear measurements needed

for the recovery of all s-sparse d-disjointed vectors? Answer: mspa&dis ≍ s ln

  • e N − d(s − 1)

s

  • .
slide-48
SLIDE 48

Resolution of the Fundamental Question

◮ The minimal number of linear measurements for the recovery

  • f all s-sparse vectors is

mspa ≍ s ln

  • e N

s

  • .

◮ The minimal number of linear measurements for the recovery

  • f all d-disjointed vectors is [Cand`

es–Fernandez-Granda 14] mdis ≍ N d .

◮ What is the minimal number of linear measurements needed

for the recovery of all s-sparse d-disjointed vectors? Answer: mspa&dis ≍ s ln

  • e N − d(s − 1)

s

  • .

◮ There is no benefit in knowing the simultaneity of sparsity and

disjointedness over knowing only one of the structures, since mspa&dis ≍ min {mspa, mdis} .

slide-49
SLIDE 49

Sparse Disjointed Supports

slide-50
SLIDE 50

Sparse Disjointed Supports

There are N − d(s − 1) s

  • e N − d(s − 1)

s s d-disjointed subsets of 1 : N with size s.

slide-51
SLIDE 51

Sparse Disjointed Supports

There are N − d(s − 1) s

  • e N − d(s − 1)

s s d-disjointed subsets of 1 : N with size s.

≥ ¡d 1 2 ≥ ¡d 3 ¡ s Length ¡N ¡ Length ¡N+d d+1 d+1 Length ¡N-­‑d(s-­‑1) insert ¡d Length ¡N+d d+1

slide-52
SLIDE 52

Sufficient Number of Measurements via IHT

slide-53
SLIDE 53

Sufficient Number of Measurements via IHT

◮ The adaptation of iterative hard thresholding is

xn+1 = Ps,d(xn + A∗(y − Axn)), where Ps,d is the projection onto s-sparse d-disjointed vectors.

slide-54
SLIDE 54

Sufficient Number of Measurements via IHT

◮ The adaptation of iterative hard thresholding is

xn+1 = Ps,d(xn + A∗(y − Axn)), where Ps,d is the projection onto s-sparse d-disjointed vectors.

◮ For any s-sparse d-disjointed x ∈ RN and any e ∈ Rm,

x − lim

n→∞ xn2 ≤ De2

as soon as the RI-like property (1−δ)z+z′+z′′2

2 ≤ A(z+z′+z′′)2 2 ≤ (1+δ)z+z′+z′′2 2

holds with δ < 1/2 for all s-sparse d-disjointed z, z′, z′′ ∈ RN.

slide-55
SLIDE 55

Sufficient Number of Measurements via IHT

◮ The adaptation of iterative hard thresholding is

xn+1 = Ps,d(xn + A∗(y − Axn)), where Ps,d is the projection onto s-sparse d-disjointed vectors.

◮ For any s-sparse d-disjointed x ∈ RN and any e ∈ Rm,

x − lim

n→∞ xn2 ≤ De2

as soon as the RI-like property (1−δ)z+z′+z′′2

2 ≤ A(z+z′+z′′)2 2 ≤ (1+δ)z+z′+z′′2 2

holds with δ < 1/2 for all s-sparse d-disjointed z, z′, z′′ ∈ RN.

◮ The latter occurs w/hp for m ≥ Cδ−2 ln(e(N − d(s − 1))/s).

slide-56
SLIDE 56

Sufficient Number of Measurements via IHT

◮ The adaptation of iterative hard thresholding is

xn+1 = Ps,d(xn + A∗(y − Axn)), where Ps,d is the projection onto s-sparse d-disjointed vectors.

◮ For any s-sparse d-disjointed x ∈ RN and any e ∈ Rm,

x − lim

n→∞ xn2 ≤ De2

as soon as the RI-like property (1−δ)z+z′+z′′2

2 ≤ A(z+z′+z′′)2 2 ≤ (1+δ)z+z′+z′′2 2

holds with δ < 1/2 for all s-sparse d-disjointed z, z′, z′′ ∈ RN.

◮ The latter occurs w/hp for m ≥ Cδ−2 ln(e(N − d(s − 1))/s). ◮ Similar results obtained earlier for the adaptation of CoSaMP.

slide-57
SLIDE 57

Computing the Projection Ps,d

slide-58
SLIDE 58

Computing the Projection Ps,d

◮ [Hedge–Duarte–Cevher 09] propose an integer program relaxed

to a linear program that is solved in O(N3.5) operations.

slide-59
SLIDE 59

Computing the Projection Ps,d

◮ [Hedge–Duarte–Cevher 09] propose an integer program relaxed

to a linear program that is solved in O(N3.5) operations.

◮ A dynamic program can be solved in O(N2) operations.

slide-60
SLIDE 60

Computing the Projection Ps,d

◮ [Hedge–Duarte–Cevher 09] propose an integer program relaxed

to a linear program that is solved in O(N3.5) operations.

◮ A dynamic program can be solved in O(N2) operations. ◮ Determine F(N, s), where

F(n, r) := min   

n

  • j=1

|xj − zj|2 : z ∈ Cn r-sparse d-disjointed    = min    F(n − 1, r) + |xn|p, F(n − d − 1, r − 1) +

n−1

  • j=n−d

|xj|p.

slide-61
SLIDE 61

Computing the Projection Ps,d, ctd.

Dynamic program for x = (1, 0, 1, 21/4, 1, 0, 2−1/2), s = 3, d = 1.

x 1 1 1.1892 1 0.7071 F(n, r) r = 0 r = 1 r = 2 r = 3 n = 1 1 n = 2 1 n = 3 2 1 n = 4 3.4142 2 1 1 n = 5 4.4142 3 2 1.4142 n = 6 4.4142 3 2 1.4142 n = 7 4.9142 3.5 2.5 1.9142

slide-62
SLIDE 62

Necessary Number of Measurements

slide-63
SLIDE 63

Necessary Number of Measurements

◮ Noninflating measurements relative to our model:

Az2 ≤ cz2 whenever z is s-sparse d-disjointed.

slide-64
SLIDE 64

Necessary Number of Measurements

◮ Noninflating measurements relative to our model:

Az2 ≤ cz2 whenever z is s-sparse d-disjointed.

◮ ∆ reconstruction map providing the robust estimate

x − ∆(Ax + e)2 ≤ De2, valid for all s-sparse d-disjointed x and for all e.

slide-65
SLIDE 65

Necessary Number of Measurements

◮ Noninflating measurements relative to our model:

Az2 ≤ cz2 whenever z is s-sparse d-disjointed.

◮ ∆ reconstruction map providing the robust estimate

x − ∆(Ax + e)2 ≤ De2, valid for all s-sparse d-disjointed x and for all e.

◮ Then

m ≥ C s ln

  • e N − d(s − 1)

s

  • .
slide-66
SLIDE 66

Key Combinatorial Lemma

slide-67
SLIDE 67

Key Combinatorial Lemma

There exist n ≥ N − d(s − 1) c1s c2s d-disjointed subsets S1, . . . , Sn of 1 : N such that card(Si) = s for all i, card(Si ∩ Sj) < s 2 for all i = j.

slide-68
SLIDE 68

Key Combinatorial Lemma

There exist n ≥ N − d(s − 1) c1s c2s d-disjointed subsets S1, . . . , Sn of 1 : N such that card(Si) = s for all i, card(Si ∩ Sj) < s 2 for all i = j. This extends a crucial result known for d = 0 (sparse vectors),

slide-69
SLIDE 69

Key Combinatorial Lemma

There exist n ≥ N − d(s − 1) c1s c2s d-disjointed subsets S1, . . . , Sn of 1 : N such that card(Si) = s for all i, card(Si ∩ Sj) < s 2 for all i = j. This extends a crucial result known for d = 0 (sparse vectors), but the counting argument must be somewhat refined.