SLIDE 1
Fun with ℓ1 and ℓ2
x1 ≤ √nx2.
SLIDE 2 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
SLIDE 3 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
SLIDE 4 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
supp(x) is non-zero indices of x.
SLIDE 5 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
supp(x) is non-zero indices of x. If concentrated mass, x1 = x2.
SLIDE 6 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
supp(x) is non-zero indices of x. If concentrated mass, x1 = x2. x = (1,0,0,...,0).
SLIDE 7 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
supp(x) is non-zero indices of x. If concentrated mass, x1 = x2. x = (1,0,0,...,0). If spreadout, √nx2 ≤ x1.
SLIDE 8 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
supp(x) is non-zero indices of x. If concentrated mass, x1 = x2. x = (1,0,0,...,0). If spreadout, √nx2 ≤ x1. x = (1,1,1,...,1).
SLIDE 9 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
supp(x) is non-zero indices of x. If concentrated mass, x1 = x2. x = (1,0,0,...,0). If spreadout, √nx2 ≤ x1. x = (1,1,1,...,1). If kind of spread out, x2 ≤
1 √ k x1.
SLIDE 10 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
supp(x) is non-zero indices of x. If concentrated mass, x1 = x2. x = (1,0,0,...,0). If spreadout, √nx2 ≤ x1. x = (1,1,1,...,1). If kind of spread out, x2 ≤
1 √ k x1.
x has k 1’s.
SLIDE 11 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
supp(x) is non-zero indices of x. If concentrated mass, x1 = x2. x = (1,0,0,...,0). If spreadout, √nx2 ≤ x1. x = (1,1,1,...,1). If kind of spread out, x2 ≤
1 √ k x1.
x has k 1’s. Fixing v2, sparse vectors have small v1 norm, dense ones have big v1 norm.
SLIDE 12
Compressed Sensing.
Find x with small number of non-zeros using linear measurements.
SLIDE 13
Compressed Sensing.
Find x with small number of non-zeros using linear measurements. Ax = b.
SLIDE 14
Compressed Sensing.
Find x with small number of non-zeros using linear measurements. Ax = b. Application: MRI.
SLIDE 15
Compressed Sensing.
Find x with small number of non-zeros using linear measurements. Ax = b. Application: MRI. Find x with k-sparse x, i.e., supp(x) ≤ k.
SLIDE 16
Compressed Sensing.
Find x with small number of non-zeros using linear measurements. Ax = b. Application: MRI. Find x with k-sparse x, i.e., supp(x) ≤ k. ℓ0-minimization.
SLIDE 17
Compressed Sensing.
Find x with small number of non-zeros using linear measurements. Ax = b. Application: MRI. Find x with k-sparse x, i.e., supp(x) ≤ k. ℓ0-minimization. Extremely “non-convex”.
SLIDE 18
Compressed Sensing.
Find x with small number of non-zeros using linear measurements. Ax = b. Application: MRI. Find x with k-sparse x, i.e., supp(x) ≤ k. ℓ0-minimization. Extremely “non-convex”. Find solution to minw1,Ax = b.
SLIDE 19
Compressed Sensing.
Find x with small number of non-zeros using linear measurements. Ax = b. Application: MRI. Find x with k-sparse x, i.e., supp(x) ≤ k. ℓ0-minimization. Extremely “non-convex”. Find solution to minw1,Ax = b. Linear Program!
SLIDE 20
Compressed Sensing.
Find x with small number of non-zeros using linear measurements. Ax = b. Application: MRI. Find x with k-sparse x, i.e., supp(x) ≤ k. ℓ0-minimization. Extremely “non-convex”. Find solution to minw1,Ax = b. Linear Program! Exercise.
SLIDE 21
Restricted Isometry Property (RIP) matrices.
Definition: A matrix A is RIP for δk if any k-sparse vector x (1−δk)x2 ≤ Ax2 ≤ (1+δk)x2.
SLIDE 22
Restricted Isometry Property (RIP) matrices.
Definition: A matrix A is RIP for δk if any k-sparse vector x (1−δk)x2 ≤ Ax2 ≤ (1+δk)x2. Theorem [Candes-Tao]: For any matrix RIP matrix A with δ2k +δ3k < 1, for Ax = b with a k-sparse solution, then the solution to miny1,Ay = b, has y = x.
SLIDE 23
Fun with ℓ1 and ℓ2
x1 ≤ √nx2.
SLIDE 24 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
SLIDE 25 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
SLIDE 26 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
supp(x) is non-zero indices of x.
SLIDE 27 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
supp(x) is non-zero indices of x. If concentrated mass, x1 = x2.
SLIDE 28 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
supp(x) is non-zero indices of x. If concentrated mass, x1 = x2. x = (1,0,0,...,0).
SLIDE 29 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
supp(x) is non-zero indices of x. If concentrated mass, x1 = x2. x = (1,0,0,...,0). If spreadout, √nx2 ≤ x1.
SLIDE 30 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
supp(x) is non-zero indices of x. If concentrated mass, x1 = x2. x = (1,0,0,...,0). If spreadout, √nx2 ≤ x1. x = (1,1,1,...,1).
SLIDE 31 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
supp(x) is non-zero indices of x. If concentrated mass, x1 = x2. x = (1,0,0,...,0). If spreadout, √nx2 ≤ x1. x = (1,1,1,...,1). If kind of spread out, x2 ≤
1 √ k x1.
SLIDE 32 Fun with ℓ1 and ℓ2
x1 ≤ √nx2. x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
x1 = x ·sgn(x) ≤ x2sgn(x)2 ≤
supp(x) is non-zero indices of x. If concentrated mass, x1 = x2. x = (1,0,0,...,0). If spreadout, √nx2 ≤ x1. x = (1,1,1,...,1). If kind of spread out, x2 ≤
1 √ k x1.
x has k 1’s.
SLIDE 33
Almost Euclidean Nullspace.
Theorem: For a random ±1, d ×n matrix A, and for any x in ker(A) some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
x2 <
√ 1 √ 16k x1. (∗)
SLIDE 34
Almost Euclidean Nullspace.
Theorem: For a random ±1, d ×n matrix A, and for any x in ker(A) some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
x2 <
√ 1 √ 16k x1. (∗)
Intuition: “Mass in x is spread out over k entries.”
SLIDE 35
Almost Euclidean Nullspace.
Theorem: For a random ±1, d ×n matrix A, and for any x in ker(A) some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
x2 <
√ 1 √ 16k x1. (∗)
Intuition: “Mass in x is spread out over k entries.” The nullspace of A, is almost euclidean.
SLIDE 36
Almost Euclidean Nullspace.
Theorem: For a random ±1, d ×n matrix A, and for any x in ker(A) some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
x2 <
√ 1 √ 16k x1. (∗)
Intuition: “Mass in x is spread out over k entries.” The nullspace of A, is almost euclidean. Typical vectors are spread out: every vector is kind of spread out.
SLIDE 37
Almost Euclidean Nullspace.
Theorem: For a random ±1, d ×n matrix A, and for any x in ker(A) some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
x2 <
√ 1 √ 16k x1. (∗)
Intuition: “Mass in x is spread out over k entries.” The nullspace of A, is almost euclidean. Typical vectors are spread out: every vector is kind of spread out. The ℓ1 ball is closer to scaling of ℓ2 ball for vectors in the null-space.
SLIDE 38
Almost Euclidean Nullspace.
Theorem: For a random ±1, d ×n matrix A, and for any x in ker(A) some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
x2 <
√ 1 √ 16k x1. (∗)
Intuition: “Mass in x is spread out over k entries.” The nullspace of A, is almost euclidean. Typical vectors are spread out: every vector is kind of spread out. The ℓ1 ball is closer to scaling of ℓ2 ball for vectors in the null-space. Idea: Consider random r ×n matrix A over GF(2).
SLIDE 39
Almost Euclidean Nullspace.
Theorem: For a random ±1, d ×n matrix A, and for any x in ker(A) some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
x2 <
√ 1 √ 16k x1. (∗)
Intuition: “Mass in x is spread out over k entries.” The nullspace of A, is almost euclidean. Typical vectors are spread out: every vector is kind of spread out. The ℓ1 ball is closer to scaling of ℓ2 ball for vectors in the null-space. Idea: Consider random r ×n matrix A over GF(2). For a vector x in GF(2).
SLIDE 40
Almost Euclidean Nullspace.
Theorem: For a random ±1, d ×n matrix A, and for any x in ker(A) some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
x2 <
√ 1 √ 16k x1. (∗)
Intuition: “Mass in x is spread out over k entries.” The nullspace of A, is almost euclidean. Typical vectors are spread out: every vector is kind of spread out. The ℓ1 ball is closer to scaling of ℓ2 ball for vectors in the null-space. Idea: Consider random r ×n matrix A over GF(2). For a vector x in GF(2). A·x = 0, with probability (1/2)r if r rows.
SLIDE 41 Almost Euclidean Nullspace.
Theorem: For a random ±1, d ×n matrix A, and for any x in ker(A) some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
x2 <
√ 1 √ 16k x1. (∗)
Intuition: “Mass in x is spread out over k entries.” The nullspace of A, is almost euclidean. Typical vectors are spread out: every vector is kind of spread out. The ℓ1 ball is closer to scaling of ℓ2 ball for vectors in the null-space. Idea: Consider random r ×n matrix A over GF(2). For a vector x in GF(2). A·x = 0, with probability (1/2)r if r rows. There are < X = 2 n
k
- vectors x with fewer than k zeros.
SLIDE 42 Almost Euclidean Nullspace.
Theorem: For a random ±1, d ×n matrix A, and for any x in ker(A) some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
x2 <
√ 1 √ 16k x1. (∗)
Intuition: “Mass in x is spread out over k entries.” The nullspace of A, is almost euclidean. Typical vectors are spread out: every vector is kind of spread out. The ℓ1 ball is closer to scaling of ℓ2 ball for vectors in the null-space. Idea: Consider random r ×n matrix A over GF(2). For a vector x in GF(2). A·x = 0, with probability (1/2)r if r rows. There are < X = 2 n
k
- vectors x with fewer than k zeros.
If r > log(2 n
k
k ), plus union bound.
SLIDE 43 Almost Euclidean Nullspace.
Theorem: For a random ±1, d ×n matrix A, and for any x in ker(A) some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
x2 <
√ 1 √ 16k x1. (∗)
Intuition: “Mass in x is spread out over k entries.” The nullspace of A, is almost euclidean. Typical vectors are spread out: every vector is kind of spread out. The ℓ1 ball is closer to scaling of ℓ2 ball for vectors in the null-space. Idea: Consider random r ×n matrix A over GF(2). For a vector x in GF(2). A·x = 0, with probability (1/2)r if r rows. There are < X = 2 n
k
- vectors x with fewer than k zeros.
If r > log(2 n
k
k ), plus union bound.
= ⇒ Ax = 0 for all vectors that are k-sparse.
SLIDE 44 Almost Euclidean Nullspace.
Theorem: For a random ±1, d ×n matrix A, and for any x in ker(A) some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
x2 <
√ 1 √ 16k x1. (∗)
Intuition: “Mass in x is spread out over k entries.” The nullspace of A, is almost euclidean. Typical vectors are spread out: every vector is kind of spread out. The ℓ1 ball is closer to scaling of ℓ2 ball for vectors in the null-space. Idea: Consider random r ×n matrix A over GF(2). For a vector x in GF(2). A·x = 0, with probability (1/2)r if r rows. There are < X = 2 n
k
- vectors x with fewer than k zeros.
If r > log(2 n
k
k ), plus union bound.
= ⇒ Ax = 0 for all vectors that are k-sparse.
SLIDE 45 Almost Euclidean Nullspace.
Theorem: For a random ±1, d ×n matrix A, and for any x in ker(A) some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
x2 <
√ 1 √ 16k x1. (∗)
Intuition: “Mass in x is spread out over k entries.” The nullspace of A, is almost euclidean. Typical vectors are spread out: every vector is kind of spread out. The ℓ1 ball is closer to scaling of ℓ2 ball for vectors in the null-space. Idea: Consider random r ×n matrix A over GF(2). For a vector x in GF(2). A·x = 0, with probability (1/2)r if r rows. There are < X = 2 n
k
- vectors x with fewer than k zeros.
If r > log(2 n
k
k ), plus union bound.
= ⇒ Ax = 0 for all vectors that are k-sparse. That is, random A has no sparse vectors in null-space.
SLIDE 46 Almost Euclidean Nullspace.
Theorem: For a random ±1, d ×n matrix A, and for any x in ker(A) some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
x2 <
√ 1 √ 16k x1. (∗)
Intuition: “Mass in x is spread out over k entries.” The nullspace of A, is almost euclidean. Typical vectors are spread out: every vector is kind of spread out. The ℓ1 ball is closer to scaling of ℓ2 ball for vectors in the null-space. Idea: Consider random r ×n matrix A over GF(2). For a vector x in GF(2). A·x = 0, with probability (1/2)r if r rows. There are < X = 2 n
k
- vectors x with fewer than k zeros.
If r > log(2 n
k
k ), plus union bound.
= ⇒ Ax = 0 for all vectors that are k-sparse. That is, random A has no sparse vectors in null-space. Note: Parity check matrix of linear code!
SLIDE 47
Small projection onto small set of coordinates.
Consider A with property, x ∈ ker(A), has x2 <
1 16 √ k x1.
SLIDE 48
Small projection onto small set of coordinates.
Consider A with property, x ∈ ker(A), has x2 <
1 16 √ k x1.
Lemma: For v ∈ ker(A), T ⊂ [n], |T| < k, vT 1 < v1
4 .
SLIDE 49
Small projection onto small set of coordinates.
Consider A with property, x ∈ ker(A), has x2 <
1 16 √ k x1.
Lemma: For v ∈ ker(A), T ⊂ [n], |T| < k, vT 1 < v1
4 .
Proof: vT 1 ≤
SLIDE 50 Small projection onto small set of coordinates.
Consider A with property, x ∈ ker(A), has x2 <
1 16 √ k x1.
Lemma: For v ∈ ker(A), T ⊂ [n], |T| < k, vT 1 < v1
4 .
Proof: vT 1 ≤
SLIDE 51 Small projection onto small set of coordinates.
Consider A with property, x ∈ ker(A), has x2 <
1 16 √ k x1.
Lemma: For v ∈ ker(A), T ⊂ [n], |T| < k, vT 1 < v1
4 .
Proof: vT 1 ≤
SLIDE 52 Small projection onto small set of coordinates.
Consider A with property, x ∈ ker(A), has x2 <
1 16 √ k x1.
Lemma: For v ∈ ker(A), T ⊂ [n], |T| < k, vT 1 < v1
4 .
Proof: vT 1 ≤
1 √ 16k v1
SLIDE 53 Small projection onto small set of coordinates.
Consider A with property, x ∈ ker(A), has x2 <
1 16 √ k x1.
Lemma: For v ∈ ker(A), T ⊂ [n], |T| < k, vT 1 < v1
4 .
Proof: vT 1 ≤
1 √ 16k v1 < 1 4v1
SLIDE 54 Small projection onto small set of coordinates.
Consider A with property, x ∈ ker(A), has x2 <
1 16 √ k x1.
Lemma: For v ∈ ker(A), T ⊂ [n], |T| < k, vT 1 < v1
4 .
Proof: vT 1 ≤
1 √ 16k v1 < 1 4v1
SLIDE 55 Small projection onto small set of coordinates.
Consider A with property, x ∈ ker(A), has x2 <
1 16 √ k x1.
Lemma: For v ∈ ker(A), T ⊂ [n], |T| < k, vT 1 < v1
4 .
Proof: vT 1 ≤
1 √ 16k v1 < 1 4v1
Intuition:
SLIDE 56 Small projection onto small set of coordinates.
Consider A with property, x ∈ ker(A), has x2 <
1 16 √ k x1.
Lemma: For v ∈ ker(A), T ⊂ [n], |T| < k, vT 1 < v1
4 .
Proof: vT 1 ≤
1 √ 16k v1 < 1 4v1
Intuition: For any v ∈ ker(A), the amount of mass in any small, k, set of coordinates is small, 1
4v1.
SLIDE 57 Small projection onto small set of coordinates.
Consider A with property, x ∈ ker(A), has x2 <
1 16 √ k x1.
Lemma: For v ∈ ker(A), T ⊂ [n], |T| < k, vT 1 < v1
4 .
Proof: vT 1 ≤
1 √ 16k v1 < 1 4v1
Intuition: For any v ∈ ker(A), the amount of mass in any small, k, set of coordinates is small, 1
4v1.
Mass is spread out over more than k coordinates.
SLIDE 58
Optimum is correct!
Want to find: k-sparse solution to Ax = b.
SLIDE 59
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b.
SLIDE 60
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
SLIDE 61
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates.
SLIDE 62
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w.
SLIDE 63
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w. w = x +v where v ∈ ker(A).
SLIDE 64
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w. w = x +v where v ∈ ker(A).
SLIDE 65
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w. w = x +v where v ∈ ker(A). Will prove: v = 0 or w = x.
SLIDE 66
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w. w = x +v where v ∈ ker(A). Will prove: v = 0 or w = x. Contradiction
SLIDE 67
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w. w = x +v where v ∈ ker(A). Will prove: v = 0 or w = x. Contradiction ?
SLIDE 68
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w. w = x +v where v ∈ ker(A). Will prove: v = 0 or w = x. Contradiction ? Hmmm.
SLIDE 69
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w. w = x +v where v ∈ ker(A). Will prove: v = 0 or w = x. Contradiction ? Hmmm. Let T be non-zero coordinates of x.
SLIDE 70
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w. w = x +v where v ∈ ker(A). Will prove: v = 0 or w = x. Contradiction ? Hmmm. Let T be non-zero coordinates of x.
SLIDE 71
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w. w = x +v where v ∈ ker(A). Will prove: v = 0 or w = x. Contradiction ? Hmmm. Let T be non-zero coordinates of x. w1 = x +v1
SLIDE 72
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w. w = x +v where v ∈ ker(A). Will prove: v = 0 or w = x. Contradiction ? Hmmm. Let T be non-zero coordinates of x. w1 = x +v1 = xT +vT 1 +vT 1
SLIDE 73
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w. w = x +v where v ∈ ker(A). Will prove: v = 0 or w = x. Contradiction ? Hmmm. Let T be non-zero coordinates of x. w1 = x +v1 = xT +vT 1 +vT 1 v ≥ vT −vT
SLIDE 74
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w. w = x +v where v ∈ ker(A). Will prove: v = 0 or w = x. Contradiction ? Hmmm. Let T be non-zero coordinates of x. w1 = x +v1 = xT +vT 1 +vT 1 v ≥ vT −vT = ⇒
SLIDE 75
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w. w = x +v where v ∈ ker(A). Will prove: v = 0 or w = x. Contradiction ? Hmmm. Let T be non-zero coordinates of x. w1 = x +v1 = xT +vT 1 +vT 1 v ≥ vT −vT = ⇒ ≥ xT 1 −vT 1 +vT
SLIDE 76
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w. w = x +v where v ∈ ker(A). Will prove: v = 0 or w = x. Contradiction ? Hmmm. Let T be non-zero coordinates of x. w1 = x +v1 = xT +vT 1 +vT 1 v ≥ vT −vT = ⇒ ≥ xT 1 −vT 1 +vT ≥ xT 1 −vT 1 −vT 1 +v1
SLIDE 77
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w. w = x +v where v ∈ ker(A). Will prove: v = 0 or w = x. Contradiction ? Hmmm. Let T be non-zero coordinates of x. w1 = x +v1 = xT +vT 1 +vT 1 v ≥ vT −vT = ⇒ ≥ xT 1 −vT 1 +vT ≥ xT 1 −vT 1 −vT 1 +v1 ≥ x1 −2vT 1 +v1
SLIDE 78
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w. w = x +v where v ∈ ker(A). Will prove: v = 0 or w = x. Contradiction ? Hmmm. Let T be non-zero coordinates of x. w1 = x +v1 = xT +vT 1 +vT 1 v ≥ vT −vT = ⇒ ≥ xT 1 −vT 1 +vT ≥ xT 1 −vT 1 −vT 1 +v1 ≥ x1 −2vT 1 +v1 > x1.
SLIDE 79
Optimum is correct!
Want to find: k-sparse solution to Ax = b. Recall: minimize w1 with Aw = b. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k, vT 1 < v1
4 .
Idea: any nonzero vector, v ∈ ker(A) has small projection onto any k coordinates. Consider solution w. w = x +v where v ∈ ker(A). Will prove: v = 0 or w = x. Contradiction ? Hmmm. Let T be non-zero coordinates of x. w1 = x +v1 = xT +vT 1 +vT 1 v ≥ vT −vT = ⇒ ≥ xT 1 −vT 1 +vT ≥ xT 1 −vT 1 −vT 1 +v1 ≥ x1 −2vT 1 +v1 > x1. If v is nonzero.
SLIDE 80
Imperfect Case.
What if x is mostly sparse?
SLIDE 81
Imperfect Case.
What if x is mostly sparse? σk(x) = min
supp(z)≤k x −z1
SLIDE 82
Imperfect Case.
What if x is mostly sparse? σk(x) = min
supp(z)≤k x −z1
“Amount of x outside of k coordinates.”
SLIDE 83
Imperfect Case.
What if x is mostly sparse? σk(x) = min
supp(z)≤k x −z1
“Amount of x outside of k coordinates.” Theorem: If v ∈ ker(A) = ⇒ v2 ≤
1 16k v1, then solution to
minw1,Ax = b, has x −w1 ≤ 4σk(x).
SLIDE 84
Imperfect Case.
What if x is mostly sparse? σk(x) = min
supp(z)≤k x −z1
“Amount of x outside of k coordinates.” Theorem: If v ∈ ker(A) = ⇒ v2 ≤
1 16k v1, then solution to
minw1,Ax = b, has x −w1 ≤ 4σk(x). Still have.
SLIDE 85
Imperfect Case.
What if x is mostly sparse? σk(x) = min
supp(z)≤k x −z1
“Amount of x outside of k coordinates.” Theorem: If v ∈ ker(A) = ⇒ v2 ≤
1 16k v1, then solution to
minw1,Ax = b, has x −w1 ≤ 4σk(x). Still have. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k
16,
vT 1 < v1
4 .
SLIDE 86
Proof of w −x ≤ 4σ(x).
Again: σk(x) = minsupp(z)≤k |x −z|1.
SLIDE 87
Proof of w −x ≤ 4σ(x).
Again: σk(x) = minsupp(z)≤k |x −z|1. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k
16, vT 1 < v1 4 .
SLIDE 88
Proof of w −x ≤ 4σ(x).
Again: σk(x) = minsupp(z)≤k |x −z|1. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k
16, vT 1 < v1 4 .
Proof of Theorem: T be k largest in magnitude coordinates of x.
SLIDE 89
Proof of w −x ≤ 4σ(x).
Again: σk(x) = minsupp(z)≤k |x −z|1. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k
16, vT 1 < v1 4 .
Proof of Theorem: T be k largest in magnitude coordinates of x. x −w1 = (x −w)T 1 +(x −w)T 1
SLIDE 90
Proof of w −x ≤ 4σ(x).
Again: σk(x) = minsupp(z)≤k |x −z|1. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k
16, vT 1 < v1 4 .
Proof of Theorem: T be k largest in magnitude coordinates of x. x −w1 = (x −w)T 1 +(x −w)T 1 ≤ (x −w)T 1 +xT 1 +wT 1
SLIDE 91
Proof of w −x ≤ 4σ(x).
Again: σk(x) = minsupp(z)≤k |x −z|1. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k
16, vT 1 < v1 4 .
Proof of Theorem: T be k largest in magnitude coordinates of x. x −w1 = (x −w)T 1 +(x −w)T 1 ≤ (x −w)T 1 +xT 1 +wT 1 ≤ (x −w)T 1 +xT 1 +w1 −wT 1 triangle inequality on w
SLIDE 92
Proof of w −x ≤ 4σ(x).
Again: σk(x) = minsupp(z)≤k |x −z|1. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k
16, vT 1 < v1 4 .
Proof of Theorem: T be k largest in magnitude coordinates of x. x −w1 = (x −w)T 1 +(x −w)T 1 ≤ (x −w)T 1 +xT 1 +wT 1 ≤ (x −w)T 1 +xT 1 +w1 −wT 1 triangle inequality on w
SLIDE 93
Proof of w −x ≤ 4σ(x).
Again: σk(x) = minsupp(z)≤k |x −z|1. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k
16, vT 1 < v1 4 .
Proof of Theorem: T be k largest in magnitude coordinates of x. x −w1 = (x −w)T 1 +(x −w)T 1 ≤ (x −w)T 1 +xT 1 +wT 1 ≤ (x −w)T 1 +xT 1 +w1 −wT 1 triangle inequality on w wT 1 = w1 −wT 1
SLIDE 94
Proof of w −x ≤ 4σ(x).
Again: σk(x) = minsupp(z)≤k |x −z|1. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k
16, vT 1 < v1 4 .
Proof of Theorem: T be k largest in magnitude coordinates of x. x −w1 = (x −w)T 1 +(x −w)T 1 ≤ (x −w)T 1 +xT 1 +wT 1 ≤ (x −w)T 1 +xT 1 +w1 −wT 1 triangle inequality on w wT 1 = w1 −wT 1 ≤ x1.
SLIDE 95
Proof of w −x ≤ 4σ(x).
Again: σk(x) = minsupp(z)≤k |x −z|1. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k
16, vT 1 < v1 4 .
Proof of Theorem: T be k largest in magnitude coordinates of x. x −w1 = (x −w)T 1 +(x −w)T 1 ≤ (x −w)T 1 +xT 1 +wT 1 ≤ (x −w)T 1 +xT 1 +w1 −wT 1 triangle inequality on w wT 1 = w1 −wT 1 ≤ x1. x −w1 ≤ (x −w)T 1 +xT 1 +x1 −wT 1.
SLIDE 96
Proof of w −x ≤ 4σ(x).
Again: σk(x) = minsupp(z)≤k |x −z|1. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k
16, vT 1 < v1 4 .
Proof of Theorem: T be k largest in magnitude coordinates of x. x −w1 = (x −w)T 1 +(x −w)T 1 ≤ (x −w)T 1 +xT 1 +wT 1 ≤ (x −w)T 1 +xT 1 +w1 −wT 1 triangle inequality on w wT 1 = w1 −wT 1 ≤ x1. x −w1 ≤ (x −w)T 1 +xT 1 +x1 −wT 1. (∗) = 2xT 1 +xT −wT 1
SLIDE 97
Proof of w −x ≤ 4σ(x).
Again: σk(x) = minsupp(z)≤k |x −z|1. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k
16, vT 1 < v1 4 .
Proof of Theorem: T be k largest in magnitude coordinates of x. x −w1 = (x −w)T 1 +(x −w)T 1 ≤ (x −w)T 1 +xT 1 +wT 1 ≤ (x −w)T 1 +xT 1 +w1 −wT 1 triangle inequality on w wT 1 = w1 −wT 1 ≤ x1. x −w1 ≤ (x −w)T 1 +xT 1 +x1 −wT 1. (∗) = 2xT 1 +xT −wT 1 ≤ 2xT 1 +xT −wT 1
SLIDE 98
Proof of w −x ≤ 4σ(x).
Again: σk(x) = minsupp(z)≤k |x −z|1. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k
16, vT 1 < v1 4 .
Proof of Theorem: T be k largest in magnitude coordinates of x. x −w1 = (x −w)T 1 +(x −w)T 1 ≤ (x −w)T 1 +xT 1 +wT 1 ≤ (x −w)T 1 +xT 1 +w1 −wT 1 triangle inequality on w wT 1 = w1 −wT 1 ≤ x1. x −w1 ≤ (x −w)T 1 +xT 1 +x1 −wT 1. (∗) = 2xT 1 +xT −wT 1 ≤ 2xT 1 +xT −wT 1 x −w1 ≤ 2(x −w)T 1 +2xT 1
SLIDE 99
Proof of w −x ≤ 4σ(x).
Again: σk(x) = minsupp(z)≤k |x −z|1. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k
16, vT 1 < v1 4 .
Proof of Theorem: T be k largest in magnitude coordinates of x. x −w1 = (x −w)T 1 +(x −w)T 1 ≤ (x −w)T 1 +xT 1 +wT 1 ≤ (x −w)T 1 +xT 1 +w1 −wT 1 triangle inequality on w wT 1 = w1 −wT 1 ≤ x1. x −w1 ≤ (x −w)T 1 +xT 1 +x1 −wT 1. (∗) = 2xT 1 +xT −wT 1 ≤ 2xT 1 +xT −wT 1 x −w1 ≤ 2(x −w)T 1 +2xT 1 ≤ 2 (x−w)
4
+2σ(x)
SLIDE 100
Proof of w −x ≤ 4σ(x).
Again: σk(x) = minsupp(z)≤k |x −z|1. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k
16, vT 1 < v1 4 .
Proof of Theorem: T be k largest in magnitude coordinates of x. x −w1 = (x −w)T 1 +(x −w)T 1 ≤ (x −w)T 1 +xT 1 +wT 1 ≤ (x −w)T 1 +xT 1 +w1 −wT 1 triangle inequality on w wT 1 = w1 −wT 1 ≤ x1. x −w1 ≤ (x −w)T 1 +xT 1 +x1 −wT 1. (∗) = 2xT 1 +xT −wT 1 ≤ 2xT 1 +xT −wT 1 x −w1 ≤ 2(x −w)T 1 +2xT 1 ≤ 2 (x−w)
4
+2σ(x)
SLIDE 101
Proof of w −x ≤ 4σ(x).
Again: σk(x) = minsupp(z)≤k |x −z|1. Lemma: For v ∈ ker(A), T ⊂ [n], |T| ≤ k
16, vT 1 < v1 4 .
Proof of Theorem: T be k largest in magnitude coordinates of x. x −w1 = (x −w)T 1 +(x −w)T 1 ≤ (x −w)T 1 +xT 1 +wT 1 ≤ (x −w)T 1 +xT 1 +w1 −wT 1 triangle inequality on w wT 1 = w1 −wT 1 ≤ x1. x −w1 ≤ (x −w)T 1 +xT 1 +x1 −wT 1. (∗) = 2xT 1 +xT −wT 1 ≤ 2xT 1 +xT −wT 1 x −w1 ≤ 2(x −w)T 1 +2xT 1 ≤ 2 (x−w)
4
+2σ(x) = ⇒ x −w1 ≤ 4σ(x).
SLIDE 102
Almost Euclidean Matrices Proof.
Theorem:
For a random ±1, d ×n matrix, and for any x in with Ax some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
xT 2 <
√ 1 √ 16k xT 1. (∗)
SLIDE 103 Almost Euclidean Matrices Proof.
Theorem:
For a random ±1, d ×n matrix, and for any x in with Ax some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
xT 2 <
√ 1 √ 16k xT 1. (∗)
Idea in GF(2): Random dot product is 0 with probability 1/2. All r rows 0: (1/2)r. Union bound over n
k
SLIDE 104 Almost Euclidean Matrices Proof.
Theorem:
For a random ±1, d ×n matrix, and for any x in with Ax some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
xT 2 <
√ 1 √ 16k xT 1. (∗)
Idea in GF(2): Random dot product is 0 with probability 1/2. All r rows 0: (1/2)r. Union bound over n
k
⇒ log n
k
SLIDE 105 Almost Euclidean Matrices Proof.
Theorem:
For a random ±1, d ×n matrix, and for any x in with Ax some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
xT 2 <
√ 1 √ 16k xT 1. (∗)
Idea in GF(2): Random dot product is 0 with probability 1/2. All r rows 0: (1/2)r. Union bound over n
k
⇒ log n
k
Too many vectors. Real proof is fancy.
SLIDE 106 Almost Euclidean Matrices Proof.
Theorem:
For a random ±1, d ×n matrix, and for any x in with Ax some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
xT 2 <
√ 1 √ 16k xT 1. (∗)
Idea in GF(2): Random dot product is 0 with probability 1/2. All r rows 0: (1/2)r. Union bound over n
k
⇒ log n
k
Too many vectors. Real proof is fancy. Discusses distribution of X ·v for a vector v
SLIDE 107 Almost Euclidean Matrices Proof.
Theorem:
For a random ±1, d ×n matrix, and for any x in with Ax some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
xT 2 <
√ 1 √ 16k xT 1. (∗)
Idea in GF(2): Random dot product is 0 with probability 1/2. All r rows 0: (1/2)r. Union bound over n
k
⇒ log n
k
Too many vectors. Real proof is fancy. Discusses distribution of X ·v for a vector v and random ±1 vector X
SLIDE 108 Almost Euclidean Matrices Proof.
Theorem:
For a random ±1, d ×n matrix, and for any x in with Ax some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
xT 2 <
√ 1 √ 16k xT 1. (∗)
Idea in GF(2): Random dot product is 0 with probability 1/2. All r rows 0: (1/2)r. Union bound over n
k
⇒ log n
k
Too many vectors. Real proof is fancy. Discusses distribution of X ·v for a vector v and random ±1 vector X Poor Man’s proof:
SLIDE 109 Almost Euclidean Matrices Proof.
Theorem:
For a random ±1, d ×n matrix, and for any x in with Ax some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
xT 2 <
√ 1 √ 16k xT 1. (∗)
Idea in GF(2): Random dot product is 0 with probability 1/2. All r rows 0: (1/2)r. Union bound over n
k
⇒ log n
k
Too many vectors. Real proof is fancy. Discusses distribution of X ·v for a vector v and random ±1 vector X Poor Man’s proof: Group coordinates of v until groups of same size.
SLIDE 110 Almost Euclidean Matrices Proof.
Theorem:
For a random ±1, d ×n matrix, and for any x in with Ax some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
xT 2 <
√ 1 √ 16k xT 1. (∗)
Idea in GF(2): Random dot product is 0 with probability 1/2. All r rows 0: (1/2)r. Union bound over n
k
⇒ log n
k
Too many vectors. Real proof is fancy. Discusses distribution of X ·v for a vector v and random ±1 vector X Poor Man’s proof: Group coordinates of v until groups of same size. ni in each group.
SLIDE 111 Almost Euclidean Matrices Proof.
Theorem:
For a random ±1, d ×n matrix, and for any x in with Ax some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
xT 2 <
√ 1 √ 16k xT 1. (∗)
Idea in GF(2): Random dot product is 0 with probability 1/2. All r rows 0: (1/2)r. Union bound over n
k
⇒ log n
k
Too many vectors. Real proof is fancy. Discusses distribution of X ·v for a vector v and random ±1 vector X Poor Man’s proof: Group coordinates of v until groups of same size. ni in each group. Deviation in group ≤ √ni/2 in each group is less than 1/2.
SLIDE 112 Almost Euclidean Matrices Proof.
Theorem:
For a random ±1, d ×n matrix, and for any x in with Ax some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
xT 2 <
√ 1 √ 16k xT 1. (∗)
Idea in GF(2): Random dot product is 0 with probability 1/2. All r rows 0: (1/2)r. Union bound over n
k
⇒ log n
k
Too many vectors. Real proof is fancy. Discusses distribution of X ·v for a vector v and random ±1 vector X Poor Man’s proof: Group coordinates of v until groups of same size. ni in each group. Deviation in group ≤ √ni/2 in each group is less than 1/2. Probability groups cancel is small.
SLIDE 113 Almost Euclidean Matrices Proof.
Theorem:
For a random ±1, d ×n matrix, and for any x in with Ax some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
xT 2 <
√ 1 √ 16k xT 1. (∗)
Idea in GF(2): Random dot product is 0 with probability 1/2. All r rows 0: (1/2)r. Union bound over n
k
⇒ log n
k
Too many vectors. Real proof is fancy. Discusses distribution of X ·v for a vector v and random ±1 vector X Poor Man’s proof: Group coordinates of v until groups of same size. ni in each group. Deviation in group ≤ √ni/2 in each group is less than 1/2. Probability groups cancel is small. Lots of rows. So, norm is good on average for each group.
SLIDE 114 Almost Euclidean Matrices Proof.
Theorem:
For a random ±1, d ×n matrix, and for any x in with Ax some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
xT 2 <
√ 1 √ 16k xT 1. (∗)
Idea in GF(2): Random dot product is 0 with probability 1/2. All r rows 0: (1/2)r. Union bound over n
k
⇒ log n
k
Too many vectors. Real proof is fancy. Discusses distribution of X ·v for a vector v and random ±1 vector X Poor Man’s proof: Group coordinates of v until groups of same size. ni in each group. Deviation in group ≤ √ni/2 in each group is less than 1/2. Probability groups cancel is small. Lots of rows. So, norm is good on average for each group. “Few” vectors with most of mass in small set of coordinates.
SLIDE 115 Almost Euclidean Matrices Proof.
Theorem:
For a random ±1, d ×n matrix, and for any x in with Ax some d = Ω(k log n
k ) rows, has for any T ⊂ [n] that
xT 2 <
√ 1 √ 16k xT 1. (∗)
Idea in GF(2): Random dot product is 0 with probability 1/2. All r rows 0: (1/2)r. Union bound over n
k
⇒ log n
k
Too many vectors. Real proof is fancy. Discusses distribution of X ·v for a vector v and random ±1 vector X Poor Man’s proof: Group coordinates of v until groups of same size. ni in each group. Deviation in group ≤ √ni/2 in each group is less than 1/2. Probability groups cancel is small. Lots of rows. So, norm is good on average for each group. “Few” vectors with most of mass in small set of coordinates. Union bound over those.
SLIDE 116
Credits
Moitra, MIT,6.854. Roughgarden, CS168, Stanford.
SLIDE 117
Credits
Moitra, MIT,6.854. Roughgarden, CS168, Stanford. See Jame Lee, TCS Blog, May 2008 for proof of Almost Euclidean Nature of random subspaces.
SLIDE 118
Possible Topics.
TODO: Long tailed distributions.
SLIDE 119
Possible Topics.
TODO: Long tailed distributions. Interior Point Algorithms.
SLIDE 120
Possible Topics.
TODO: Long tailed distributions. Interior Point Algorithms. Matrix Concentration/Matrix Experts/Semidefinite Programs.
SLIDE 121
Possible Topics.
TODO: Long tailed distributions. Interior Point Algorithms. Matrix Concentration/Matrix Experts/Semidefinite Programs. Coding Theory: Low Density Parity Check Codes or Expander codes.
SLIDE 122 Possible Topics.
TODO: Long tailed distributions. Interior Point Algorithms. Matrix Concentration/Matrix Experts/Semidefinite Programs. Coding Theory: Low Density Parity Check Codes or Expander codes.
- Auctions. Mechanism Design.
SLIDE 123 Possible Topics.
TODO: Long tailed distributions. Interior Point Algorithms. Matrix Concentration/Matrix Experts/Semidefinite Programs. Coding Theory: Low Density Parity Check Codes or Expander codes.
- Auctions. Mechanism Design.