The Sparsity Gap
❦
Joel A. Tropp
Computing & Mathematical Sciences California Institute of Technology jtropp@acm.caltech.edu
Research supported in part by ONR 1
The Sparsity Gap Joel A. Tropp Computing & Mathematical - - PowerPoint PPT Presentation
The Sparsity Gap Joel A. Tropp Computing & Mathematical Sciences California Institute of Technology jtropp@acm.caltech.edu Research supported in part by ONR 1 Introduction The Sparsity Gap (Casazza Birthday Conference, College Park,
❦
Joel A. Tropp
Computing & Mathematical Sciences California Institute of Technology jtropp@acm.caltech.edu
Research supported in part by ONR 1
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 2
Systems of Linear Equations
We consider linear systems of the form m Φ
x = b
Assume that
❧ Φ has dimensions m × N with N > m ❧ Φ has full row rank ❧ The columns of Φ have unit ℓ2 norm
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 3
The Trichotomy Theorem
Theorem 1. For a linear system Φx = b, exactly one of the following situations occurs.
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 4
Regularization via Sparsity
A principled approach to underdetermined systems: min x0 subject to Φx = b (P0) where x0 = # supp(x) = #{j : xj = 0} ❧ When x0 ≤ s, then x is called s-sparse ❧ If Φx = b and x is s-sparse, then x is an s-sparse representation of b ❧ Since Φ has full row-rank, every b has an m-sparse representation ❧ Question: What can we say about sparser representations?
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 5
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 6
Key Insight Sparse representations are well behaved when the matrix Φ is sufficiently nice
(Column submatrices should not be singular and individual columns should not look like sparse signals)
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 7
Quantifying Niceness I
❧ We call Φ a tight frame when ΦΦ∗ = N m · I ❧ Equivalently, the rows of Φ form an orthonormal family (up to scaling) ❧ Observe that N/m is the redundancy of the frame ❧ Tight frames have minimal spectral norm among conformal matrices
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 8
Quantifying Niceness II
❧ The coherence of Φ is the quantity µ = max
j=k |ϕj, ϕk|
❧ Measures the angle between columns ❧ When N ≥ 2m, the coherence satisfies µ 1 √m ❧ Incoherent matrices appear often in signal processing applications
References: [Welch 1974, Mallat–Zhang 1993, Donoho–Huo 2001, Gribonval–Nielsen 2003, Strohmer–Heath 2003]
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 9
Example: Identity + Fourier
1
Impulses Complex Exponentials
A very incoherent tight frame
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 10
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 11
Uncertainty implies Uniqueness
Theorem 2. Suppose that a vector b has two representations: Φx = b = Φy. Then x0 + y0 > µ−1. Corollary 3. Suppose that b = Φx where x0 < 1 2 · µ−1. Then x is the unique solution to (P0). ❧ Very strict requirement since µ−1 √m
References: [Donoho–Stark 1989, Donoho–Huo 2001, Gribonval–Nielsen 2003, Donoho–Elad 2003]
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 12
The Square-Root Threshold
❧ Sparse representations are not necessarily unique past the √m threshold
Example: The Dirac Comb
❧ Consider the Identity + Fourier matrix with m = p2 ❧ There is a vector b that can be written as either p spikes or p sines ❧ By the Poisson summation formula, b(t) =
p−1
δpj(t) = 1 √m
p−1
e−2πipjt/m for t = 0, 1, . . . , m
References: [Donoho–Stark 1989]
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 13
Enter Probability Insight: The bad vectors are atypical
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 14
An Uncertainty Principle for Generic Signals
Theorem 4. [T 2010] Suppose that b = Φx where the nonzero components of x have a continuous distribution. With probability one, the vector b has no representation b = Φy where supp(x) ∩ supp(y) = ∅ unless x0 + y0 > µ−1 x1/2 . Corollary 5. When µ ≤ m−1/2, condition becomes y0 > x0 m x0 − 1
❧ Problem: Some supports could be bad
References: [The Sparsity Gap]
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 15
Enter More Probability Insight: The bad supports are atypical
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 16
A Simple Model for Random Sparse Vectors
Model (M0) for b = Φx The matrix Φ is a unit-norm tight frame of size m × N with coherence µ ≤ c/ log N. The support of x has cardinality s ≤ cm/ log N and is uniformly random. The nonzero entries of x have a continuous distribution.
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 17
Random Submatrices & Sparse Representation
Theorem 6. [T 2006, 2008] Assume all parameters satisfy Model (M0). Draw a uniformly random set S of s columns from Φ, and define the random column submatrix A = ΦS. Then Prob
2
and Prob
∈S A∗ϕn2 < 1
2
References: [Random Subdictionaries, Random Submatrices]
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 18
The Sparsity Gap
Theorem 7. [T 2008, 2010] Let b = Φx be a vector drawn from Model (M0). With probability at least 99.44%, the following statements hold.
b = Φy where supp(x) ∩ supp(y) = ∅ unless y0 > x0
N
References: [Cand` es–Romberg 2006, Random Subdictionaries, Random Submatrices, The Sparsity Gap]
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 19
To learn more...
Web: http://www.acm.caltech.edu/~jtropp E-mail: jtropp@acm.caltech.edu
Relevant papers: ❧ “Conditioning of random subdictionaries,” ACHA, 2008 ❧ “Norms of random submatrices,” CRAS, 2008 ❧ “Spikes and sines,” JFAA, 2008 ❧ “The sparsity gap,” Proc. CISS, 2010
The Sparsity Gap (Casazza Birthday Conference, College Park, 20 May 2010) 20