sparse solutions of underdetermined linear equations by
play

Sparse Solutions of Underdetermined Linear Equations by Linear - PowerPoint PPT Presentation

Sparse Solutions of Underdetermined Linear Equations by Linear Programming David Donoho & Jared Tanner Stanford University, Department of Statistics University of Utah, Department of Mathematics Arizona State University: March 6 th 2006


  1. Sparse Solutions of Underdetermined Linear Equations by Linear Programming David Donoho & Jared Tanner Stanford University, Department of Statistics University of Utah, Department of Mathematics Arizona State University: March 6 th 2006

  2. Underdetermined systems, dictionary perspective ◮ Underdetermined system, infinite number of solutions A ∈ R d × n , Ax = b , d < n

  3. Underdetermined systems, dictionary perspective ◮ Underdetermined system, infinite number of solutions A ∈ R d × n , Ax = b , d < n ◮ Least squares solution via “canonical dual” ( A T A ) − 1 A T • Linear reconstruction, not signal adaptive • Solution vector full, n nonzero elements in x

  4. Underdetermined systems, dictionary perspective ◮ Underdetermined system, infinite number of solutions A ∈ R d × n , Ax = b , d < n ◮ Least squares solution via “canonical dual” ( A T A ) − 1 A T • Linear reconstruction, not signal adaptive • Solution vector full, n nonzero elements in x ◮ Eschew redundancy, find simple model of data from A

  5. Underdetermined systems, dictionary perspective ◮ Underdetermined system, infinite number of solutions A ∈ R d × n , Ax = b , d < n ◮ Least squares solution via “canonical dual” ( A T A ) − 1 A T • Linear reconstruction, not signal adaptive • Solution vector full, n nonzero elements in x ◮ Eschew redundancy, find simple model of data from A ◮ Seek sparsest solution, � x � ℓ 0 := # nonzero elements min � x � ℓ 0 subject to Ax = b

  6. Underdetermined systems, dictionary perspective ◮ Underdetermined system, infinite number of solutions A ∈ R d × n , Ax = b , d < n ◮ Least squares solution via “canonical dual” ( A T A ) − 1 A T • Linear reconstruction, not signal adaptive • Solution vector full, n nonzero elements in x ◮ Eschew redundancy, find simple model of data from A ◮ Seek sparsest solution, � x � ℓ 0 := # nonzero elements min � x � ℓ 0 subject to Ax = b ◮ Combinatorial cost for naive approach

  7. Underdetermined systems, dictionary perspective ◮ Underdetermined system, infinite number of solutions A ∈ R d × n , Ax = b , d < n ◮ Least squares solution via “canonical dual” ( A T A ) − 1 A T • Linear reconstruction, not signal adaptive • Solution vector full, n nonzero elements in x ◮ Eschew redundancy, find simple model of data from A ◮ Seek sparsest solution, � x � ℓ 0 := # nonzero elements min � x � ℓ 0 subject to Ax = b ◮ Combinatorial cost for naive approach ◮ Efficient nonlinear (signal adaptive) methods • Greedy (local) and Basis Pursuit (global)

  8. Greedy [Temlyakov, DeVore, Tropp, ...] ◮ Orthogonal Matching Pursuit: initial r = b , ˜ A = [] while r � = 0 ℓ ∞ A T r =: a T max j r

  9. Greedy [Temlyakov, DeVore, Tropp, ...] ◮ Orthogonal Matching Pursuit: initial r = b , ˜ A = [] while r � = 0 A = [˜ ˜ ℓ ∞ A T r =: a T max j r A a j ]

  10. Greedy [Temlyakov, DeVore, Tropp, ...] ◮ Orthogonal Matching Pursuit: initial r = b , ˜ A = [] while r � = 0 A T ˜ A = [˜ ˜ r = b − A (˜ A ) − 1 ˜ ℓ ∞ A T r =: a T A T b max j r A a j ]

  11. Greedy [Temlyakov, DeVore, Tropp, ...] ◮ Orthogonal Matching Pursuit: initial r = b , ˜ A = [] while r � = 0 A T ˜ A = [˜ ˜ r = b − A (˜ A ) − 1 ˜ ℓ ∞ A T r =: a T A T b max j r A a j ] A T ˜ ◮ Nonlinear selection of basis, x = (˜ A ) − 1 ˜ A T b ; � x � ℓ 0 ≤ d

  12. Greedy [Temlyakov, DeVore, Tropp, ...] ◮ Orthogonal Matching Pursuit: initial r = b , ˜ A = [] while r � = 0 A T ˜ A = [˜ ˜ r = b − A (˜ A ) − 1 ˜ ℓ ∞ A T r =: a T A T b max j r A a j ] A T ˜ ◮ Nonlinear selection of basis, x = (˜ A ) − 1 ˜ A T b ; � x � ℓ 0 ≤ d ◮ Highly redundant dictionary often give fast decay of residual

  13. Greedy [Temlyakov, DeVore, Tropp, ...] ◮ Orthogonal Matching Pursuit: initial r = b , ˜ A = [] while r � = 0 A T ˜ A = [˜ ˜ r = b − A (˜ A ) − 1 ˜ ℓ ∞ A T r =: a T A T b max j r A a j ] A T ˜ ◮ Nonlinear selection of basis, x = (˜ A ) − 1 ˜ A T b ; � x � ℓ 0 ≤ d ◮ Highly redundant dictionary often give fast decay of residual ◮ Recovery of sparsest solution? • examples of arbitrary sub-optimality for a fixed dictionary A [Temlyakov, DeVore, S. Chen, Tropp, . . . ]

  14. Greedy [Temlyakov, DeVore, Tropp, ...] ◮ Orthogonal Matching Pursuit: initial r = b , ˜ A = [] while r � = 0 A T ˜ A = [˜ ˜ r = b − A (˜ A ) − 1 ˜ ℓ ∞ A T r =: a T A T b max j r A a j ] A T ˜ ◮ Nonlinear selection of basis, x = (˜ A ) − 1 ˜ A T b ; � x � ℓ 0 ≤ d ◮ Highly redundant dictionary often give fast decay of residual ◮ Recovery of sparsest solution? • examples of arbitrary sub-optimality for a fixed dictionary A [Temlyakov, DeVore, S. Chen, Tropp, . . . ] • residual nonzero for steps < d , irregardless of sparsity [Chen]

  15. Greedy [Temlyakov, DeVore, Tropp, ...] ◮ Orthogonal Matching Pursuit: initial r = b , ˜ A = [] while r � = 0 A T ˜ A = [˜ ˜ r = b − A (˜ A ) − 1 ˜ ℓ ∞ A T r =: a T A T b max j r A a j ] A T ˜ ◮ Nonlinear selection of basis, x = (˜ A ) − 1 ˜ A T b ; � x � ℓ 0 ≤ d ◮ Highly redundant dictionary often give fast decay of residual ◮ Recovery of sparsest solution? • examples of arbitrary sub-optimality for a fixed dictionary A [Temlyakov, DeVore, S. Chen, Tropp, . . . ] • residual nonzero for steps < d , irregardless of sparsity [Chen] √ ◮ Recover sparsest if sufficiently sparse, O ( d ) [Tropp]

  16. Greedy [Temlyakov, DeVore, Tropp, ...] ◮ Orthogonal Matching Pursuit: initial r = b , ˜ A = [] while r � = 0 A T ˜ A = [˜ ˜ r = b − A (˜ A ) − 1 ˜ ℓ ∞ A T r =: a T A T b max j r A a j ] A T ˜ ◮ Nonlinear selection of basis, x = (˜ A ) − 1 ˜ A T b ; � x � ℓ 0 ≤ d ◮ Highly redundant dictionary often give fast decay of residual ◮ Recovery of sparsest solution? • examples of arbitrary sub-optimality for a fixed dictionary A [Temlyakov, DeVore, S. Chen, Tropp, . . . ] • residual nonzero for steps < d , irregardless of sparsity [Chen] √ ◮ Recover sparsest if sufficiently sparse, O ( d ) [Tropp] ◮ More sophisticated variants; weak greedy, swapping, etc ...

  17. Greedy [Temlyakov, DeVore, Tropp, ...] ◮ Orthogonal Matching Pursuit: initial r = b , ˜ A = [] while r � = 0 A T ˜ A = [˜ ˜ r = b − A (˜ A ) − 1 ˜ ℓ ∞ A T r =: a T A T b max j r A a j ] A T ˜ ◮ Nonlinear selection of basis, x = (˜ A ) − 1 ˜ A T b ; � x � ℓ 0 ≤ d ◮ Highly redundant dictionary often give fast decay of residual ◮ Recovery of sparsest solution? • examples of arbitrary sub-optimality for a fixed dictionary A [Temlyakov, DeVore, S. Chen, Tropp, . . . ] • residual nonzero for steps < d , irregardless of sparsity [Chen] √ ◮ Recover sparsest if sufficiently sparse, O ( d ) [Tropp] ◮ More sophisticated variants; weak greedy, swapping, etc ... ◮ More about OMP for random sampling later

  18. Basis Pursuit ◮ Rather than solve ℓ 0 (combinatorial), solve ℓ 1 , (use LP) min � x � ℓ 1 subject to Ax = b • Global basis selection rather than greedy local selection

  19. Basis Pursuit ◮ Rather than solve ℓ 0 (combinatorial), solve ℓ 1 , (use LP) min � x � ℓ 1 subject to Ax = b • Global basis selection rather than greedy local selection ◮ Example, A = [ A 1 A 2 ] two ONB with coherence max ij ( a i , a j ) • If � x � ℓ 0 � . 914(1 + µ − 1 ) then ℓ 1 → ℓ 0 [Elad, Bruckstein] √ • Coherence, µ := max ij ( a i , a j ) ≥ 1 / d , [Candes, Romberg]

  20. Basis Pursuit ◮ Rather than solve ℓ 0 (combinatorial), solve ℓ 1 , (use LP) min � x � ℓ 1 subject to Ax = b • Global basis selection rather than greedy local selection ◮ Example, A = [ A 1 A 2 ] two ONB with coherence max ij ( a i , a j ) • If � x � ℓ 0 � . 914(1 + µ − 1 ) then ℓ 1 → ℓ 0 [Elad, Bruckstein] √ • Coherence, µ := max ij ( a i , a j ) ≥ 1 / d , [Candes, Romberg] √ ◮ Ensures convergence for only the most sparse, O ( d ), signals.

  21. Basis Pursuit ◮ Rather than solve ℓ 0 (combinatorial), solve ℓ 1 , (use LP) min � x � ℓ 1 subject to Ax = b • Global basis selection rather than greedy local selection ◮ Example, A = [ A 1 A 2 ] two ONB with coherence max ij ( a i , a j ) • If � x � ℓ 0 � . 914(1 + µ − 1 ) then ℓ 1 → ℓ 0 [Elad, Bruckstein] √ • Coherence, µ := max ij ( a i , a j ) ≥ 1 / d , [Candes, Romberg] √ ◮ Ensures convergence for only the most sparse, O ( d ), signals. ◮ Examples of failure: Dirac’s Comb [Candes, . . . ]

  22. Basis Pursuit ◮ Rather than solve ℓ 0 (combinatorial), solve ℓ 1 , (use LP) min � x � ℓ 1 subject to Ax = b • Global basis selection rather than greedy local selection ◮ Example, A = [ A 1 A 2 ] two ONB with coherence max ij ( a i , a j ) • If � x � ℓ 0 � . 914(1 + µ − 1 ) then ℓ 1 → ℓ 0 [Elad, Bruckstein] √ • Coherence, µ := max ij ( a i , a j ) ≥ 1 / d , [Candes, Romberg] √ ◮ Ensures convergence for only the most sparse, O ( d ), signals. ◮ Examples of failure: Dirac’s Comb [Candes, . . . ] √ ◮ Is the story over? Can O ( d ) threshold be overcome? yes!

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend