com pressive sensing
play

Com pressive Sensing and Applications Volkan Cevher - PowerPoint PPT Presentation

Com pressive Sensing and Applications Volkan Cevher volkan@rice.edu Rice University Acknowledgements Rice DSP Group (Slides) Richard Baraniuk Mark Davenport, Marco Duarte, Chinmay Hegde, Jason Laska, Shri


  1. Com pressive Sensing and Applications Volkan Cevher volkan@rice.edu Rice University

  2. Acknowledgements • Rice DSP Group (Slides) – Richard Baraniuk � Mark Davenport, � Marco Duarte, � Chinmay Hegde, � Jason Laska, � Shri Sarvotham, � Mona Sheikh � Stephen Schnelle… – Mike Wakin, Justin Romberg, Petros Boufounos, Dror Baron

  3. Outline • Introduction to Compressive Sensing (CS) – motivation – basic concepts • CS Theoretical Foundation – geometry of sparse and compressible signals – coded acquisition – restricted isometry property (RIP) – signal recovery • CS in Action • Summary

  4. Sensing

  5. Digital Revolution

  6. Pressure is on Digital Sensors • Success of digital data acquisition is placing increasing pressure on signal/ image processing hardware and software to support higher resolution / denser sam pling » ADCs, cameras, imaging systems, microarrays, … large num bers of sensors » image data bases, camera arrays, distributed wireless sensor networks, … increasing num bers of m odalities » acoustic, RF, visual, IR, UV, x-ray, gamma ray, …

  7. Pressure is on Digital Sensors • Success of digital data acquisition is placing increasing pressure on signal/ image processing hardware and software to support higher resolution / denser sam pling » ADCs, cameras, imaging systems, microarrays, … x large num bers of sensors » image data bases, camera arrays, distributed wireless sensor networks, … x increasing num bers of m odalities » acoustic, RF, visual, IR, UV deluge of data » how to acquire , store , fuse , process efficiently?

  8. Digital Data Acquisition • Foundation: Shannon/ Nyquist sampling theorem “if you sample densely enough (at the Nyquist rate), you can perfectly reconstruct the original analog data” time space

  9. Sensing by Sampling • Long-established paradigm for digital data acquisition – uniformly sam ple data at Nyquist rate (2x Fourier bandwidth) sample

  10. Sensing by Sampling • Long-established paradigm for digital data acquisition – uniformly sam ple data at Nyquist rate (2x Fourier bandwidth) too sample m uch data!

  11. Sensing by Sampling • Long-established paradigm for digital data acquisition – uniformly sam ple data at Nyquist rate (2x Fourier bandwidth) – com press data transmit/ store com press sample JPEG JPEG2 0 0 0 … receive decompress

  12. Sparsity / Compressibility large pixels wavelet coefficients (blue = 0) frequency wideband large signal Gabor (TF) samples coefficients time

  13. Sample / Compress • Long-established paradigm for digital data acquisition – uniformly sam ple data at Nyquist rate – com press data transmit/ store sample com press sparse / com pressible wavelet transform receive decompress

  14. What’s Wrong with this Picture? • W hy go to all the w ork to acquire N sam ples only to discard all but K pieces of data? transmit/ store sample com press sparse / com pressible wavelet transform receive decompress

  15. What’s Wrong with this Picture? linear processing nonlinear processing linear signal model nonlinear signal model (bandlimited subspace) (union of subspaces) transmit/ store sample com press sparse / com pressible wavelet transform receive decompress

  16. Compressive Sensing • Directly acquire “ com pressed ” data • Replace samples by more general “measurements” transmit/ store com pressive sensing receive reconstruct

  17. Com pressive Sensing Theory I Geom etrical Perspective

  18. Sampling • Signal is - sparse in basis/ dictionary – WLOG assume sparse in space domain sparse signal nonzero entries

  19. Sampling • Signal is - sparse in basis/ dictionary – WLOG assume sparse in space domain • Sam ples sparse measurements signal nonzero entries

  20. Compressive Sampling • When data is sparse/ compressible, can directly acquire a condensed representation with no/ little information loss through linear dim ensionality reduction sparse measurements signal nonzero entries

  21. How Can It Work? • Projection not full rank … … and so loses inform ation in general • Ex: Infinitely many ’s map to the same

  22. How Can It Work? • Projection not full rank… columns … and so loses information in general • But we are only interested in sparse vectors

  23. How Can It Work? • Projection not full rank… columns … and so loses information in general • But we are only interested in sparse vectors • is effectively M x K

  24. How Can It Work? • Projection not full rank… columns … and so loses information in general • But we are only interested in sparse vectors • Design so that each of its M x K submatrices are full rank

  25. How Can It Work? columns • Goal: Design so that its M x2 K submatrices are full rank – difference between two K -sparse vectors is 2 K sparse in general – preserve information in K -sparse signals – Restricted I som etry Property (RIP) of order 2 K

  26. Unfortunately… columns • Goal: Design so that its M x2 K submatrices are full rank (Restricted Isometry Property – RIP) • Unfortunately, a combinatorial, NP-com plete design problem

  27. Insight from the 80’s [ Kashin, Gluskin] • Draw at random – iid Gaussian – iid Bernoulli … columns • Then has the RIP with high probability as long as – M x2 K submatrices are full rank – stable embedding for sparse signals – extends to compressible signals in balls

  28. Compressive Data Acquisition • Measurements = random linear com binations of the entries of • WHP does not distort structure of sparse signals – no information loss sparse measurements signal nonzero entries

  29. CS Signal Recovery • Goal : Recover signal from measurements • Challenge : Random projection not full rank (ill-posed inverse problem) • Solution : Exploit the sparse/ compressible geom etry of acquired signal

  30. Concise Signal Structure • Sparse signal: only K out of N coordinates nonzero sorted index

  31. Concise Signal Structure • Sparse signal: only K out of N coordinates nonzero – model: union of K -dimensional subspaces aligned w/ coordinate axes sorted index

  32. Concise Signal Structure • Sparse signal: only K out of N coordinates nonzero – model: union of K -dimensional subspaces • Com pressible signal: sorted coordinates decay rapidly to zero power-law decay sorted index

  33. Concise Signal Structure • Sparse signal: only K out of N coordinates nonzero – model: union of K -dimensional subspaces • Com pressible signal: sorted coordinates decay rapidly to zero – model: ball: power-law decay sorted index

  34. CS Signal Recovery • Random projection not full rank • Recovery problem: given find • Null space • So search in null space for the “best” according to some criterion – ex: least squares

  35. CS Signal Recovery • Recovery: given find (sparse) (ill-posed inverse problem) • fast pseudoinverse

  36. CS Signal Recovery • Recovery: given find (sparse) (ill-posed inverse problem) • fast, w rong pseudoinverse

  37. Why Doesn’t Work for signals sparse in the space/ tim e dom ain least squares, minimum solution null space of is almost never sparse translated to (random angle)

  38. CS Signal Recovery • Reconstruction/ decoding: given find (ill-posed inverse problem) • fast, wrong • num ber of nonzero entries “find sparsest in translated nullspace”

  39. CS Signal Recovery • Reconstruction/ decoding: given find (ill-posed inverse problem) • fast, wrong correct: • only M = 2 K measurements required to reconstruct number of K -sparse signal nonzero entries

  40. CS Signal Recovery • Reconstruction/ decoding: given find (ill-posed inverse problem) • fast, wrong correct: • only M = 2 K measurements required to reconstruct number of K -sparse signal nonzero entries slow : NP-complete algorithm

  41. CS Signal Recovery • Recovery: given find (sparse) (ill-posed inverse problem) • fast, wrong • correct, slow correct, efficient • m ild oversam pling [ Candes, Romberg, Tao; Donoho] linear program number of measurements required

  42. Why Works for signals sparse in the space/ tim e dom ain minimum solution = sparsest solution (with high probability) if

  43. Universality • Random measurements can be used for signals sparse in any basis

  44. Universality • Random measurements can be used for signals sparse in any basis

  45. Universality • Random measurements can be used for signals sparse in any basis sparse coefficient vector nonzero entries

  46. Compressive Sensing • Directly acquire “ com pressed ” data • Replace N samples by M random projections transmit/ store random m easurem ents … receive linear pgm

  47. Com pressive Sensing Theory I I Stable Em bedding

  48. Johnson-Lindenstrauss Lemma • JL Lemma: random projection stably embeds a cloud of Q points whp provided Q points • Proved via concentration inequality • Same techniques link JLL to RIP [ Baraniuk, Davenport, DeVore, Wakin, Constructive Approximation , 2008]

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend