Com pressive Sensing and Applications Volkan Cevher - - PowerPoint PPT Presentation

com pressive sensing
SMART_READER_LITE
LIVE PREVIEW

Com pressive Sensing and Applications Volkan Cevher - - PowerPoint PPT Presentation

Com pressive Sensing and Applications Volkan Cevher volkan@rice.edu Rice University Acknowledgements Rice DSP Group (Slides) Richard Baraniuk Mark Davenport, Marco Duarte, Chinmay Hegde, Jason Laska, Shri


slide-1
SLIDE 1

Volkan Cevher

volkan@rice.edu Rice University

Com pressive Sensing

and Applications

slide-2
SLIDE 2

Acknowledgements

  • Rice DSP Group (Slides)

– Richard Baraniuk Mark Davenport, Marco Duarte, Chinmay Hegde, Jason Laska, Shri Sarvotham, Mona Sheikh Stephen Schnelle… – Mike Wakin, Justin Romberg, Petros Boufounos, Dror Baron

slide-3
SLIDE 3

Outline

  • Introduction to Compressive Sensing (CS)

– motivation – basic concepts

  • CS Theoretical Foundation

– geometry of sparse and compressible signals – coded acquisition – restricted isometry property (RIP) – signal recovery

  • CS in Action
  • Summary
slide-4
SLIDE 4

Sensing

slide-5
SLIDE 5

Digital Revolution

slide-6
SLIDE 6

Pressure is on Digital Sensors

  • Success of digital data acquisition is placing increasing pressure
  • n signal/ image processing hardware and software to support

higher resolution / denser sam pling

» ADCs, cameras, imaging systems, microarrays, …

large num bers of sensors

» image data bases, camera arrays, distributed wireless sensor networks, …

increasing num bers of m odalities

» acoustic, RF, visual, IR, UV, x-ray, gamma ray, …

slide-7
SLIDE 7

Pressure is on Digital Sensors

  • Success of digital data acquisition is placing increasing pressure
  • n signal/ image processing hardware and software to support

higher resolution / denser sam pling

» ADCs, cameras, imaging systems, microarrays, …

x large num bers of sensors

» image data bases, camera arrays, distributed wireless sensor networks, …

x increasing num bers of m odalities

» acoustic, RF, visual, IR, UV

deluge of data

» how to acquire, store, fuse, process efficiently?

slide-8
SLIDE 8

Digital Data Acquisition

  • Foundation: Shannon/ Nyquist sampling theorem

“if you sample densely enough (at the Nyquist rate), you can perfectly reconstruct the original analog data”

time space

slide-9
SLIDE 9

Sensing by Sampling

  • Long-established paradigm for digital data acquisition

– uniformly sam ple data at Nyquist rate (2x Fourier bandwidth)

sample

slide-10
SLIDE 10

Sensing by Sampling

  • Long-established paradigm for digital data acquisition

– uniformly sam ple data at Nyquist rate (2x Fourier bandwidth)

sample too m uch data!

slide-11
SLIDE 11

Sensing by Sampling

  • Long-established paradigm for digital data acquisition

– uniformly sam ple data at Nyquist rate (2x Fourier bandwidth) – com press data

com press transmit/ store receive decompress sample JPEG JPEG2 0 0 0 …

slide-12
SLIDE 12

Sparsity / Compressibility

pixels large wavelet coefficients

(blue = 0)

wideband signal samples large Gabor (TF) coefficients

time frequency

slide-13
SLIDE 13

Sample / Compress

com press transmit/ store receive decompress sample

sparse / com pressible wavelet transform

  • Long-established paradigm for digital data acquisition

– uniformly sam ple data at Nyquist rate – com press data

slide-14
SLIDE 14

What’s Wrong with this Picture?

  • W hy go to all the w ork to acquire

N sam ples only to discard all but K pieces of data? com press transmit/ store receive decompress sample

sparse / com pressible wavelet transform

slide-15
SLIDE 15

What’s Wrong with this Picture?

linear processing linear signal model (bandlimited subspace)

com press transmit/ store receive decompress sample

sparse / com pressible wavelet transform

nonlinear processing nonlinear signal model (union of subspaces)

slide-16
SLIDE 16

Compressive Sensing

  • Directly acquire “com pressed” data
  • Replace samples by more general “measurements”

com pressive sensing transmit/ store receive reconstruct

slide-17
SLIDE 17

Com pressive Sensing Theory I Geom etrical Perspective

slide-18
SLIDE 18
  • Signal is -sparse in basis/ dictionary

– WLOG assume sparse in space domain

Sampling

sparse signal nonzero entries

slide-19
SLIDE 19
  • Signal is -sparse in basis/ dictionary

– WLOG assume sparse in space domain

  • Sam ples

sparse signal nonzero entries measurements

Sampling

slide-20
SLIDE 20

Compressive Sampling

  • When data is sparse/ compressible, can directly

acquire a condensed representation with no/ little information loss through linear dim ensionality reduction

measurements sparse signal nonzero entries

slide-21
SLIDE 21

How Can It Work?

  • Projection

not full rank… … and so loses inform ation in general

  • Ex: Infinitely many ’s map to the same
slide-22
SLIDE 22

How Can It Work?

  • Projection

not full rank… … and so loses information in general

  • But we are only interested in sparse vectors

columns

slide-23
SLIDE 23

How Can It Work?

  • Projection

not full rank… … and so loses information in general

  • But we are only interested in sparse vectors
  • is effectively MxK

columns

slide-24
SLIDE 24

How Can It Work?

  • Projection

not full rank… … and so loses information in general

  • But we are only interested in sparse vectors
  • Design

so that each of its MxK submatrices are full rank

columns

slide-25
SLIDE 25

How Can It Work?

  • Goal: Design so that its

Mx2K submatrices are full rank

– difference between two K-sparse vectors is 2K sparse in general – preserve information in K-sparse signals – Restricted I som etry Property (RIP) of order 2K

columns

slide-26
SLIDE 26

Unfortunately…

  • Goal: Design so that its

Mx2K submatrices are full rank

(Restricted Isometry Property – RIP)

  • Unfortunately, a combinatorial,

NP-com plete design problem

columns

slide-27
SLIDE 27

Insight from the 80’s [ Kashin, Gluskin]

  • Draw at random

– iid Gaussian – iid Bernoulli …

  • Then has the RIP with high probability

as long as

– Mx2K submatrices are full rank – stable embedding for sparse signals – extends to compressible signals in balls

columns

slide-28
SLIDE 28

Compressive Data Acquisition

  • Measurements = random linear com binations
  • f the entries of
  • WHP does not distort structure of sparse signals

– no information loss

measurements sparse signal nonzero entries

slide-29
SLIDE 29

CS Signal Recovery

  • Goal: Recover signal

from measurements

  • Challenge: Random

projection not full rank (ill-posed inverse problem)

  • Solution: Exploit the sparse/ compressible

geom etry of acquired signal

slide-30
SLIDE 30
  • Sparse signal:
  • nly K out of N

coordinates nonzero

Concise Signal Structure

sorted index

slide-31
SLIDE 31
  • Sparse signal:
  • nly K out of N

coordinates nonzero

– model: union of K-dimensional subspaces aligned w/ coordinate axes

Concise Signal Structure

sorted index

slide-32
SLIDE 32
  • Sparse signal:
  • nly K out of N

coordinates nonzero

– model: union of K-dimensional subspaces

  • Com pressible signal:

sorted coordinates decay rapidly to zero

Concise Signal Structure

sorted index power-law decay

slide-33
SLIDE 33
  • Sparse signal:
  • nly K out of N

coordinates nonzero

– model: union of K-dimensional subspaces

  • Com pressible signal:

sorted coordinates decay rapidly to zero

– model: ball:

Concise Signal Structure

sorted index power-law decay

slide-34
SLIDE 34

CS Signal Recovery

  • Random projection

not full rank

  • Recovery problem:

given find

  • Null space
  • So search in null space

for the “best” according to some criterion

– ex: least squares

slide-35
SLIDE 35
  • Recovery:

given

(ill-posed inverse problem)

find (sparse)

  • fast

CS Signal Recovery

pseudoinverse

slide-36
SLIDE 36
  • Recovery:

given

(ill-posed inverse problem)

find (sparse)

  • fast, w rong

CS Signal Recovery

pseudoinverse

slide-37
SLIDE 37

Why Doesn’t Work

least squares, minimum solution is almost never sparse

null space of translated to (random angle)

for signals sparse in the space/ tim e dom ain

slide-38
SLIDE 38
  • Reconstruction/ decoding:

given

(ill-posed inverse problem)

find

  • fast, wrong
  • CS Signal Recovery

num ber of nonzero entries “find sparsest in translated nullspace”

slide-39
SLIDE 39
  • Reconstruction/ decoding:

given

(ill-posed inverse problem)

find

  • fast, wrong
  • correct:
  • nly M= 2 K

measurements required to reconstruct K-sparse signal

CS Signal Recovery

number of nonzero entries

slide-40
SLIDE 40
  • Reconstruction/ decoding:

given

(ill-posed inverse problem)

find

  • fast, wrong
  • correct:
  • nly M= 2K

measurements required to reconstruct K-sparse signal slow : NP-complete algorithm

CS Signal Recovery

number of nonzero entries

slide-41
SLIDE 41
  • Recovery:

given

(ill-posed inverse problem)

find (sparse)

  • fast, wrong
  • correct, slow
  • correct, efficient

m ild oversam pling

[ Candes, Romberg, Tao; Donoho]

number of measurements required

CS Signal Recovery

linear program

slide-42
SLIDE 42

Why Works

minimum solution = sparsest solution (with high probability) if for signals sparse in the space/ tim e dom ain

slide-43
SLIDE 43

Universality

  • Random measurements can be used for signals

sparse in any basis

slide-44
SLIDE 44

Universality

  • Random measurements can be used for signals

sparse in any basis

slide-45
SLIDE 45

Universality

  • Random measurements can be used for signals

sparse in any basis

sparse coefficient vector nonzero entries

slide-46
SLIDE 46

Compressive Sensing

  • Directly acquire “com pressed” data
  • Replace N samples by M random projections

transmit/ store receive linear pgm

random m easurem ents

slide-47
SLIDE 47

Com pressive Sensing Theory I I Stable Em bedding

slide-48
SLIDE 48

Johnson-Lindenstrauss Lemma

  • JL Lemma: random projection stably embeds

a cloud of Q points whp provided

  • Proved via concentration inequality
  • Same techniques link JLL to RIP

[ Baraniuk, Davenport, DeVore, Wakin, Constructive Approximation, 2008]

Q points

slide-49
SLIDE 49

Connecting JL to RIP

Consider effect of random JL Φ on each K-plane

– construct covering of points Q on unit sphere – JL: isometry for each point with high probability – union bound isometry for all points q in Q – extend to isometry for all x in K-plane K-plane

slide-50
SLIDE 50

Connecting JL to RIP

Consider effect of random JL Φ on each K-plane

– construct covering of points Q on unit sphere – JL: isometry for each point with high probability – union bound isometry for all points q in Q – extend to isometry for all x in K-plane – union bound isometry for all K-planes K-planes

slide-51
SLIDE 51
  • Gaussian
  • Bernoulli/ Rademacher [ Achlioptas]
  • “Database-friendly” [ Achlioptas]
  • Random Orthoprojection to RM [ Gupta, Dasgupta]

Favorable JL Distributions

slide-52
SLIDE 52

RIP as a “Stable” Embedding

  • RIP of order 2K implies: for all K-sparse x1 and x2,

K-planes

slide-53
SLIDE 53

Com pressive Sensing Recovery Algorithm s

slide-54
SLIDE 54

CS Recovery Algorithms

  • Convex optimization:

– noise-free signals Linear programming (Basis pursuit) FPC Bregman iteration, … – noisy signals Basis Pursuit De-Noising (BPDN) Second-Order Cone Programming (SOCP) Dantzig selector GPSR, …

  • Iterative greedy algorithms

– Matching Pursuit (MP) – Orthogonal Matching Pursuit (OMP) – StOMP – CoSaMP – Iterative Hard Thresholding (IHT), …

softw are @ dsp.rice.edu/ cs

slide-55
SLIDE 55

SOCP

  • Standard LP recovery
  • Noisy measurements
  • Second-Order Cone Program
  • Convex, quadratic program
slide-56
SLIDE 56

BPDN

  • Standard LP recovery
  • Noisy measurements
  • Basis Pursuit De-Noising
  • Convex, quadratic program
slide-57
SLIDE 57

Matching Pursuit

  • Greedy algorithm
  • Key ideas:

(1) measurements composed of sum

  • f K columns of

(2) identify which K columns sequentially according to size of contribution to

columns

slide-58
SLIDE 58

Matching Pursuit

  • For each column

compute

  • Choose largest

(greedy)

  • Update estimate by adding in
  • Form residual measurement

and iterate until convergence

slide-59
SLIDE 59

Orthogonal Matching Pursuit

  • Same procedure

as Matching Pursuit

  • Except at each iteration:

– remove selected column – re-orthogonalize the remaining columns of

  • Converges in K iterations
slide-60
SLIDE 60

Com pressive Sensing I n Action Cam eras

slide-61
SLIDE 61

“Single-Pixel” CS Camera

random pattern on DMD array

DMD DMD

single photon detector im age reconstruction

  • r

processing

w/ Kevin Kelly

scene

slide-62
SLIDE 62

“Single-Pixel” CS Camera

random pattern on DMD array

DMD DMD

single photon detector im age reconstruction

  • r

processing scene

  • Flip mirror array M times to acquire M measurements
  • Sparsity-based (linear programming) recovery

slide-63
SLIDE 63

Single Pixel Camera

Object LED (light source) DMD+ ALP Board Lens 1 Lens 2 Photodiode circuit

slide-64
SLIDE 64

Single Pixel Camera

Object LED (light source) DMD+ ALP Board Lens 1 Lens 2 Photodiode circuit

slide-65
SLIDE 65

Single Pixel Camera

Object LED (light source) DMD+ ALP Board Lens 1 Lens 2 Photodiode circuit

slide-66
SLIDE 66

Single Pixel Camera

Object LED (light source) DMD+ ALP Board Lens 1 Lens 2 Photodiode circuit

slide-67
SLIDE 67

First Image Acquisition

target 65536 pixels 1300 measurements (2% ) 11000 measurements (16% )

slide-68
SLIDE 68

Second Image Acquisition

500 random measurements 4096 pixels

slide-69
SLIDE 69

CS Low-Light Imaging with PMT

true color low-light imaging 256 x 256 image with 10: 1 compression

[ Nature Photonics, April 2007]

slide-70
SLIDE 70

Hyperspectral Imaging

spectrom eter

blue red near I R

slide-71
SLIDE 71

Com pressive Sensing I n Action A/ D Converters

slide-72
SLIDE 72

Analog-to-Digital Conversion

  • Nyquist rate limits reach of today’s ADCs
  • “Moore’s Law” for ADCs:

– technology Figure of Merit incorporating sampling rate and dynamic range doubles every 6 -8 years

  • DARPA Analog-to-Information (A2I) program

– wideband signals have high Nyquist rate but are often sparse/ compressible – develop new ADC technologies to exploit – new tradeoffs among Nyquist rate, sampling rate, dynamic range, …

frequency hopper spectrogram

time frequency

slide-73
SLIDE 73

Analog-to-Information Conversion

  • Sample near signal’s (low) “information rate”

rather than its (high) Nyquist rate

  • Practical hardware:

randomized demodulator (CDMA receiver)

A2I sampling rate number of tones / window Nyquist bandwidth

slide-74
SLIDE 74

Example: Frequency Hopper

20x sub-Nyquist sampling spectrogram sparsogram Nyquist rate sampling

slide-75
SLIDE 75

Com pressive Sensing I n Action Data Processing

slide-76
SLIDE 76

Information Scalability

  • Many applications involve signal inference

and not reconstruction

detection < classification < estim ation < reconstruction

fairly computationally intense

slide-77
SLIDE 77

Information Scalability

  • Many applications involve signal inference

and not reconstruction

detection < classification < estim ation < reconstruction

  • Good new s:

CS supports efficient learning, inference, processing directly

  • n compressive measurements
  • Random projections ~ sufficient statistics

for signals with concise geometrical structure

slide-78
SLIDE 78

Matched Filter

  • Detection/ classification with K unknown

articulation param eters

– Ex: position and pose of a vehicle in an image – Ex: time delay of a radar signal return

  • Matched filter: joint parameter estimation and

detection/ classification

– compute sufficient statistic for each potential target and articulation – compare “best” statistics to detect/ classify

slide-79
SLIDE 79

Matched Filter Geometry

  • Detection/ classification with K unknown

articulation parameters

  • Images are points in
  • Classify by finding closest

target template to data for each class (AWG noise)

– distance or inner product data

target tem plates

from generative model

  • r

training data (points)

slide-80
SLIDE 80

Matched Filter Geometry

  • Detection/ classification with K unknown

articulation parameters

  • Images are points in
  • Classify by finding closest

target template to data

  • As template articulation

parameter changes, points map out a K-dim nonlinear m anifold

  • Matched filter classification

= closest m anifold search

articulation parameter space

data

slide-81
SLIDE 81

CS for Manifolds

  • Theorem :

random measurements stably em bed m anifold whp

[ Baraniuk, Wakin, FOCM ’08] related work: [ Indyk and Naor, Agarwal et al., Dasgupta and Freund]

  • Stable embedding
  • Proved via concentration

inequality arguments (JLL/ CS relation)

slide-82
SLIDE 82

CS for Manifolds

  • Theorem :

random measurements stably embed manifold whp

  • Enables parameter

estimation and MF detection/ classification directly on com pressive m easurem ents

– K very small in many applications (# articulations)

slide-83
SLIDE 83

Example: Matched Filter

  • Detection/ classification with K= 3 unknown

articulation param eters

  • 1. horizontal translation
  • 2. vertical translation
  • 3. rotation
slide-84
SLIDE 84

Smashed Filter

  • Detection/ classification with K= 3 unknown

articulation parameters ( m anifold structure)

  • Dimensionally reduced matched filter directly on

compressive measurements

slide-85
SLIDE 85

Smashed Filter

  • Random shift and rotation (K= 3 dim. manifold)
  • Noise added to measurements
  • Goal:

identify most likely position for each image class

identify most likely class using nearest-neighbor test

number of measurements M number of measurements M

  • avg. shift estimate error

classification rate (% ) more noise more noise

slide-86
SLIDE 86

Com pressive Sensing Sum m ary

slide-87
SLIDE 87

CS Hallmarks

  • CS changes the rules of the data acquisition game

– exploits a priori signal sparsity information

  • Stable

– acquisition/ recovery process is numerically stable

  • Universal

– same random projections / hardware can be used for any compressible signal class (generic)

  • Asym m etrical (most processing at decoder)

– conventional: smart encoder, dumb decoder – CS: dumb encoder, smart decoder

  • Random projections weakly encrypted
slide-88
SLIDE 88

CS Hallmarks

  • Dem ocratic

– each measurement carries the same amount of information – robust to measurement loss and quantization simple encoding

  • Ex: wireless streaming application with data loss

– conventional: complicated (unequal) error protection of compressed data DCT/ wavelet low frequency coefficients – CS: merely stream additional measurements and reconstruct using those that arrive safely (fountain-like)

slide-89
SLIDE 89

After the Break

Beyond Sparsity with structured sparsity models.

slide-90
SLIDE 90

dsp.rice.edu/ cs

volkan@rice.edu