Sparsity, Randomness and Compressed Sensing Petros Boufounos - - PowerPoint PPT Presentation

sparsity randomness and compressed sensing
SMART_READER_LITE
LIVE PREVIEW

Sparsity, Randomness and Compressed Sensing Petros Boufounos - - PowerPoint PPT Presentation

Sparsity, Randomness and Compressed Sensing Petros Boufounos Mitsubishi Electric Research Labs petrosb@merl.com Sparsity Why Sparsity Naturaldataandsignalsexhibit structure Sparsity o2encapturesthat


slide-1
SLIDE 1

Sparsity, Randomness and Compressed Sensing

Petros Boufounos Mitsubishi Electric Research Labs

petrosb@merl.com

slide-2
SLIDE 2

Sparsity


slide-3
SLIDE 3

Why Sparsity

  • Natural
data
and
signals
exhibit
structure

  • Sparsity
o2en
captures
that
structure

  • Very
general
signal
model

  • Computa9onally
tractable

  • Wide
range
of
applica9ons
in
signal
acquisi2on,


processing,
and
transmission


slide-4
SLIDE 4

Signal Representations

slide-5
SLIDE 5

Signal example: Images

  • 2‐D
func9on

  • Idealized
view



 
some
func9on
 
 
space
defined
 
 
over



slide-6
SLIDE 6

Signal example: Images

  • 2‐D
func9on

  • Idealized
view



 
some
func9on
 
 
space
defined
 
 
over



  • In
prac9ce



 
ie:
an














matrix


slide-7
SLIDE 7

Signal example: Images

  • 2‐D
func9on

  • Idealized
view



 
some
func9on
 
 
space
defined
 
 
over



  • In
prac9ce



 
ie:
an














matrix

(pixel
average)


slide-8
SLIDE 8

Signal Models

Classical Model: Signal lies in a linear vector space (e.g. bandlimited functions) Sparse Model: Signals of interest are often sparse

  • r compressible

Signal Transform Image Bat Sonar Chirp Wavelet Gabor/ STFT i.e., very few large coefficients, many close to zero.

x2 x3 x1

X

slide-9
SLIDE 9

Sparse Signal Models

x2 x3 x1 X x1 x2 x3 X 1-sparse 2-sparse Compressible (p ball, p<1)

Sparse signals have few non-zero coefficients Compressible signals have few significant coefficients. The coefficients decay as a power law.

x1 x2 x3

slide-10
SLIDE 10

Sparse Approximation

slide-11
SLIDE 11

Computational Harmonic Analysis

  • Representa9on

  • Analysis:







study




through
structure
of




 
















should
extract features
of
interest


  • Approxima2on: 






uses
just
a
few
terms



 
 
 
exploit
sparsity of



basis,
frame
 coefficients


slide-12
SLIDE 12

Wavelet Transform Sparsity

  • Many













(blue)


slide-13
SLIDE 13

Sparseness ⇒ Approximation

sorted
index


few big many small

slide-14
SLIDE 14

Linear Approximation

index


slide-15
SLIDE 15

Linear Approximation

  • ‐term approxima6on:

use
“first”








index


slide-16
SLIDE 16

Nonlinear Approximation

  • ‐term approxima6on:





 
 
 
use
largest






independently



  • Greedy
/
thresholding

sorted
index


few big

slide-17
SLIDE 17

Error Approximation Rates

  • Op9mize
asympto9c
error decay rate
  • Nonlinear
approxima9on
works
beQer
than
linear


as


slide-18
SLIDE 18

Compression is Approximation

  • Lossy
compression
of
an
image
creates
an


approxima9on


basis,
frame
 coefficients
 quan6ze to total bits

slide-19
SLIDE 19
  • Sparse
approxima9on
chooses
coefficients
but
does


not quan6ze
or
worry
about
their
loca6ons

threshold

Sparse approximation ≠ Compression

slide-20
SLIDE 20

Location, Location, Location

  • Nonlinear
approxima9on


selects





largest
 to
minimize
error

 (easy
–
threshold)


  • Compression
algorithm


must
encode
both
a
set


  • f





and
their
loca9ons


(harder)







slide-21
SLIDE 21

Exposing Sparsity

slide-22
SLIDE 22

Spikes and Sinusoids example

Example Signal Model: Sinusoidal with a few spikes. DCT Basis:

B f a

slide-23
SLIDE 23

Spikes and Sinusoids Dictionary

DCT basis

D f a

Impulses Lost Uniqueness!!

slide-24
SLIDE 24

Overcomplete Dictionaries D f a

Strategy: Improve sparse approximation by constructing a large dictionary. How do we design a dictionary?

slide-25
SLIDE 25

Dictionary Design

DCT, DFT Impulse Basis Wavelets Edgelets, curvelets, … Oversampling Frame Dictionary D Can we just throw in the bucket everything we know? … …

slide-26
SLIDE 26

Dictionary Design Considerations

  • Dic9onary
Size:



– Computa2on
and
storage
increases
with
size


  • Fast
Transforms:


– FFT,
DCT,
FWT,
etc.
drama9cally
decrease
computa2on
and
 storage


  • Coherence:


– Similarity
in
elements
makes
solu9on
harder


slide-27
SLIDE 27

Dictionary Coherence D1 D2

Two candidate dictionaries: BAD! Intuition: D2 has too many similar elements. It is very coherent. Coherence (similarity) between elements: 〈d1,d2〉 Dictionary coherence: μ=maxi,j〈di,dj〉

slide-28
SLIDE 28

Incoherent Bases

  • “Mix”
well
the
signal
components


– Impulses
and
Fourier
Basis
 – Anything
and
Random
Gaussian
 – Anything
and
Random
0‐1
basis


slide-29
SLIDE 29

Computing Sparse Representations

slide-30
SLIDE 30

Thresholding

Compute set of coefficients

D f a

a=D†f

Zero out small ones Computationally efficient Good for small and very incoherent dictionaries

slide-31
SLIDE 31

Matching Pursuit

DT

ρ f

Measure image against dictionary Select largest correlation

ρk

Add to representation

ak ← ak+ρk

Compute residual

f ← f - ρkdk

Iterate using residual

〈dk,f 〉 = ρk

slide-32
SLIDE 32

Greedy Pursuits Family

  • Several
Varia9ons
of
MP:


OMP,
StOMP,
ROMP,
CoSaMP,
Tree
MP,
…
 (You
can
create
an
AndrewMP
if

you
work
on
it…)






















  • Some
have
provable
guarantees

  • Some
improve
dic2onary
search

  • Some
improve
coefficient
selec2on

slide-33
SLIDE 33

CoSaMP (Compressive Sampling MP)

DT

ρ f

Measure image against dictionary Iterate using residual

〈dk,f 〉 = ρk

Select location

  • f largest

2K correlations

supp(ρ|2K)

Add to support set Truncate and compute residual

Ω = supp(ρ|2K) ∪ T

b = D†

Ωf

Invert over support

T = supp(b|K) a = b|K r ← f − Da

slide-34
SLIDE 34

Optimization (Basis Pursuit)

Sparse approximation: Minimize non-zeros in representation s.t.: representation is close to signal

min
‖a‖0

s.t.

f ≈ Da

Number of non-zeros (sparsity measure) Data Fidelity (approximation quality) Combinatorial complexity. Very hard problem!

slide-35
SLIDE 35

Optimization (Basis Pursuit)

Sparse approximation: Minimize non-zeros in representation s.t.: representation is close to signal

min
‖a‖0

s.t.

f ≈ Da min
‖a‖1

s.t.

f ≈ Da

Convex Relaxation Ploynomial complexity. Solved using linear programming.

slide-36
SLIDE 36

Why l1 relaxation works

f = Da min
‖a‖1

s.t.

f ≈ Da

l1 “ball”

Sparse solution

slide-37
SLIDE 37

Basis Pursuits

  • Have
provable
guarantees


– Finds
sparsest
solu9on
for
incoherent
dic9onaries


  • Several
variants
in
formula9on:


BPDN,
LASSO,
Dantzig
selector,
…


  • Varia9ons
on
fidelity
term
and
relaxa2on
choice

  • Several
fast
algorithms:


FPC,
GPSR,
SPGL,
…


slide-38
SLIDE 38

Compressed
Sensing:
 Sensing,
Sampling
and

 Data
Processing


slide-39
SLIDE 39

Data Acquisition

  • Usual
acquisi9on
methods
sample
signals
uniformly


– Time:
A/D
with
microphones,
geophones,
hydrophones.
 – Space:
CCD
cameras,
sensor
arrays.


  • Founda9on:
Nyquist/Shannon
sampling
theory


– Sample
at
twice
the
signal
bandwidth.
 – Generally
a
projec2on
to
a
complete
basis
that
spans
the
 signal
space.


slide-40
SLIDE 40

Data Processing and Transmission

  • Data
processing
steps:


– Sample
Densely
 – Transform
to
an
informa9ve
domain
(Fourier,
Wavelet)
 – Process/Compress/Transmit


Sets
small
coefficients
to
zero
(sparsifica9on)


Signal x, N coefficients K<<N significant coefficients

slide-41
SLIDE 41

Sparsity Model

  • Signals
can
usually
be
compressed
in
some
basis

  • Sparsity:
good
prior
in
picking
from
a
lot
of
candidates

slide-42
SLIDE 42

x2 x3 x1 X x1 x2 x3 X 1-sparse 2-sparse

Compressive Sensing Principles

If a signal is sparse, do not waste effort sampling the empty space. Instead, use fewer samples and allow ambiguity. Use the sparsity model to reconstruct and uniquely resolve the ambiguity.

slide-43
SLIDE 43

Measuring Sparse Signals

slide-44
SLIDE 44

Compressive Measurements

x2 x3 x1 X

N = Signal dimensionality K = Signal sparsity

x1 x2 x3 y1 y2 X

Measurement (Projection) Reconstruction Ambiguity

N ≫
M ≳ K

Φ has rank M≪N M = Number of measurements (dimensionality of y)

slide-45
SLIDE 45

One Simple Question

slide-46
SLIDE 46

Geometry of Sparse Signal Sets

slide-47
SLIDE 47

Geometry: Embedding in RM

slide-48
SLIDE 48

Illustrative Example

slide-49
SLIDE 49

Example: 1-sparse signal

x2 x3 x1 X N=3 K=1 M=2K=2 x1=0 y1=x2 y2 =x3 X Bad! y1=x1=x2 y2 =x3 X Bad!

slide-50
SLIDE 50

Example: 1-sparse signal

x2 x3 x1 X N=3 K=1 M=2K=2 x1 y1=x2 y2 =x3 X Good! x2 y1=x1 y2 X x3 Better!

slide-51
SLIDE 51

Restricted Isometry Property

slide-52
SLIDE 52

RIP as a “Stable” Embedding

slide-53
SLIDE 53

Verifying RIP

slide-54
SLIDE 54

Universality Property

slide-55
SLIDE 55

Universality Property

slide-56
SLIDE 56

Democracy

measurements
 Bad/lost/dropped
measurements


Φ


  • Measurements
are
democra2c
[Davenport,
Laska,
Boufounos,
Baraniuk]

  • They
are
all
equally
important

  • We
can
loose
some
arbitrarily,
(i.e.
an
adversary
can
choose


which
ones)


  • The
Φ
s9ll
sa9sfies
RIP
(as
long
as
we
don’t
drop
too
many)


~


Φ


~


~

slide-57
SLIDE 57

Reconstruction

slide-58
SLIDE 58

Requirements for Reconstruction

  • Let
x1, x2 be
K‐sparse
signals
(I.e.
x1-x2 is 2K‐sparse):

  • Mapping
y=Φx
is
inver2ble
for
K‐sparse
signals:


Φ(x1-x2)≠0
if
x1≠x2

  • Mapping
is
robust
for
K‐sparse
signals:


||Φ(x1-x2)||2≈|| x1-x2||2


– Restricted
Isometry
Property
(RIP):

 Φ
preserves
distance
when
projec9ng
K‐sparse
signals


  • Guarantees
there
exists
a
unique
K‐sparse
signal
explains
the


measurements,
and
is
robust
to
noise.


slide-59
SLIDE 59

Reconstruction Ambiguity

  • Solu9on
should
be
consistent
with
measurements

  • Projec9ons
imply
that
an
infinite
number
of
solu9ons
are
consistent!

  • Classical
approach:
use
the
pseudoinverse
(minimize
l2
norm)

  • Compressive
sensing
approach:
pick
the
sparsest.

  • RIP
guarantee:
sparsest
solu9on
unique
and
reconstructs
the
signal.


ˆ x s.t. y = Φˆ x or y ≈ Φˆ x

Becomes a sparse approximation problem!

slide-60
SLIDE 60

Putting everything together

slide-61
SLIDE 61

Compressed Sensing Coming Together

  • Signal
model:
Provides
prior
informa9on;
allows
undersampling

  • Randomness:
Provides
robustness/stability;
makes
proofs
easier

  • Non‐linear
reconstruc2on:
Incorporates
informa9on
through
computa9on


Measurement y=Φx Reconstruc9on
 using
sparse
 approxima9on


x y x ~

Signal
Structure
(sparsity)
 Stable
Embedding
 (random
projec9ons)
 Non‐linear
Reconstruc2on
 (Basis
Pursuit,
Matching
 Pursuit,
CoSaMP,
etc…)


slide-62
SLIDE 62

Beyond:
Extensions,

 Connec2ons,
Generaliza2ons


slide-63
SLIDE 63

Sparsity Models

slide-64
SLIDE 64

Block Sparsity

measurements
 sparse
 signal
 nonzero
 blocks
of
L

Mixed l1/l2 norm—sum of l2 norms: Basis pursuit becomes: Blocks are not allowed to overlap

Φ

  • i

xBi2 min

x

  • i

xBi2 s.t. y ≈ Φx

slide-65
SLIDE 65

y x

Joint Sparsity

Φ

Mixed l1/l2 norm—sum of l2 norms: Basis pursuit becomes: min

x

  • i

x(i,·)2 s.t. y ≈ Φx

  • i

x(i,·)2

M × L

measurements


N L

L
sparse
signals
in
RN Sparse
components
per
signal

 with
common
support


slide-66
SLIDE 66

Randomized Embeddings

slide-67
SLIDE 67

Stable Embeddings

slide-68
SLIDE 68

Johnson-Lindenstrauss Lemma

slide-69
SLIDE 69

Favorable JL Distributions

slide-70
SLIDE 70

Connecting JL to RIP

slide-71
SLIDE 71

Connecting JL to RIP

slide-72
SLIDE 72
slide-73
SLIDE 73

More?


slide-74
SLIDE 74

The tip of the iceberg

Today’s lecture Compressive Sensing Repository dsp.rice.edu/cs Blog on CS nuit-blanche.blogspot.com/ Yet to be discovered… Start working on it 