Com pressed Sensing m eets I nform ation Theory Dror Baron ECE - - PowerPoint PPT Presentation

com pressed sensing
SMART_READER_LITE
LIVE PREVIEW

Com pressed Sensing m eets I nform ation Theory Dror Baron ECE - - PowerPoint PPT Presentation

Measurem ents and Bits: Com pressed Sensing m eets I nform ation Theory Dror Baron ECE Department Rice University dsp.rice.edu/ cs Sensing by Sampling Sam ple data at Nyquist rate Com press data using model (e.g., sparsity)


slide-1
SLIDE 1

Dror Baron

ECE Department Rice University dsp.rice.edu/ cs

Measurem ents and Bits:

Com pressed Sensing

m eets

I nform ation Theory

slide-2
SLIDE 2

Sensing by Sampling

  • Sam ple data at Nyquist rate
  • Com press data using model (e.g., sparsity)

– encode coefficient locations and values

  • Lots of work to throw away > 80% of the coefficients
  • Most computation at sensor (asymmetrical)
  • Brick wall to performance of modern acquisition systems

com press transmit/ store receive decompress sample

sparse wavelet transform

slide-3
SLIDE 3

Sparsity / Compressibility

pixels large wavelet coefficients wideband signal samples large Gabor coefficients

  • Many signals are sparse or compressible in

some representation/ basis (Fourier, wavelets, … )

slide-4
SLIDE 4

Compressed Sensing

  • Shannon/ Nyquist sampling theorem

– worst case bound for any bandlimited signal – too pessimistic for some classes of signals – does not exploit signal sparsity/ compressibility

  • Seek direct sensing of compressible information
  • Compressed Sensing (CS)

– sparse signals can be recovered from a small number

  • f nonadaptive (fixed) linear measurements

– [ Candes et al.; Donoho; Kashin; Gluskin; Rice… ]

– based on new uncertainty principles beyond Heisenberg (“incoherency”)

slide-5
SLIDE 5

Incoherent Bases (matrices)

  • Spikes and sines (Fourier)
slide-6
SLIDE 6

Incoherent Bases

  • Spikes and “random noise”
slide-7
SLIDE 7
  • Measure linear projections onto incoherent basis

where data is not sparse/ compressible

– random projections are universally incoherent – fewer measurements – no location information

  • Reconstruct via optimization
  • Highly asymmetrical (most computation at receiver)

Compressed Sensing via Random Projections

project transmit/ store receive reconstruct

slide-8
SLIDE 8

CS Encoding

  • Replace sam ples by more general encoder

based on a few linear projections (inner products)

  • Matrix vector multiplication – potentially analog

measurements sparse signal # non-zeros

slide-9
SLIDE 9
  • Random projections
  • Universally incoherent with any compressible/ sparse

signal class

measurements sparse signal

Universality via Random Projections

# non-zeros

slide-10
SLIDE 10

Reconstruction Before-CS –

  • Goal: Given measurements find signal
  • Fewer rows than columns in measurement matrix
  • Ill-posed: infinitely many solutions
  • Classical solution:

least squares

slide-11
SLIDE 11
  • Goal: Given measurements find signal
  • Fewer rows than columns in measurement matrix
  • Ill-posed: infinitely many solutions
  • Classical solution:

least squares

  • Problem:

small L2 doesn’t imply sparsity

Reconstruction Before-CS –

slide-12
SLIDE 12

Ideal Solution –

  • Ideal solution: exploit sparsity of
  • Of the infinitely many solutions seek sparsest one

number of nonzero entries

slide-13
SLIDE 13

Ideal Solution –

  • Ideal solution: exploit sparsity of
  • Of the infinitely many solutions seek sparsest one
  • If M · K then w/ high probability this can’t be done
  • If M ¸ K+ 1 then perfect reconstruction

w/ high probability [ Bresler et al.; Wakin et al.]

  • But not robust and combinatorial complexity
slide-14
SLIDE 14

The CS Revelation –

  • Of the infinitely many solutions seek the one

with smallest L1 norm

slide-15
SLIDE 15
  • Of the infinitely many solutions seek the one

with smallest L1 norm

  • If

then perfect reconstruction w/ high probability [ Candes et al.; Donoho]

  • Robust to measurement noise
  • Linear programming

The CS Revelation –

slide-16
SLIDE 16

CS Hallmarks

  • CS changes the rules of data acquisition game

– exploits a priori signal sparsity information (signal is compressible)

  • Hardw are:

Universality

– same random projections / hardware for any compressible signal class – simplifies hardware and algorithm design

  • Processing:

I nform ation scalability

– random projections ~ sufficient statistics – same random projections for range of tasks reconstruction > estimation > recognition > detection – far fewer measurements required to detect/ recognize

  • Next generation data acquisition

– new imaging devices and A/ D converters [ DARPA] – new reconstruction algorithms – new distributed source coding algorithms [ Baron et al.]

slide-17
SLIDE 17

Random Projections in Analog

slide-18
SLIDE 18

Optical Computation of Random Projections

  • CS encoder integrates sensing, compression, processing
  • Example: new cameras and imaging algorithms
slide-19
SLIDE 19

First Image Acquisition (M= 0.38N)

ideal 64x64 image (4096 pixels) 400 wavelets image on DMD array 1600 random meas.

slide-20
SLIDE 20

A/ D Conversion Below Nyquist Rate

  • Challenge:

– wideband signals (radar, communications, … ) – currently impossible to sample at Nyquist rate

  • Proposed CS-based solution:

– sample at “information rate” – simple hardware components – good reconstruction performance

Downsample Filter Modulator

slide-21
SLIDE 21

Connections Betw een Com pressed Sensing and I nform ation Theory

slide-22
SLIDE 22

Measurement Reduction via CS

  • CS reconstruction via

– If then perfect reconstruction w/ high probability [ Candes et al.; Donoho] – Linear programming

  • Compressible signals (signal components decay)

– also requires – polynomial complexity (BPDN) [ Candes et al.] – cannot reduce order of [ Kashin,Gluskin]

slide-23
SLIDE 23

Fundamental Goal: Minimize

  • Compressed sensing aims to minimize resource

consumption due to measurements

  • Donoho:

“Why go to so much effort to acquire all the data when most of what we get will be thrown away?”

slide-24
SLIDE 24

Fundamental Goal: Minimize

  • Compressed sensing aims to minimize resource

consumption due to measurements

  • Donoho:

“Why go to so much effort to acquire all the data when most of what we get will be thrown away?”

  • Recall sparse signals

– only measurements for reconstruction – not robust and combinatorial complexity

slide-25
SLIDE 25

Rich Design Space

  • What performance metric to use?

– Determine support set of nonzero entries [ Wainwright] this is distortion metric but why let tiny nonzero entries spoil the fun? – metric? ?

slide-26
SLIDE 26

Rich Design Space

  • What performance metric to use?

– Determine support set of nonzero entries [ Wainwright] this is distortion metric but why let tiny nonzero entries spoil the fun? – metric? ?

  • What complexity class of reconstruction algorithms?

– any algorithms? – polynomial complexity? – near-linear or better?

slide-27
SLIDE 27

Rich Design Space

  • What performance metric to use?

– Determine support set of nonzero entries [ Wainwright] this is distortion metric but why let tiny nonzero entries wreck spoil the fun? – metric? ?

  • What complexity class of reconstruction algorithms?

– any algorithms? – polynomial complexity? – near-linear or better?

  • How to account for imprecisions?

– noise in measurements? – compressible signal model?

slide-28
SLIDE 28

Low er Bound on Num ber of Measurem ents

slide-29
SLIDE 29

Measurement Noise

  • Measurement process is analog
  • Analog systems add noise, non-linearities, etc.
  • Assume Gaussian noise for ease of analysis
slide-30
SLIDE 30

Setup

  • Signal is iid
  • Additive white Gaussian noise
  • Noisy measurement process
slide-31
SLIDE 31

Setup

  • Signal is iid
  • Additive white Gaussian noise
  • Noisy measurement process
  • Random projection of tiny coefficients (compressible

signals) similar to measurement noise

slide-32
SLIDE 32

Measurement and Reconstruction Quality

  • Measurement signal to noise ratio
  • Reconstruct using decoder mapping
  • Reconstruction distortion metric
  • Goal: minimize CS measurement rate
slide-33
SLIDE 33

Measurement Channel

  • Model process

as measurement channel

  • Capacity of measurement channel
  • Measurem ents are bits!
slide-34
SLIDE 34

Lower Bound [ Sarvotham et al.]

  • Theorem : For a sparse signal with rate-distortion

function , lower bound on measurement rate subject to measurement quality and reconstruction distortion satisfies

  • Direct relationship to rate-distortion content
  • Applies to any linear signal acquisition system
slide-35
SLIDE 35

Lower Bound [ Sarvotham et al.]

  • Theorem : For a sparse signal with rate-distortion

function , lower bound on measurement rate subject to measurement quality and reconstruction distortion satisfies

  • Proof sketch:

– each measurement provides bits – information content of source bits – source-channel separation for continuous amplitude sources – minimal number of measurements – obtain measurement rate via normalization by

slide-36
SLIDE 36

Example

  • Spike process -

spikes of uniform amplitude

  • Rate-distortion function
  • Lower bound
  • Numbers:

– signal of length 107 – 103 spikes – SNR= 10 dB ⇒ – SNR= -20 dB ⇒

  • If interesting portion of signal has relatively small

energy then need significantly more measurements!

  • Upper bound (achievable) in progress…
slide-37
SLIDE 37

CS Reconstruction Meets Channel Coding

slide-38
SLIDE 38

Why is Reconstruction Expensive?

measurements sparse signal nonzero entries

Culprit: dense, unstructured

slide-39
SLIDE 39

Fast CS Reconstruction

measurements sparse signal nonzero entries

  • LDPC measurement matrix (sparse)
  • Only 0/ 1 in
  • Each row of contains randomly placed 1’s
  • Fast matrix multiplication

fast encoding fast reconstruction

slide-40
SLIDE 40

Ongoing Work: CS Using BP

  • Considering noisy CS signals
  • Application of Belief Propagation

– BP over real number field – sparsity is modeled as prior in graph

Measurements Y States Coefficients X Q

slide-41
SLIDE 41

Promising Results

500 1000 1500 2000 2500 3000 100 200 300 400 500 l2 norm of (x−x_hat) vs M Number of measurements (M) l2 reconstruction error l=5 l=10 l=15 l=20 l=25 norm(x) norm(x_n)

200 400 600 −40 −20 20 40 60 80 j x(j)

slide-42
SLIDE 42

Theoretical Advantages of CS-BP

  • Low complexity
  • Provable reconstruction with noisy measurements

using

  • Success of LDPC+ BP in channel coding carried over

to CS!

slide-43
SLIDE 43

Distributed Com pressed Sensing ( DCS)

CS for distributed signal ensembles

slide-44
SLIDE 44

Why Distributed?

  • Networks of many sensor nodes

– sensor, microprocessor for computation, wireless communication, networking, battery – can be spread over large geographical area

  • Must be energy efficient

– minimize communication at expense of computation – motivates distributed compression

slide-45
SLIDE 45

destination raw data

Distributed Sensing

  • Transmitting raw data typically inefficient
slide-46
SLIDE 46
  • Can we exploit

intra-sensor and inter-sensor correlation to jointly compress?

  • Ongoing challenge in information

theory (distributed source coding)

Correlation

destination

?

slide-47
SLIDE 47

destination

Collaborative Sensing

  • Collaboration introduces

– inter-sensor communication overhead – complexity at sensors

compressed data

slide-48
SLIDE 48

destination

Distributed Compressed Sensing

  • Exploit intra- and inter-sensor

correlations with

– zero inter-sensor communication overhead – low complexity at sensors

  • Distributed source coding via CS

compressed data

slide-49
SLIDE 49

Model 1 : Com m on + I nnovations

slide-50
SLIDE 50

Common + Innovations Model

  • Motivation: measuring signals in smooth field

– “average” temperature value common at multiple locations – “innovations” driven by wind, rain, clouds, etc.

  • Joint sparsity model:

– length-N sequences x1 and x2 – zC is length-N common component – z1, z2 length-N innovations components – zC, z1, z2 have sparsity KC, K1, K2

  • Measurements
slide-51
SLIDE 51

Measurement Rate Region with Separate Reconstruction

separate encoding & recon

Decoder g1 Decoder g2 Encoder f1 Encoder f2

slide-52
SLIDE 52

Slepian-Wolf Theorem

(Distributed lossless coding)

  • Theorem : [ Slepian and Wolf 1973]

R1 > H(X1| X2)

(conditional entropy)

R2 > H(X2| X1)

(conditional entropy)

R1+ R2 > H(X1,X2)

(joint entropy) R1 R2 H(X2| X1) H(X2) H(X1) H(X1| X2) Slepian-Wolf joint recon separate encoding & separate recon

slide-53
SLIDE 53

separate encoding & joint recon

Measurement Rate Region with Joint Reconstruction

Encoder f1 Decoder g Encoder f2

  • Inspired by Slepian-Wolf coding
slide-54
SLIDE 54

sim ulation separate reconstruction converse achievable

Measurement Rate Region [ Baron et al.]

slide-55
SLIDE 55

Multiple Sensors

slide-56
SLIDE 56

Model 2 : Com m on Sparse Supports

slide-57
SLIDE 57

Ex: Many audio signals

  • sparse in Fourier Domain
  • same frequencies received

by each node

  • different attenuations and delays

(magnitudes and phases)

Common Sparse Supports Model

slide-58
SLIDE 58
  • Signals share sparse components but

different coefficients

  • Intuition: Each measurement vector holds clues

about coefficient support set

Common Sparse Supports Model

slide-59
SLIDE 59

Required Number of Measurements

[ Baron et al. 2005]

  • Theorem : M= K measurements per sensor do not

suffice to reconstruct signal ensemble

  • Theorem : As number of sensors J increases, M= K+ 1

measurements suffice to reconstruct

  • Joint reconstruction with reasonable computational

complexity

slide-60
SLIDE 60

Results for Common Sparse Supports

K= 5 N= 50

Separate Joint Reconstruction

slide-61
SLIDE 61

Real Data Example

  • Light levels in Intel Berkeley Lab
  • 49 sensors, 1024 samples each
  • Compare:

– wavelet approx 100 terms per sensor – separate CS 400 measurements per sensor – joint CS (SOMP) 400 measurements per sensor

  • Correlated signal ensemble
slide-62
SLIDE 62

Light Intensity at Node 19

slide-63
SLIDE 63

Model 3 : Non-Sparse Com m on Com ponent

slide-64
SLIDE 64

Non-Sparse Common Model

  • Motivation: non-sparse video frame + sparse motion
  • Length-N common component zC is non-sparse

⇒Each signal is incompressible

  • Innovation sequences zj may share supports
  • Intuition: each measurement vector contains clues

about common component zC

not sparse sparse

slide-65
SLIDE 65

Results for Non-Sparse Common

(same supports)

K= 5 N= 50

Impact of zC vanishes as J ! 1

slide-66
SLIDE 66

Summary

  • Com pressed Sensing

– “random projections” – process sparse signals using far fewer measurements – universality and information scalability

  • Determination of measurement rates in CS

– measurements are bits – lower bound on measurement rate direct relationship to rate-distortion content

  • Promising results with LDPC measurement matrices
  • Distributed CS

– new models for joint sparsity – analogy with Slepian-Wolf coding from information theory – compression of sources w/ intra- and inter-sensor correlation

  • Much potential and m uch m ore to be done
  • Com pressed sensing m eets inform ation theory

dsp.rice.edu/ cs

slide-67
SLIDE 67

THE END

slide-68
SLIDE 68

“With High Probability”