Beyond Compressed Sensing: The Effectiveness of Convex Programming - - PowerPoint PPT Presentation

beyond compressed sensing the effectiveness of convex
SMART_READER_LITE
LIVE PREVIEW

Beyond Compressed Sensing: The Effectiveness of Convex Programming - - PowerPoint PPT Presentation

Beyond Compressed Sensing: The Effectiveness of Convex Programming in the Information and Physical Sciences Emmanuel Cand` es EUSIPCO 2015, Nice, September 2015 Three stories about signal recovery from missing information Today I want to


slide-1
SLIDE 1

Beyond Compressed Sensing: The Effectiveness of Convex Programming in the Information and Physical Sciences

Emmanuel Cand` es EUSIPCO 2015, Nice, September 2015

slide-2
SLIDE 2

Three stories about signal recovery from missing information

Today I want to tell you three stories from my life. That’s it. No big deal. Just three stories Steve Jobs

slide-3
SLIDE 3

Three stories about signal recovery from missing information

Today I want to tell you three stories from my life. That’s it. No big deal. Just three stories Steve Jobs (1) Missing phase (phase retrieval) (2) Missing and/or corrupted entries in data matrix (robust PCA) (3) Missing high-frequency spectrum (super-resolution) Makes signal/data recovery difficult

This lecture

Convex programming usually (but not always) returns the right answer!

slide-4
SLIDE 4

Story # 1: Phase Retrieval

Collaborators: Y. Eldar, X. Li, T. Strohmer, V. Voroninski

slide-5
SLIDE 5

X-ray crystallography

Method for determining atomic structure within a crystal principle typical setup 10 Nobel Prizes in X-ray crystallography, and counting...

slide-6
SLIDE 6

Importance

principle Franklin’s photograph

slide-7
SLIDE 7

Missing phase problem

Detectors record intensities of diffracted rays → phaseless data only! Fraunhofer diffraction − → intensity of electrical field |ˆ x(f1, f2)|2 =

  • x(t1, t2)e−i2π(f1t1+f2t2) dt1dt2
  • 2
slide-8
SLIDE 8

Missing phase problem

Detectors record intensities of diffracted rays → phaseless data only! Fraunhofer diffraction − → intensity of electrical field |ˆ x(f1, f2)|2 =

  • x(t1, t2)e−i2π(f1t1+f2t2) dt1dt2
  • 2

Phase retrieval problem (inversion)

How can we recover the phase (or equivalently signal x(t1, t2)) from |ˆ x(f1, f2)|?

slide-9
SLIDE 9

About the importance of phase...

slide-10
SLIDE 10

About the importance of phase...

DFT DFT

slide-11
SLIDE 11

About the importance of phase...

DFT DFT keep magnitude swap phase

slide-12
SLIDE 12

About the importance of phase...

DFT DFT keep magnitude swap phase

slide-13
SLIDE 13

X-ray imaging: now and then

  • ntgen (1895)

Dierolf (2010)

slide-14
SLIDE 14

Ultrashort X-ray pulses

Imaging single large protein complexes

slide-15
SLIDE 15

Discrete mathematical model

Phaseless measurements about x0 ∈ Cn bk = |ak, x0|2 k ∈ {1, . . . , m} = [m] Phase retrieval is feasibility problem find x subject to |ak, x|2 = bk k ∈ [m] Solving quadratic equations is NP hard in general

slide-16
SLIDE 16

Discrete mathematical model

Phaseless measurements about x0 ∈ Cn bk = |ak, x0|2 k ∈ {1, . . . , m} = [m] Phase retrieval is feasibility problem find x subject to |ak, x|2 = bk k ∈ [m] Solving quadratic equations is NP hard in general 1985 Nobel Prize for Hauptman & Karle: use very specific prior knowledge

slide-17
SLIDE 17

Discrete mathematical model

Phaseless measurements about x0 ∈ Cn bk = |ak, x0|2 k ∈ {1, . . . , m} = [m] Phase retrieval is feasibility problem find x subject to |ak, x|2 = bk k ∈ [m] Solving quadratic equations is NP hard in general 1985 Nobel Prize for Hauptman & Karle: use very specific prior knowledge Standard approach: Gerchberg–Saxton (or Fienup) iterative algorithm Sometimes works well Sometimes does not

slide-18
SLIDE 18

Quadratic equations: geometric view I

slide-19
SLIDE 19

Quadratic equations: geometric view I

slide-20
SLIDE 20

Quadratic equations: geometric view I

slide-21
SLIDE 21

Quadratic equations: geometric view I

slide-22
SLIDE 22

Quadratic equations: geometric view I

slide-23
SLIDE 23

Quadratic equations: geometric view I

slide-24
SLIDE 24

Quadratic equations: geometric view II

slide-25
SLIDE 25

PhaseLift

Lifting: X = xx∗ bk = |ak, x|2 = a∗

kxx∗ak = a∗ kXak

Turns quadratic measurements into linear measurements b := A(X) about xx∗

slide-26
SLIDE 26

PhaseLift

Lifting: X = xx∗ bk = |ak, x|2 = a∗

kxx∗ak = a∗ kXak

Turns quadratic measurements into linear measurements b := A(X) about xx∗

Phase retrieval: equivalent formulation

find X

  • s. t.

A(X) = b X 0, rank(X) = 1 ⇐ ⇒ min rank(X)

  • s. t.

A(X) = b X 0 Combinatorially hard

slide-27
SLIDE 27

PhaseLift

Lifting: X = xx∗ bk = |ak, x|2 = a∗

kxx∗ak = a∗ kXak

Turns quadratic measurements into linear measurements b := A(X) about xx∗

PhaseLift: tractable semidefinite relaxation

minimize Tr(X) subject to A(X) = b X 0 This is a semidefinite program (SDP) Trace is convex proxy for rank

slide-28
SLIDE 28

Semidefinite programming (SDP)

Special class of convex optimization problems Relatively natural extension of linear programming (LP) ‘Efficient’ numerical solvers (interior point methods)

LP (std. form): x ∈ Rn

minimize c, x subject to aT

k x = bk k = 1, . . .

x ≥ 0

SDP (std. form): X ∈ Rn×n

minimize C, X subject to Ak, X = bk k = 1, . . . X 0 Standard inner product: C, X = Tr(C∗X)

slide-29
SLIDE 29

This is not really new...

Relaxation of quadratically constrained QP’s Shor (87) [Lower bounds on nonconvex quadratic optimization problems] Goemans and Williamson (95) [MAX-CUT] Ben-Tal and Nemirovskii (01) [Monograph] ... Similar approach for array imaging: Chai, Moscoso, Papanicolaou (11)

slide-30
SLIDE 30

From overdetermined to highly underdetermined

Quadratic equations b = A(xx∗) bk = |ak, x|2 k ∈ [m] Lift b = A(X) minimize Tr(X) subject to A(X) = b X 0

Have we made things worse?

  • verdetermined (m > n)

→ highly underdetermined (m ≪ n2)

slide-31
SLIDE 31

Exact phase retrieval via SDP

Quadratic equations b = A(xx∗) bk = |ak, x|2 k ∈ [m] Simplest model: ak independently and uniformly sampled on unit sphere

  • f Cn if x ∈ Cn (complex-valued problem)
  • f Rn if x ∈ Rn (real-valued problem)
slide-32
SLIDE 32

Exact phase retrieval via SDP

Quadratic equations b = A(xx∗) bk = |ak, x|2 k ∈ [m] Simplest model: ak independently and uniformly sampled on unit sphere

  • f Cn if x ∈ Cn (complex-valued problem)
  • f Rn if x ∈ Rn (real-valued problem)

Theorem (C. and Li (’12); C., Strohmer and Voroninski (’11))

Assume m n. With prob. 1 − O(e−γm), for all x ∈ Cn, only point in feasible set {X : A(X) = b and X 0} is xx∗

slide-33
SLIDE 33

Exact phase retrieval via SDP

Quadratic equations b = A(xx∗) bk = |ak, x|2 k ∈ [m] Simplest model: ak independently and uniformly sampled on unit sphere

  • f Cn if x ∈ Cn (complex-valued problem)
  • f Rn if x ∈ Rn (real-valued problem)

Theorem (C. and Li (’12); C., Strohmer and Voroninski (’11))

Assume m n. With prob. 1 − O(e−γm), for all x ∈ Cn, only point in feasible set {X : A(X) = b and X 0} is xx∗ Injectivity if m ≥ 4n − 2 (Balan, Bodmann, Casazza, Edidin ’09)

slide-34
SLIDE 34

Extensions to physical setups

Random masking + diffraction Similar theory: C. , Li and Soltanolkotabi (’13)

slide-35
SLIDE 35

Numerical results: noiseless recovery

(a) Smooth signal (real part) (b) Random signal (real part)

Figure: Recovery (with reweighting) of n-dimensional complex signal (2n unknowns) from 4n quadratic measurements (random binary masks)

slide-36
SLIDE 36

How is this possible?

How can feasible set {X 0} ∩ {A(X) = b} have a unique point? Intersection of

  • x

y y z

  • 0 with affine space
slide-37
SLIDE 37

Correct representation

Rank-1 matrices are on the boundary (extreme rays) of PSD cone

slide-38
SLIDE 38

My mental representation

slide-39
SLIDE 39

My mental representation

slide-40
SLIDE 40

My mental representation

slide-41
SLIDE 41

My mental representation

slide-42
SLIDE 42

My mental representation

slide-43
SLIDE 43

Numerical results: noiseless 2D images

50 100 150 200 250 50 100 150 200 250 50 100 150 200 250 50 100 150 200 250

  • riginal image

3 Gaussian masks

50 100 150 200 250 50 100 150 200 250 50 100 150 200 250 50 100 150 200 250

8 binary masks error with 8 binary masks

Courtesy

  • S. Marchesini (LBL)
slide-44
SLIDE 44

Numerical results: noisy 2D data

Low SNR (20dB) High SNR (60 dB)

Reconstructions from noisy data using 32 Gaussian random masks Similar output with 16 masks

slide-45
SLIDE 45

With noise

Figure: Relative MSE vs. signal-to-noise ratio on dB scale (binary masks)

slide-46
SLIDE 46

Story #2: Robust Principal Component Analysis

Collaborators: X. Li, Y. Ma, J. Wright

slide-47
SLIDE 47

The separation problem (Chandrasekahran et al.)

M = L + S M: data matrix (observed) L: low-rank (unobserved) S: sparse (unobserved)

slide-48
SLIDE 48

The separation problem (Chandrasekahran et al.)

M = L + S M: data matrix (observed) L: low-rank (unobserved) S: sparse (unobserved)

Problem: can we recover L and S accurately?

Again, missing information

slide-49
SLIDE 49

Motivation: robust principal component analysis (RPCA)

PCA sensitive to outliers: breaks down with one (badly) corrupted data point      x11 x12 . . . x1n x21 x22 . . . x2n . . . . . . . . . . . . xd1 xd2 . . . xdn      ❆

slide-50
SLIDE 50

Motivation: robust principal component analysis (RPCA)

PCA sensitive to outliers: breaks down with one (badly) corrupted data point      x11 x12 . . . x1n x21 x22 . . . x2n . . . . . . . . . . . . xd1 xd2 . . . xdn      = ⇒      x11 x12 . . . x1n x21 x22 . . . x2n . . . . . . . . . . . . xd1 ❆ . . . xdn     

slide-51
SLIDE 51

Robust PCA

Data increasingly high dimensional Gross errors frequently occur in many applications

Image processing Web data analysis Bioinformatics ... Occlusions Malicious tampering Sensor failures ...

     ❆ x12 . . . x1n x21 x22 . . . ❆ . . . . . . . . . . . . xd1 ❆ . . . xdn      Important to make PCA robust

slide-52
SLIDE 52

Gross errors

Movies Users         × × ❆ ❆ × ❆ × × × ❆ ×         Observe corrupted entries Yij = Lij + Sij (i, j) ∈ Ωobs L low-rank matrix S entries that have been tampered with (impulsive noise)

Problem

Recover L from missing and corrupted samples

slide-53
SLIDE 53

When does separation make sense?

M = L + S Sparse component cannot be low rank: S =      ∗ · · · ∗ · · · . . . . . . . . . . . . . . . . . . ∗ · · ·      Sparsity pattern will be assumed (uniform) random

slide-54
SLIDE 54

When does separation make sense?

M = L + S Sparse component cannot be low rank: S =      ∗ · · · ∗ · · · . . . . . . . . . . . . . . . . . . ∗ · · ·      Sparsity pattern will be assumed (uniform) random Low-rank component cannot be sparse: L =      ∗ ∗ ∗ ∗ · · · ∗ ∗ · · · . . . . . . . . . . . . . . . . . . · · ·     

slide-55
SLIDE 55

Sparse component cannot be low-rank

L =      x1 x2 · · · xn−1 xn x1 x2 · · · xn−1 xn . . . . . . . . . . . . . . . x1 x2 · · · xn−1 xn     

  • 1x∗

⇒ L + S =      ❆ x2 · · · xn−1 xn ❆ x2 · · · xn−1 xn . . . . . . . . . . . . . . . ❆ x2 · · · xn−1 xn     

slide-56
SLIDE 56

Low-rank component cannot be sparse

L =          x1 x2 x3 x4 · · · xn−1 xn x1 x2 x3 x4 · · · xn−1 xn · · · · · · . . . . . . . . . . . . . . . . . . · · ·          Incoherent condition [C. and Recht (’08)]: column and row spaces not aligned with coordinate axes (cannot have small subsets of rows and/or columns that are singular)

slide-57
SLIDE 57

Low-rank component cannot be sparse

M =           x1 x2 ❆ x4 · · · xn−1 xn x1 x2 ❆ x4 · · · xn−1 ❆ ❆ ❆ · · · ❆ · · · . . . . . . . . . . . . . . . . . . ❆ · · · ❆           Incoherent condition [C. and Recht (’08)]: column and row spaces not aligned with coordinate axes (cannot have small subsets of rows and/or columns that are singular)

slide-58
SLIDE 58

Demixing by convex programming

M = L + S L unknown (rank unknown) S unknown (# of entries = 0, locations, magnitudes all unknown)

slide-59
SLIDE 59

Demixing by convex programming

M = L + S L unknown (rank unknown) S unknown (# of entries = 0, locations, magnitudes all unknown)

Recovery via SDP

minimize ˆ L∗ + λ ˆ S1 subject to ˆ L + ˆ S = M See also Chandrasekaran, Sanghavi, Parrilo, Willsky (’09) nuclear norm: L∗ =

i σi(L) (sum of sing. values)

ℓ1 norm: S1 =

ij |Sij| (sum of abs. values)

slide-60
SLIDE 60

Exact recovery via SDP

min ˆ L∗ + λ ˆ S1

  • s. t.

ˆ L + ˆ S = M ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆

slide-61
SLIDE 61

Exact recovery via SDP

min ˆ L∗ + λ ˆ S1

  • s. t.

ˆ L + ˆ S = M

Theorem

L is n × n of rank(L) ≤ ρrn (log n)−2 and incoherent S is n × n, random sparsity pattern of cardinality at most ρsn2 Then with probability 1 − O(n−10), SDP with λ = 1/√n is exact: ˆ L = L, ˆ S = S Same conclusion for rectangular matrices with λ = 1/ √ max dim ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆ ❆

slide-62
SLIDE 62

Exact recovery via SDP

min ˆ L∗ + λ ˆ S1

  • s. t.

ˆ L + ˆ S = M

Theorem

L is n × n of rank(L) ≤ ρrn (log n)−2 and incoherent S is n × n, random sparsity pattern of cardinality at most ρsn2 Then with probability 1 − O(n−10), SDP with λ = 1/√n is exact: ˆ L = L, ˆ S = S Same conclusion for rectangular matrices with λ = 1/ √ max dim No tuning parameter! Whatever the magnitudes of L and S          × ❆ ❆ ❆ × ❆ ❆ ❆ × × ❆ ❆ × ❆ ❆ × ❆ ❆ ❆ ❆ × ❆ ❆ × × ❆ ❆ ❆ ❆ ❆ ❆ ❆ × × ❆ ❆         

slide-63
SLIDE 63

Phase transitions in probability of success

(a) RPCA, Random Signs (b) RPCA, Coherent Signs (c) Matrix Completion

L = XY T is a product of independent n × r i.i.d. N(0, 1/n) matrices

slide-64
SLIDE 64

Missing and corrupted

RPCA

min ˆ L∗ + λ ˆ S1

  • s. t.

ˆ Lij + ˆ Sij = Lij + Sij (i, j) ∈ Ωobs

        × ❆ ? ? × ? ? ? × ❆ ? ? × ? ? × ? ? ? ? × ? ? ❆ × ? ❆ ? ? ? ? ? × ❆ ? ?        

Same theorem: with high prob. λ = 1 √

  • frac. observed × max dim

= ⇒ ˆ L = L!

slide-65
SLIDE 65

Video surveillance

Sequence of 200 video frames (144 × 172 pixels) with a static background Problem: detect any activity in the foreground

… …

RPCA

slide-66
SLIDE 66

L + S background subtraction

slide-67
SLIDE 67

L + S background subtraction

From GoDec

slide-68
SLIDE 68

L + S reconstruction of MR angiography

L + S L S automatic and improved background suppression

Joint with R. Otazo and D. Sodickson

slide-69
SLIDE 69

Free-breathing MRI of the liver

NUFFT Standard L + S Motion-Guided L + S 12.8 fold acceleration min L∗ + λS1

  • s. t.

A(L + S) = y

Joint with R. Otazo and D. Sodickson

slide-70
SLIDE 70

Free-breathing MRI of the liver

NUFFT Standard L + S Motion-Guided L + S Temporal blurring

Joint with R. Otazo and D. Sodickson

slide-71
SLIDE 71

Free-breathing MRI of the kidneys

NUFFT Standard L + S Motion-Guided L + S 12.8 fold acceleration min L∗ + λS1

  • s. t.

A(L + S) = y

Joint with R. Otazo and D. Sodickson

slide-72
SLIDE 72

Free-breathing MRI of the kidneys

NUFFT Standard L + S Motion-Guided L + S

Joint with R. Otazo and D. Sodickson

slide-73
SLIDE 73

Story #3: Super-resolution

Collaborator: C. Fernandez-Granda

slide-74
SLIDE 74

Limits of resolution

In any optical imaging system, diffraction imposes fundamental limit on resolution

The physical phenomenon called diffraction is of the utmost importance in the theory of optical imaging systems (Joseph Goodman)

Interested in usual bandlimited imaging systems

slide-75
SLIDE 75

Pupil Airy disk Cross section

slide-76
SLIDE 76

Rayleigh resolution limit

Lord Rayleigh

slide-77
SLIDE 77

The super-resolution problem

  • bjective

data Retrieve fine scale information from low-pass data

Equivalent description: extrapolate spectrum Fundamental problem Radar Microscopy Spectroscopy Medical imaging Astronomy Geophysics ...

slide-78
SLIDE 78

Single molecule imaging

Microscope receives light from fluorescent molecules

Problem

Resolution is much coarser than size of individual molecules (low-pass data) Can we ‘beat’ the diffraction limit and super-resolve those molecules? Higher molecule density − → faster imaging

slide-79
SLIDE 79

Mathematical model

Signal x =

j ajδτj

aj ∈ C, τj ∈ [0, 1] Data y = Fnx: n = 2flo + 1 low-frequency coefficients (Nyquist sampling) y(k) = 1 e−i2πktx(dt) =

  • j

aje−i2πkτj k ∈ Z, |k| ≤ flo Resolution limit: (λlo/2 is Rayleigh distance) 1/flo = λlo

slide-80
SLIDE 80

Mathematical model

Signal x =

j ajδτj

aj ∈ C, τj ∈ [0, 1] Data y = Fnx: n = 2flo + 1 low-frequency coefficients (Nyquist sampling) y(k) = 1 e−i2πktx(dt) =

  • j

aje−i2πkτj k ∈ Z, |k| ≤ flo Resolution limit: (λlo/2 is Rayleigh distance) 1/flo = λlo

Question

Can we resolve the signal beyond this limit?

slide-81
SLIDE 81

Can you find the spikes?

Low-frequency data about spike train

slide-82
SLIDE 82

Can you find the spikes?

Low-frequency data about spike train

slide-83
SLIDE 83

Recovery by minimum total variation

Recovery by cvx prog.

min ˜ xTV subject to Fn ˜ x = y xTV =

  • |x(dt)| is continuous analog of discrete ℓ1 norm xℓ1 =

t |xt|

x =

  • j

ajδτj = ⇒ xTV =

  • j

|aj| Work on ℓ1 minimization: Logan, Donoho, Stark, Tropp, Elad, C. , Tao...

slide-84
SLIDE 84

Recovery by convex programming

y(k) = 1 e−i2πktx(dt) |k| ≤ flo

slide-85
SLIDE 85

Recovery by convex programming

y(k) = 1 e−i2πktx(dt) |k| ≤ flo

Theorem (C. and Fernandez Granda (2012))

If spikes are separated by at least 1.86 /flo := 1.86 λlo then min TV solution is exact! (Current state of the art 1.4λlo)

slide-86
SLIDE 86

Recovery by convex programming

y(k) = 1 e−i2πktx(dt) |k| ≤ flo

Theorem (C. and Fernandez Granda (2012))

If spikes are separated by at least 1.86 /flo := 1.86 λlo then min TV solution is exact! (Current state of the art 1.4λlo) Infinite precision! (Whatever the amplitudes)

slide-87
SLIDE 87

Recovery by convex programming

y(k) = 1 e−i2πktx(dt) |k| ≤ flo

Theorem (C. and Fernandez Granda (2012))

If spikes are separated by at least 1.86 /flo := 1.86 λlo then min TV solution is exact! (Current state of the art 1.4λlo) Infinite precision! (Whatever the amplitudes) Cannot go below λlo

slide-88
SLIDE 88

Recovery by convex programming

y(k) = 1 e−i2πktx(dt) |k| ≤ flo

Theorem (C. and Fernandez Granda (2012))

If spikes are separated by at least 1.86 /flo := 1.86 λlo then min TV solution is exact! (Current state of the art 1.4λlo) Infinite precision! (Whatever the amplitudes) Cannot go below λlo Can recover (2λlo)−1 = flo/2 = n/4 spikes from n low-freq. samples

slide-89
SLIDE 89

Recovery by convex programming

y(k) = 1 e−i2πktx(dt) |k| ≤ flo

Theorem (C. and Fernandez Granda (2012))

If spikes are separated by at least 1.86 /flo := 1.86 λlo then min TV solution is exact! (Current state of the art 1.4λlo) Infinite precision! (Whatever the amplitudes) Cannot go below λlo Can recover (2λlo)−1 = flo/2 = n/4 spikes from n low-freq. samples Essentially same result in higher dimensions

slide-90
SLIDE 90

Formulation as a finite-dimensional problem

Primal problem min xTV s. t. Fnx = y Infinite-dimensional variable x Finitely many constraints Dual problem max Rey, c s. t. F∗

nc∞ ≤ 1

Finite-dimensional variable c Infinitely many constraints (F∗

n c)(t) =

  • |k|≤flo

ckei2πkt

slide-91
SLIDE 91

Formulation as a finite-dimensional problem

Primal problem min xTV s. t. Fnx = y Infinite-dimensional variable x Finitely many constraints Dual problem max Rey, c s. t. F∗

nc∞ ≤ 1

Finite-dimensional variable c Infinitely many constraints (F∗

n c)(t) =

  • |k|≤flo

ckei2πkt

Semidefinite representability

|(F∗

n c)(t)| ≤ 1 for all t ∈ [0, 1] equivalent to

(1) there is Q Hermitian s. t.

  • Q

c c∗ 1

  • (2) Tr(Q) = 1

(3) sums along superdiagonals vanish: n−j

i=1 Qi,i+j = 0 for 1 ≤ j ≤ n − 1

slide-92
SLIDE 92

Dual solution

c: coeffs. of low-pass trig. poly.

k ckei2πkt interpolating sign of primal solution

slide-93
SLIDE 93

Super-resolution via semidefinite programming

slide-94
SLIDE 94

Super-resolution via semidefinite programming

slide-95
SLIDE 95

Super-resolution via semidefinite programming

  • 1. Solve semidefinite program to obtain dual solution
slide-96
SLIDE 96

Super-resolution via semidefinite programming

  • 2. Locate points at which corresponding polynomial has unit magnitude
slide-97
SLIDE 97

Super-resolution via semidefinite programming

Signal Estimate

  • 3. Estimate amplitudes via least squares
slide-98
SLIDE 98

Support-location accuracy

fc 25 50 75 100 Average error 6.66 10−9 1.70 10−9 5.58 10−10 2.96 10−10 Maximum error 1.83 10−7 8.14 10−8 2.55 10−8 2.31 10−8 For each fc, 100 random signals with |T| = fc/4 and ∆(T) ≥ 2/fc

slide-99
SLIDE 99

Example

Minimum separation : 1.5 λc

slide-100
SLIDE 100

Example

SNR 20 dB

Noisy Noiseless

slide-101
SLIDE 101

Example

SNR 20 dB

Signal Estimate

slide-102
SLIDE 102

Example

SNR 15 dB

Noisy Noiseless

slide-103
SLIDE 103

Example

SNR 15 dB

Signal Estimate

slide-104
SLIDE 104

Example

SNR 5 dB

Noisy Noiseless

slide-105
SLIDE 105

Example

SNR 5 dB

Signal Estimate

slide-106
SLIDE 106

Noise is amenable to convex modeling

Phase retrieval minimize b − A(X)1 subject to X 0 Robust PCA minimize L∗ + λS1 + γY − L − S2

F

Super-resolution minimize xTV subject to y − Fnx2

2 ≤ σ2

Stability in all cases, sometimes near the information theoretic limit

slide-107
SLIDE 107

Apologies: things I have not talked

Noise Algorithms Applications Avalanche of related works

slide-108
SLIDE 108

A small sample of papers I have greatly enjoyed

Phase retrieval

Netrapalli, Jain, Sanghavi, Phase retrieval using alternating minimization (’13) Waldspurger, d’Aspremont, Mallat, Phase recovery, MaxCut and complex semidefinite programming (’12)

Robust PCA

Gross, Recovering low-rank matrices from few coefficients in any basis (’09) Chandrasekaran, Parrilo and Willsky, Latent variable graphical model selection via convex optimization (’11) Hsu, Kakade and Zhang, Robust matrix decomposition with outliers (’11)

Super-resolution

Kahane, Analyse et synth` ese harmoniques (’11) Slepian, Prolate spheroidal wave functions, Fourier analysis, and uncertainty. V - The discrete case (’78)

slide-109
SLIDE 109

Summary

Three important problems with missing data

Phase retrieval Matrix completion/RPCA Super-resolution

Three simple and model-free recovery procedures via convex programming Three near-perfect solutions — Three theorems Three fundamental modeling/computational tools Matrices Probability (Convex) Optimization

slide-110
SLIDE 110

Numerical results: noiseless recovery

(a) Smooth signal (real part) (b) Random signal (real part)

Figure: Recovery (with reweighting) of n-dimensional complex signal (2n unknowns) from 4n quadratic measurements (random binary masks)

slide-111
SLIDE 111

About separation: sparsity is not enough!

CS: sparse signals are ‘away’ from null space of sampling operator Super-res: this is not the case Signal Spectrum

10

−10

10

−5

10

slide-112
SLIDE 112

About separation: sparsity is not enough!

CS: sparse signals are ‘away’ from null space of sampling operator Super-res: this is not the case Signal Spectrum

10

−10

10

−5

10 10

−10

10

−5

10

slide-113
SLIDE 113

Analysis via prolate spheroidal functions

David Slepian If distance between spikes less than λlo/2 (Rayleigh), problem hopelessly ill posed