Super-resolution and sensor calibration in imaging Wenjing Liao - - PowerPoint PPT Presentation

super resolution and sensor calibration in imaging
SMART_READER_LITE
LIVE PREVIEW

Super-resolution and sensor calibration in imaging Wenjing Liao - - PowerPoint PPT Presentation

Super-resolution and sensor calibration in imaging Wenjing Liao School of Mathematics, Georgia Institute of Technology ICERM September 25, 2017 My collaborators Albert Fannjiang Weilin Li UC Davis UMD Yonina C. Eldar Sui Tang Tel-Aviv


slide-1
SLIDE 1

Super-resolution and sensor calibration in imaging

Wenjing Liao School of Mathematics, Georgia Institute of Technology ICERM September 25, 2017

slide-2
SLIDE 2

My collaborators

Albert Fannjiang Weilin Li UC Davis UMD Yonina C. Eldar Sui Tang Tel-Aviv University, Israel JHU

2 / 32

slide-3
SLIDE 3

Outline

1

Super-resolution Resolution in imaging Super-resolution limit and min-max error Super-resolution algorithms

2

Sensor calibration Problem formulation Uniqueness An optimization approach Numerical simulations

3 / 32

slide-4
SLIDE 4

Source localization with sensor array

S point sources located at ωj ∈ [0, 1) with amplitudes xj far field ✄ ✄ ✄ ✄ M Sensors aperture Point sources: x(t) = S

j=1 xjδ(t − ωj),

ωj ∈ [0, 1) Measurement at the mth sensor, m = 0, . . . , M − 1: ym =

S

  • j=1

xje−2πimωj + em Measurements: {ym : m = 0, . . . , M − 1} To recover: source locations {ωj}S

j=1 and source amplitudes {xj}S j=1.

4 / 32

slide-5
SLIDE 5

Rayleigh criterion

ˆ x(ω) =

M−1

  • m=0

ym e2πimω M Rayleigh length = 1/M

5 / 32

slide-6
SLIDE 6

Inverse Fourier transform and the MUSIC algorithm

Multiple Signal Classification (MUSIC): [Schmidt 1981]

42 44 46 48 50 52 54 56 2 4 6 8 10 12 14 1013 42 44 46 48 50 52 54 56 100 200 300 400 500 600

noise-free noisy

6 / 32

slide-7
SLIDE 7

Interesting questions

1 What is the super-resolution limit of the “best” algorithm? 2 What is the super-resolution limit of a specific algorithm? ◮ MUSIC [Schmidt 1981] ◮ ESPRIT [Roy and Kailath 1989] ◮ the matrix pencil method [Hua and Sarkar 1990] 7 / 32

slide-8
SLIDE 8

Existing works

1 Super-resolution limit with continuous measurements ◮ Donoho 1992, Demanet and Nguyen 2015 2 Performance guarantees for well separated point sources ◮ Total variation minimization [Cand`

es and Fernandez-Granda 2013,2014, Tang, Bhaskar, Shah and Recht 2013, Duval and Peyr´ e 2015, Li 2017]

◮ Greedy algorithms [Duarte and Baraniuk 2013, Fannjiang and L. 2012] ◮ MUSIC [L. and Fannjiang 2016] ◮ The matrix pencil method [Moitra 2015] 3 Performance guarantees for super-resolution ◮ Total variation min for positive sources [Morgenshtern and Cand`

es 2016] or sources with certain sign pattern [Benedetto and Li 2016]

◮ Lasso for positive sources [Denoyelle, Duval and Peyr´

e 2016]

8 / 32

slide-9
SLIDE 9

Discretization on a fine grid

1 ω1 ω2 ω3 ω4

1 N ≪ 1 M

Point sources: µ = N−1

n=0 xnδn/N with x ∈ CN S

Measurement vector y = Φx + e where Φ ∈ CM×N is the first M rows of the N × N DFT matrix: Φm,n = e−2πimn/N and e2 ≤ δ. Super-resolution factor (SRF) := N

M

9 / 32

slide-10
SLIDE 10

Connection to compressive sensing

Sensing matrices contain certain rows of the DFT matrix.

(a) compressive sensing (b) super-resolution

10 / 32

slide-11
SLIDE 11

Min-max error

Definition (S-min-max error)

Fix positive integers M, N, S such that S ≤ M ≤ N and let δ > 0. The S-min-max error is E(M, N, S, δ) = inf

˜ x=˜ x(y,M,N,S,δ)∈CN y=Φx+e

sup

x∈CN

S

sup

e∈CM:e2≤δ

˜ x − x2.

11 / 32

slide-12
SLIDE 12

Sharp bound on the min-max error

Theorem (Li and L. 2017)

There exist constants A(S), B(S), C(S) such that:

1 Lower bound. If M ≥ 2S and N ≥ C(2S)M3/2, then

E(M, N, S, δ) ≥ δ 2B(2S) √ M SRF2S−1.

2 Upper bound. If M ≥ 4S(2S + 1) and N ≥ M2/(2S2), then

E(M, N, S, δ) ≤ 2δ A(2S) √ M SRF2S−1. The best algorithm in the upper bound: min z0 subject toΦz − y2 ≤ δ

  • W. Li and W. Liao, “Stable super-resolution limit and smallest singular value of restricted Fourier matrices,”preprint,

arXiv:1709.03146. 12 / 32

slide-13
SLIDE 13

Multiple Signal Classification (MUSIC)

Pioneering work: Prony 1795 MUSIC in signal processing: Schmidt 1981 MUSIC in imaging: Devaney 2000, Devaney, Marengo and Gruber 2005, Cheney 2001, Kirsch 2002 Related: the linear sampling method [Cakoni, Colton and Monk 2011], factorization method [Kirsch and Grinsberg 2008]

13 / 32

slide-14
SLIDE 14

MUSIC

Assumption: S is known. ym =

S

  • j=1

xje−2πimωj, m = 0, . . . , M − 1. H = Hankel(y) =      y0 y1 . . . yM−L y1 y2 . . . yM−L+1 . . . . . . . . . . . . yL−1 yL . . . yM−1      = ΦL

  • L×S

X

  • S×S

(ΦM−L+1)T

  • S×(M−L+1)

where X = diag(x1, . . . , xS) φL(ω) =

  • 1

e−2πiω . . . e−2πi(L−1)ωT ∈ CL ΦL = [φL(ω1) . . . φL(ωS)] ∈ CL×S.

14 / 32

slide-15
SLIDE 15

MUSIC with noiseless measurements

H = ΦLX(ΦM−L+1)T Suppose {ωj}S

j=1 are distinct.

1 If L ≥ S, rank(ΦL) = S. 2 If M − L + 1 ≥ S, Range(H) = Range(ΦL). 3 If L ≥ S + 1, rank([ΦL φL(ω)]) = S + 1 if and only if ω /

∈ {ωj}S

j=1.

Theorem

If L ≥ S + 1 and M − L + 1 ≥ S, ω ∈ {ωj}S

j=1 iff φL(ω) ∈ Range(H).

Exact recovery with M ≥ 2S regardless of the support .

15 / 32

slide-16
SLIDE 16

φL(ω1) φL(ω2) φL(ω3) φL(ω), ω / ∈ {ωj}S

j=1

Range(H) = Range(ΦL) signal space noise space Noise-space correlation function: N(ω) = PnoiseφL(ω)2

φL(ω)2

Imaging function: J (ω) =

1 N(ω)

N(ωj) = 0 and J (ωj) = ∞, j = 1, . . . , S.

16 / 32

slide-17
SLIDE 17

MUSIC with noisy measurements

Three sources separated by 0.5 RL, e ∼ N(0, σ2IM)

19.5 20 20.5 21 21.5 22 22.5 dist = 2.7756e-15 RL 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 ×1014 Imaging function imaging function exact frequencies

(c) σ = 0

19.5 20 20.5 21 21.5 22 22.5 dist = 0.06 RL 200 400 600 800 1000 1200 1400 1600 Imaging function imaging function exact frequencies

(d) σ = 0.001

19.5 20 20.5 21 21.5 22 22.5 dist = 3.96 RL 20 40 60 80 100 120 140 160 180 Imaging function imaging function exact frequencies

(e) σ = 0.01

Recall upper bound of the min-max error E(M, N, S, δ) δ √ M SRF2S−1 The noise that the “best” algorithm can handle is δ ∼

  • 1

SRF

2S−1.

17 / 32

slide-18
SLIDE 18

Phase transition

S consecutive point sources on the grid with spacing 1/N Support error: d({ωj}S

j=1, {ˆ

ωj}S

j=1)

Noise e ∼ N(0, σ2IM) + i · N(0, σ2IM), so Ee2 = √ 2Mσ.

log2(bottleneck error / separation)

  • 1
  • 0.8
  • 0.6
  • 0.4
  • 0.2
  • log10SRF
  • 3.5
  • 3
  • 2.5
  • 2
  • 1.5

noise level log10σ

  • 8
  • 6
  • 4
  • 2

2 4 6 8

log2(bottleneck error / separation)

  • 1
  • 0.8
  • 0.6
  • 0.4
  • 0.2
  • log10SRF
  • 3.5
  • 3
  • 2.5
  • 2
  • 1.5

noise level log10σ

  • 8
  • 6
  • 4
  • 2

2 4 6 8

log2(bottleneck error / separation)

  • 1
  • 0.8
  • 0.6
  • 0.4
  • 0.2
  • log10SRF
  • 3.5
  • 3
  • 2.5
  • 2
  • 1.5

noise level log10σ

  • 8
  • 6
  • 4
  • 2

2 4 6 8

S = 2 S = 3 S = 4

Figure: The average log2[ Support error

1/N

] over 100 trials with respect to log10

1 SRF

(x-axis) and log10 σ (y-axis).

18 / 32

slide-19
SLIDE 19

Super-resolution limit of MUSIC

The phase transition curve is σ ∼

  • 1

SRF p(S) where 2S − 1 ≤ p(S) ≤ 2S.

  • 1
  • 0.9
  • 0.8
  • 0.7
  • 0.6
  • 0.5
  • 0.4
  • 0.3
  • 0.2
  • 0.1
  • log10SRF
  • 3.6
  • 3.4
  • 3.2
  • 3
  • 2.8
  • 2.6
  • 2.4
  • 2.2
  • 2
  • 1.8
  • 1.6

noise level log10σ Phase transition curve

S=2 slope = 3.3137 S=3 slope = 5.2149 S=4 slope = 7.6786 S=5 slope = 9.9286

Future work: Support error by MUSC SRFp(S) · σ.

19 / 32

slide-20
SLIDE 20

Outline

1

Super-resolution Resolution in imaging Super-resolution limit and min-max error Super-resolution algorithms

2

Sensor calibration Problem formulation Uniqueness An optimization approach Numerical simulations

20 / 32

slide-21
SLIDE 21

Sensor calibration

Measurement at the m-th sensor, m = 0, . . . , M − 1: ym(t) = gm

S

  • j=1

xj(t)e−2πimωj + em(t) Multiple snapshots of measurements: {ym(t), m = 0, . . . , M − 1, t ∈ Γ} To recover: Calibration parameters g = {gm}M−1

m=0 ∈ CM

Source locations {ωj}S

j=1 and source amplitudes xj(t)

21 / 32

slide-22
SLIDE 22

Assumptions

Matrix form: y(t)

  • CM

= diag(g)

CM×M

A

  • CM×S

x(t)

  • CS

+ e(t)

  • CM

An,j = e−2πimωj

x(t) = [x1(t) . . . xS(t)]T, y(t) = [y0(t) . . . yM−1(t)]T, e(t) = [e0(t) . . . eM−1(t)]T

Assumptions:

1 |gm| = 0, m = 0, . . . , M − 1; 2 Ex(t) = 0 and Ee(t) = 0; 3 Rx := Ex(t)x∗(t) = diag({γ2

j }S j=1);

4 Ex(t)e∗(t) = 0; 5 Ee(t)e∗(t) = σ2IM where σ represents noise level. 22 / 32

slide-23
SLIDE 23

Uniqueness up to a trivial ambiguity

Trivial ambiguity: {˜ g, {˜ ωj}S

j=1, ˜

x(t)} is called equivalent to {g, {ωj}S

j=1, x(t)} up to a trivial ambiguity if there exist c0 > 0, c1, c2 ∈ R:

˜ g = {˜ gm = c0ei(c1+mc2)gm}M−1

m=0

˜ S = {˜ ωj : ˜ ωj = ωj − c2/(2π)}S

j=1

˜ x(t) = x(t)c−1

0 e−ic1.

Uniqueness with infinite snapshots of noiseless measurements: Let fm = s

j=1 γ2 j e2πimωj, m = 0, . . . , M − 1.

Theorem

Suppose |f1| > 0 and M ≥ S + 1. Let {g, {ωj}S

j=1, x(t)} be a solution to

the calibration problem. If there is another solution {˜ g, {˜ ωj}S

j=1, ˜

x(t)}, then {˜ g, {˜ ωj}S

j=1, ˜

x(t)} is equivalent to {g, {ωj}S

j=1, x(t)}.

23 / 32

slide-24
SLIDE 24

Covariance matrix

Pioneering work: Full algebraic method [Paulraj and Kailath, 1985], Partial algebraic method [Wylie, Roy and Schmitt, 1993] Ry := Ey(t)y∗(t) = diag(g)ARxA∗diag(¯ g) H : CM → CM×M : H(f ) :=        f0 ¯ f1 ... ¯ fN−1 f1 f0 ... ¯ fN−2 ... ... ... ... fN−1 fN−2 ... f0        = ARxA∗. Then Ry = diag(g)H(f )diag(¯ g) Ry

m,n = gm¯

gnfm−n When f1 = 0, the diagonal and subdiagonal entries in Ry determine the solution up to a trivial ambiguity.

24 / 32

slide-25
SLIDE 25

Algebraic methods

Sensitivity of the partial algebraic method: N ≥ s + 1, |f1| > 0 and sources are separated by 1/M. Empirical covariance matrix is computed with L snapshots of measurements. We proved that, E min

c0>0,c1,c2∈R max m |c0

gm − ei(c1+mc2)gm| ≤ O max(σ, σ2) √ L

  • ,

Partial algebraic method: only diagonal and subdiagonal entries in the covariance matrix are used Full algebraic method: problem of phase wrapping

25 / 32

slide-26
SLIDE 26

An optimization approach

Ry = GARxA∗G ∗ = diag(g)H(f )diag(¯ g) Optimization problem: min

g,f∈CM L(g, f) :=

  • diag(g)H(f)diag(¯

g) − Ry

  • 2

F .

If Ry = Ry, the global minimizer of L is equivalent to the ground truth (g, f ).

26 / 32

slide-27
SLIDE 27

Regularized optimization

Goal: prevent g → ∞ and f → 0 (or vice versa)

  • n0 is an estimator of n0 := g2f from the partial algebraic method.

Regularized optimization: min

g,f∈CN ˜

L(g, f) := L(g, f) + G(g, f) G(g, f) = ρ

  • G0

f2 2 n0

  • + G0

g2 √2 n0

  • where G0(z) = (max(z − 1, 0))2 and ρ ≥ 3

n0+Ry− RyF ( √ 2−1)2

Initialization: (g0, f0) : g02 ≤ √2 n0, f0 ≤ √2 n0 Feasible set: N

n0 = {(g, f) : g2 ≤ 2√

n0, f ≤ 2√ n0}

27 / 32

slide-28
SLIDE 28

Wirtinger gradient descent

for k = 1, 2, . . . , gk = gk−1 − ηk∇g ˜ L(gk−1, fk−1) fk = fk−1 − ηk∇f ˜ L(gk−1, fk−1) end

Theorem (Eldar, L. and Tang)

If the step length is chosen such that

ηk ≤ 2 146 n0 max( √

  • n0,

4

  • n0) + 8

n0 + 16 max( √

  • n0,

4

  • n0)Ry −

RyF +

8ρ min( n0,√

  • n0)

,

then Wirtinger gradient descent gives rise to (gk, fk) ∈ N

n0, and

∇ ˜ L(gk, fk) → 0, as k → ∞.

  • Y. C. Eldar, W. Liao and S. Tang, “Sensor calibration for off-the-grid spectral estimation,”preprint, arXiv: 1707.03378.

28 / 32

slide-29
SLIDE 29

Sensitivity to the number of snapshots

1 the partial algebraic method 2 our optimization approach 3 an alternating minimization: [Friedlander and Weiss 1990]

20 sources separated by 2/M and noise level σ = 2

2 2.5 3 3.5 log10(#snapshot)

  • 2
  • 1.8
  • 1.6
  • 1.4
  • 1.2
  • 1
  • 0.8
  • 0.6
  • 0.4
  • 0.2

log10(Relative calibration error) log10(Calibration error) versus log10(#snapshot)

FriedlanderWeiss σ = 2 algebraic σ = 2: slope = -0.51352 wirtinger σ = 2: slope = -0.53044 2 2.5 3 3.5

log10(#snapshot) 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 success rate of support recovery Success rate versus log10(#snapshot)

FriedlanderWeiss σ = 2 algebraic σ = 2 wirtinger σ = 2

Relative calibration error versus L Support success rate versus L

Observation: Calibration error = O(L− 1

2 ) 29 / 32

slide-30
SLIDE 30

Sensitivity to noise level σ

20 sources separated by 2/M and L = 500

  • 0.5

0.5 1 log10(σ)

  • 1.5
  • 1
  • 0.5

log10(calibration error in the unit of RL) log10(calibration error) versus log10(σ)

FriedlanderWeiss L=500 slope = 1.3439 algebraic L=500 slope = 1.0248 wirtinger L=500 slope = 1.0565

  • 0.5

0.5 1 log10(σ) 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 log10(success rate of support recovery) log10(success rate) versus log10( σ) FriedlanderWeiss L=500 algebraic L=500 wirtinger L=500

Relative calibration error versus σ Support success rate versus σ

Observation: Calibration error = O(σ)

30 / 32

slide-31
SLIDE 31

Conclusion

1 Super-resolution ◮ Resolution limit and a sharp bound on the min-max error ◮ Resolution limit of the MUSIC algorithm 2 Sensor calibration ◮ Uniqueness with infinite snapshots of noiseless data ◮ The partial algebraic method and a stability analysis ◮ An optimization approach and convergence to a stationary point 31 / 32

slide-32
SLIDE 32

Thank you for your attention! Wenjing Liao Georgia Institute of Technology wliao60@gatech.edu http://people.math.gatech.edu/~wliao60

32 / 32