Generalized Row-Action Methods for Tomographic Imaging Sparse Tomo - - PowerPoint PPT Presentation

generalized row action methods for tomographic imaging
SMART_READER_LITE
LIVE PREVIEW

Generalized Row-Action Methods for Tomographic Imaging Sparse Tomo - - PowerPoint PPT Presentation

Generalized Row-Action Methods for Tomographic Imaging Sparse Tomo Days, Technical University of Denmark, March 27, 2014 Martin S. Andersen joint work with Per Christian Hansen Scientific Computing Section Outline Algebraic methods for X-ray


slide-1
SLIDE 1

Generalized Row-Action Methods for Tomographic Imaging

Sparse Tomo Days, Technical University of Denmark, March 27, 2014 Martin S. Andersen joint work with Per Christian Hansen Scientific Computing Section

slide-2
SLIDE 2

Outline

1

Algebraic methods for X-ray computed tomography

2

Relaxed incremental proximal methods

3

Numerical experiments

1 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-3
SLIDE 3

X-ray Computed Tomography

Measurement model I1 = I0 exp

  • l

µ(u1, u2) ds

  • log(I0/I1) ≈ aT

i x

b = Ax + e I0 I1 Parallel beam measurement geometry

Detector X-ray source

2 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-4
SLIDE 4

Algebraic reconstruction technique

Projection on hyperplane Hi = {x | aT

i x = bi}

PHi(xk) = argmin

x∈Hi

x − xk = xk − ai(aT

i xk − bi)

ai2 Kaczmarz’s method / ART xk+1 = ρPHik(xk) + (1 − ρ)xk, ik ∈ {1, . . . , m} Relaxation parameter ρ ∈ (0, 2)

3 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-5
SLIDE 5

Example — consistent system

17 equations, ρ = 1.0 35 equations, ρ = 1.0

4 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-6
SLIDE 6

Example — consistent system

17 equations, ρ = 0.5 35 equations, ρ = 0.5

4 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-7
SLIDE 7

Example — inconsistent system

17 equations, ρ = 1.0 35 equations, ρ = 1.0

5 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-8
SLIDE 8

Example — inconsistent system

17 equations, ρ = 0.5 35 equations, ρ = 0.5

5 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-9
SLIDE 9

Example — inconsistent system

17 equations, ρ = 0.2 35 equations, ρ = 0.2

5 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-10
SLIDE 10

Example — inconsistent system, randomization

17 equations, ρ = 1.0 35 equations, ρ = 1.0

6 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-11
SLIDE 11

Example — inconsistent system, randomization

17 equations, ρ = 0.5 35 equations, ρ = 0.5

6 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-12
SLIDE 12

Example — inconsistent system, randomization

17 equations, ρ = 0.2 35 equations, ρ = 0.2

6 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-13
SLIDE 13

Example — underdetermined system

2 1 1 2 x 2 1 1 2 y 5 5 z

7 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-14
SLIDE 14

Example — consistent system

ρ = 1.0

8 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-15
SLIDE 15

Example — consistent system

ρ = 0.5

8 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-16
SLIDE 16

Example — consistent system

ρ = 1.5

8 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-17
SLIDE 17

Tomographic image reconstruction

Ill-posed inverse problem — we need regularization! Incorporate a priori knowledge in reconstruction problem

  • spatial information (smoothness, piecewise constant/affine, ...)
  • bounds (nonnegativity, box constraints, ...)
  • sparsity
  • ...

Large-scale optimization problem

  • gradient computation is expensive
  • may not be differentiable

9 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-18
SLIDE 18

Superiorization

Kaczmarz’s method is pertubation resilient (Herman et al., 2009) xk+1 = P

  • xk + tkvk
  • ,

P = PHm ◦ · · · ◦ PH2 ◦ PH1 Converges if Ax = b is consistent and tk → 0 for k → ∞ Perturbed iteration yields a “superior” solution in some sense

10 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-19
SLIDE 19

Incremental methods (I)

minimize

m

  • i=1

fi(x) subject to x ∈ C Incremental (sub)gradient iteration: xk+1 = PC

  • xk − tk

∇fik(xk)

  • ,
  • ∇fik(xk) ∈ ∂fik(xk)
  • sublinear rate of convergence — initial convergence often very fast
  • diminishing stepsize or “oscillation” that depends on stepsize
  • goes back to Kibardin (1980), Litvakov (1966)
  • Bertsekas (1996): incremental least-squares and extended Kalman filter

11 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-20
SLIDE 20

Incremental methods (II)

Incremental proximal iteration: xk+1 = argmin

x∈C

  • fik(x) + 1

2tk x − xk2 equivalently xk+1 = xk − tkgk+1, gk+1 ∈ ∂fik(xk+1) + NC(xk+1) Linearized proximal iteration: let gk ∈ ∂fik(xk) xk+1 = PC

  • argmin

x∈Rn

  • fik(xk) + gT

k x + 1

2tk x − xk2 = PC

  • xk − tkgk
  • 12 / 28 DTU Compute, Technical University of Denmark

Generalized Row-Action Methods March 27, 2014

slide-21
SLIDE 21

Incremental proximal gradient methods (I)

minimize m

i=1

  • gi(x) + hi(x)
  • subject to

x ∈ C Algorithm 1 (Bertsekas, 2011) zk = argmin

u∈Rn

  • gik(u) + 1

2tk u − xk2 xk+1 = PC

  • zk − tk

∇hik(zk)

  • Interpretation

xk+1 = PC

  • xk − tk

∇gik(zk) − tk ∇hik(zk)

  • 13 / 28 DTU Compute, Technical University of Denmark

Generalized Row-Action Methods March 27, 2014

slide-22
SLIDE 22

Incremental proximal gradient methods (II)

Algorithm 2 (Bertsekas, 2011) zk = xk − tk ∇hik(xk) xk+1 = argmin

u∈C

  • gik(u) + 1

2tk u − zk2 Interpretation xk+1 = argmin

u∈C

  • gik(u) +

∇hik(xk)T u + 1 2tk u − xk2 = PC

  • xk − tk

∇gik(xk+1) − tk ∇hik(xk)

  • 14 / 28 DTU Compute, Technical University of Denmark

Generalized Row-Action Methods March 27, 2014

slide-23
SLIDE 23

Relaxed incremental proximal gradient methods

Algorithm 1 wk = argmin

u∈Rn

  • gik(u) + 1

2tk u − xk2 zk = wk − tk ∇hik(wk) xk+1 = PC

  • ρ zk + (1 − ρ)xk
  • Algorithm 2

wk = wk − tk ∇hik(wk) zk = argmin

u∈Rn

  • gik(u) + 1

2tk u − wk2 xk+1 = PC

  • ρ zk + (1 − ρ)xk
  • 15 / 28 DTU Compute, Technical University of Denmark

Generalized Row-Action Methods March 27, 2014

slide-24
SLIDE 24

Convergence results

Cyclic order or random order?

  • cyclic order yields worst-case performance bounds
  • random order yields expected performance bounds

Constant stepsize or diminishing stepize?

  • constant stepsize yields convergence within an error bound
  • diminishing stepsize yields exact convergence

Randomized cyclic order works well in practice

16 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-25
SLIDE 25

R-IPG vs. ART

minimize (1/2) m

i=1(aT i x − bi)2

subject to x ∈ C Let gi(x) = (1/2)(aT

i x − bi)2 and hi(x) = 0:

xk+1 = PC

  • xk − ρaik(aT

ikxk − bik)

t−1

k

+ aik2

  • Interpretation: ART with damped step size

17 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-26
SLIDE 26

R-IPG vs. ART — numerical example

2D tomography problem

  • 256×256 Shepp–Logan phantom
  • n = 2562 = 65536 variables
  • p = 120 uniformly spaced angles
  • r = 362 parallel rays
  • m = pr = 43440 measurements
  • Gaussian noise: ei ∼ N(0, 0.02 · b∞)

Algorithm: R-IPG with constant step size

18 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-27
SLIDE 27

R-IPG vs. ART — numerical example

2 4 6 8 10 0.2 0.4 0.6 0.8 1 Norm of relative error tk = 0.001 2 4 6 8 10 0.2 0.4 0.6 0.8 1 tk = 0.01 2 4 6 8 10 0.2 0.4 0.6 0.8 1 Cycles tk = 1.0 2 4 6 8 10 0.2 0.4 0.6 0.8 1 Cycles Norm of relative error tk = 0.1 ρ = 0.1 ρ = 0.5 ρ = 1.0 ρ = 1.5 ρ = 1.9

19 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-28
SLIDE 28

Regularized data fitting

minimize f(x) ≡ (1/2)Ax − b2 + λ h(x) subject to x ∈ C Example: express f(x) as f(x) =

m

  • i=1

(1/2)(aT

i x − bi)2

  • gi(x)

+

m

  • i=1

λ mh(x)

hi(x)

20 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-29
SLIDE 29

Total-variation regularization

minimize (1/2)Ax − b2

2 + λ n i=1 Dix2

subject to x ∈ C Let gi(x) = Aix − bi2

2 and hi(x) = (λ/p) n i=1 Dix2

zk = argmin

u∈Rn

  • gik(u) + 1

2tk u − xk2 = xk − AT

ik(AikAT ik + t−1 k I)−1(Aikxk − bik)

xk+1 = PC

  • ρ(zk − tk

∇hik(zk)) + (1 − ρ)xk

  • 21 / 28 DTU Compute, Technical University of Denmark

Generalized Row-Action Methods March 27, 2014

slide-30
SLIDE 30

Total-variation regularization — numerical example

2D tomography problem

  • 512×512 Shepp–Logan phantom
  • n = 5122 = 262144 variables
  • p = 60 uniformly spaced angles
  • r = 724 parallel rays
  • m = pr = 43440 measurements
  • Gaussian noise: ei ∼ N(0, 0.01 · b∞)

Algorithm: R-IPG1 with diminishing step-size sequence

22 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-31
SLIDE 31

Regularization curve

100 101 102 0.12 0.14 0.16 0.18 0.2 λ Norm of relative error

23 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-32
SLIDE 32

Reference

Original Filtered backprojection Total variation regularized LS

24 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-33
SLIDE 33

Relaxed incremental proximal subgradient method

5 10 15 20 10−1 100 Norm of relative error t0 = 0.001 5 10 15 20 10−1 100 t0 = 0.01 5 10 15 20 10−1 100 Cycles t0 = 1.0 5 10 15 20 10−1 100 Cycles Norm of relative error t0 = 0.1 ρ = 0.1 ρ = 0.5 ρ = 1.0 ρ = 1.5 ρ = 1.9

25 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-34
SLIDE 34

Primal–dual first-order method (Chambolle & Pock)

50 100 150 200 10−1 100 Iterations Norm of relative error ρ = 1.9 ρ = 1.5 ρ = 1.0 ρ = 0.5 ρ = 0.1

26 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-35
SLIDE 35

2 5 10 20 50

Chambolle–Pock

2 5 10 20 50

Relaxed IPG

27 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-36
SLIDE 36

Summary

  • relaxed incremental proximal gradient methods
  • slow global rate of convergence
  • often fast initial rate rate of convergence
  • hybrid methods that transition from incremental to non-incremental method

28 / 28 DTU Compute, Technical University of Denmark Generalized Row-Action Methods March 27, 2014

slide-37
SLIDE 37

Thank you for listening! mskan@dtu.dk http://compute.dtu.dk/~mskan