Numerical Methods for Ill-Posed Problems I Lothar Reichel Summer - - PowerPoint PPT Presentation

numerical methods for ill posed problems i lothar reichel
SMART_READER_LITE
LIVE PREVIEW

Numerical Methods for Ill-Posed Problems I Lothar Reichel Summer - - PowerPoint PPT Presentation

Numerical Methods for Ill-Posed Problems I Lothar Reichel Summer School on Applied Analysis TU Chemnitz, Oct. 4-8, 2010 Outline of Lecture 1: Inverse and ill-posed problems The singular value decomposition Tikhonov regularization


slide-1
SLIDE 1

Numerical Methods for Ill-Posed Problems I Lothar Reichel Summer School on Applied Analysis TU Chemnitz, Oct. 4-8, 2010

slide-2
SLIDE 2

Outline of Lecture 1:

  • Inverse and ill-posed problems
  • The singular value decomposition
  • Tikhonov regularization
slide-3
SLIDE 3

Inverse problems

  • Inverse problems arise when one seeks to determine

the cause of an observed effect. – Helioseismology: Determine the structure of the sun by measurements from earth or space. – Medical imaging, e.g., electrocardiographic imaging, computerized tomography. – Image restoration: Determine the unavailable exact image from an available contaminated version.

  • Inverse problems often are ill-posed.
slide-4
SLIDE 4

Ill-posed problems

A problem is said to be ill-posed if it has at least one of the properties:

  • the problem does not have a solution,
  • the problem does not have a unique solution,
  • the solution does not depend continuously on the

data.

slide-5
SLIDE 5

Linear discrete ill-posed problems

Ax = b arise from the discretization of linear ill-posed problems (Fredholm integral equations of the 1st kid) or, naturally, in discrete form (image restoration).

  • The matrix A is of ill-determined rank, possibly
  • singular. System may be inconsistent.
  • The right-hand side b represents available data that

generally is contaminated by an error.

slide-6
SLIDE 6

Available contaminated, possibly inconsistents, linear system Ax = b (1) Unavailable associated consistent linear system with error-free right-hand side Ax = ˆ b (2) Let ˆ x denote the desired solution of (2), e.g., the minimal-norm solution. Task: Determine an approximate solution of (1) that is a good approximation of ˆ x.

slide-7
SLIDE 7

Define the error e := ˆ b − b noise and let ǫ := e The choice of solution method depends on whether an estimate of ǫ is available. How much damage can a little noise really do? How much noise requires the use of special techniques?

slide-8
SLIDE 8

Example 1: Fredholm integral equation of the 1st kind

π

0 exp(−st)x(t)dt = 2sinh(s)

s , 0 ≤ s ≤ π 2. Determine solution x(t) = sin(t). Discretize integral by Galerkin method using piecewise constant functions. Code baart from Regularization Tools.

slide-9
SLIDE 9

This gives a linear system of equations Ax = ˆ b, A ∈ R200×200, ˆ b ∈ R200. A is numerically singular. Let the “noise” vector e in b have normally distributed entries with mean zero and ǫ = e = 10−3b b := ˆ b + e i.e., 0.1% relative noise

slide-10
SLIDE 10

20 40 60 80 100 120 140 160 180 200 0.17 0.18 0.19 0.2 0.21 0.22 0.23 0.24 0.25 0.26 right−hand side no added noise added noise/signal=1e−3

slide-11
SLIDE 11

20 40 60 80 100 120 140 160 180 200 −1.5 −1 −0.5 0.5 1 1.5 x 10

13

Solution Ax=b: noise level 10−3

slide-12
SLIDE 12

20 40 60 80 100 120 140 160 180 200 −100 −50 50 100 150 solution of Ax=b : no added noise

slide-13
SLIDE 13

20 40 60 80 100 120 140 160 180 200 0.02 0.04 0.06 0.08 0.1 0.12 0.14 exact solution

slide-14
SLIDE 14

The singular value decomposition

slide-15
SLIDE 15

The SVD of the m × n matrix A, m ≥ n: A = UΣV T U = [u1, u2, . . . , um]

  • rthogonal,

m × m, V = [v1, v2, . . . , vn]

  • rthogonal,

n × n, Σ = diag[σ1, σ2, . . . , σn], m × n with σ1 ≥ σ2 ≥ . . . ≥ σn ≥ 0.

slide-16
SLIDE 16

Application: Least-squares approximation Let the matrix A ∈ Rm×n, m ≥ n, represent the “model” and the vector b ∈ Rm the data. Solve min

x Ax−b2 = min x UΣV Tx−b2 = min x ΣV Tx−U Tb2.

Let y = [y1, y2, . . . , yn]T = V Tx and b′ = [b′

1, b′ 2, . . . , b′ m]T = U Tb. Then

min

x Ax−b2 = min y

Σy−b′2 =

n

  • j=1

(σjyj−b′

j)2+ m

  • j=n+1

(b′

j)2.

slide-17
SLIDE 17

If A is of full rank, then all σj > 0 and yj = b′

j

σj , 1 ≤ j ≤ n, yields the solution x = V y. If some σj = 0, then least-squares solution not unique. Often one is interested in the least-squares solution of minimal norm. Arbitrary components yj are set to zero.

slide-18
SLIDE 18

Assume that σ1 ≥ σ2 . . . ≥ σℓ > σℓ+1 = . . . = σn = 0. Then A is of rank ℓ. Introduce the diagonal matrix Σ† = diag[1/σ1, 1/σ2, . . . , 1/σℓ, 0, . . . , 0], n × m. The matrix A† = V Σ†U T is known as the Moore-Penrose pseudoinverse of A.

slide-19
SLIDE 19

The solution of the least-squares problem min

x Ax − b

  • f minimal Eucliden norm can be expressed as

x = A†b. Moreover, A†A = I, AA† = PR(A).

slide-20
SLIDE 20

Note A = UΣV T =

n

  • j=1

σjujvT

j .

Define Ak :=

k

  • j=1

σjujvT

j ,

1 ≤ k ≤ ℓ. Then Ak is of rank k; Ak is the sum of k rank-one matrices σjujvT

j .

Moreover, A − Ak = min

rank(B)≤k A − B = σk+1,

i.e., Ak is the closest matrix of rank ≤ k to A.

slide-21
SLIDE 21

Let b = ˆ b + e, where e denotes an error. Then x := A†b =

  • j=1

uT

j b

σj vj =

  • j=1

uT

j ˆ

b σj vj +

  • j=1

uT

j e

σj vj = ˆ x +

  • j=1

uT

j e

σj vj. If σℓ > 0 tiny, then uT

ℓ e

σℓ might be huge and x a meaningless approximation of ˆ x.

slide-22
SLIDE 22

Recall Ak =

k

  • j=1

σjujvT

j

best rank-k approximation of A. Pseudoinverse of Ak: A†

k := k

  • j=1

σ−1

j vjuT j ,

σk > 0

slide-23
SLIDE 23

Approximate ˆ x by xk := A†

kb

=

k

  • j=1

uT

j b

σj vj =

k

  • j=1

uT

j ˆ

b σj vj +

k

  • j=1

uT

j e

σj vj. for some k ≤ ℓ. How to choose k?

slide-24
SLIDE 24

Example 1 cont’d: Singular values of A

20 40 60 80 100 120 140 160 180 200 10

−20

10

−15

10

−10

10

−5

10 10

5

slide-25
SLIDE 25

Example 1 cont’d: Right-hand side without noise

5 10 15 20 25 30 35 40 45 50 0.8 1 1.2 1.4 1.6 1.8 2 2.2 2.4 2.6 2.8 Norm of computed approximate solution k ||xk||

slide-26
SLIDE 26

Example 1 cont’d: Right-hand side without noise

5 10 15 20 25 30 35 40 45 50 −14 −12 −10 −8 −6 −4 −2 k log10||b−Axk|| Norm of residual vector

slide-27
SLIDE 27

Example 1 cont’d: Right-hand side without noise

−0.1 0.1 0.2 0.3 0.4 0.5 0.6 −14 −12 −10 −8 −6 −4 −2 L−curve log10||b−Axk|| log10||xk||

slide-28
SLIDE 28

Example 1 cont’d: Right-hand side without noise: Blow-up

0.06 0.07 0.08 0.09 0.1 0.11 0.12 0.13 0.14 −14 −13.8 −13.6 −13.4 −13.2 −13 −12.8 −12.6 −12.4 L−curve log10||b−Axk|| log10||xk||

slide-29
SLIDE 29

Example 1 cont’d: Exact and computed solutions

20 40 60 80 100 120 140 160 180 200 0.2 0.4 0.6 0.8 1 1.2 1.4 exact x9 x10 x11

slide-30
SLIDE 30

Example 1 cont’d: Right-hand side with relative noise 10−3

2 4 6 8 10 12 14 16 18 20 0.8 0.9 1 1.1 1.2 1.3 1.4 1.5 ||xk|| k Norm of computed approximate solution

slide-31
SLIDE 31

Example 1 cont’d: Right-hand side with relative noise 10−3

2 4 6 8 10 12 14 16 18 20 10

−4

10

−3

10

−2

10

−1

10 log10||b−Axk|| k Norm of residual

slide-32
SLIDE 32

Example 1 cont’d: Right-hand side with relative noise 10−3

−0.06 −0.04 −0.02 0.02 0.04 0.06 0.08 0.1 0.12 0.14 −3.5 −3 −2.5 −2 −1.5 −1 −0.5 log10||xk|| log10||b−Axk|| L−curve

slide-33
SLIDE 33

Example 1 cont’d: Right-hand side with relative noise 10−3

20 40 60 80 100 120 140 160 180 200 0.2 0.4 0.6 0.8 1 1.2 1.4 exact x3 x4 x5

slide-34
SLIDE 34

The SVD in Hilbert space

slide-35
SLIDE 35

A : X → Y compact linear operator X, Y Hilbert spaces, ·, · inner products in X and Y, · associated norms. Compute minimal-norm solution ˆ x ∈ X of Ax = ˆ y, ˆ y ∈ R(A), by the SVD of A.

slide-36
SLIDE 36

Singular triplets {σj, uj, vj}∞

j=1 of A:

σj singular value, σ1 ≥ σ2 ≥ σ3 ≥ . . . > 0, uj ∈ Y left singular function, vj ∈ X right singular function, satisfy uj, uℓ = vj, vℓ =

    

1, j = ℓ, 0, j = ℓ, and

slide-37
SLIDE 37

Avj = σjuj, A∗uj = σjvj, j = 1, 2, 3, . . . , Ax =

  • j=1

σjx, vjuj, ∀x ∈ X, A∗y =

  • j=1

σjy, ujvj, ∀y ∈ Y, where A∗ : Y → X the adjoint of A. A compact ⇒ σj cluster at zero.

slide-38
SLIDE 38

Minimal-norm solution ˆ x =

  • j=1

ˆ y, uj σj vj. ˆ x ∈ X implies Picard condition:

  • j=1

|ˆ y, uj|2 σ2

j

< ∞. Fourier coefficients ˆ cj = ˆ y, uj, j = 1, 2, 3, . . . , have to converge to zero rapidly.

slide-39
SLIDE 39

y = ˆ y + e available contaminated rhs Determine approximation of ˆ x by TSVD: xk =

k

  • j=1

y, uj σj vj. ˆ k ≥ 1 smallest index, such that xˆ

k − ˆ

x = min

k≥1 xk − ˆ

x.

slide-40
SLIDE 40

Assume norm of error in rhs known: δ = y − ˆ y. z is said to satisfy the discrepancy principle if Az − y ≤ τδ, where τ > 1 constant independent of δ. The discrepancy principle selects the smallest index k = kδ, such that Axkδ − y ≤ τδ.

slide-41
SLIDE 41

Then

  • Axk − y decreases monotonically as k increases.
  • kδ increases monotonically as δ ց 0.
  • xkδ → ˆ

x as δ ց 0.

slide-42
SLIDE 42

Tikhonov regularization

slide-43
SLIDE 43

Solve the minimization problem min

x {Ax − b2 + µx2},

(3) where µ > 0 is the (fixed) regularization parameter and L ∈ Rp×n, p ≤ n, the regularization operator. Normal equations associated with (3): (ATA + µI)x = ATb.

slide-44
SLIDE 44

Solution of the Tikhonov minimization problem xµ := (ATA + µI)−1ATb, µ > 0. Note that: lim

µց0 xµ = A†b,

lim

µ→∞ xµ = 0.

A proper choice of the value of the regularization parameter µ is important.

slide-45
SLIDE 45

If ǫ := e is known, then we can use the discrepancy principle. For now, assume that ǫ is not known. To see how the value of µ affects the solution xµ and the residual error b − Axµ plot the curve (log10 xµ, log10 b − Axµ), µ > 0 known as the L-curve.

slide-46
SLIDE 46

Choose the value of µ, denoted µL , that corresponds to the vertex of the L-curve. Reasons for choosing µ = µL:

  • When µ is too small, the solution is contaminated by

errors, hence of large norm.

  • When µ is too large, xµ solves a faraway problems,

hence the norm of the residual b − Axµ is large.

  • The solution corresponding to µL is balances these

errors.

slide-47
SLIDE 47

Example 1 cont’d: Right-hand side with relative noise 10−3

0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 −2.6 −2.5 −2.4 −2.3 −2.2 −2.1 −2 −1.9 −1.8 −1.7 −1.6 L−curve log10||xµ|| log10||b−Axµ||

slide-48
SLIDE 48

Example 1 cont’d: Right-hand side with relative noise 10−3

−9 −8 −7 −6 −5 −4 −3 −2 −0.8 −0.6 −0.4 −0.2 0.2 0.4 0.6 log10µ log10||xµ−xexact|| Error in Tikhonov solutions

slide-49
SLIDE 49
  • For small to medium-sized problems, we compute the

SVD of A. It is then inexpensive to determine points (log10 xµ, log10 b − Axµ)

  • n the L-curve for several values of µ.
  • For large problems it is expensive to determine

points on the L-curve.

slide-50
SLIDE 50

Extrapolation enhanced SVD

slide-51
SLIDE 51

Let A = A1 ⊗ A2 with A1, A2 ∈ R1500×1500 defined by the MATLAB functions baart and foxgood. Then A is 2.25 · 106 × 2.25 · 106. The SVD of A can be computed from the SVDs of A1 and A2. Extrapolation is equivalent to post-processing of the singular values.

slide-52
SLIDE 52

Error in TSVD solutions (dashed graph) and in extrapolated TSVD solutions (solid graph). The errors are measured in the Frobenius norm.

5 10 15 20 25 30 −2 −1 1 2 3 4 truncation index relative error norm Error RRETSVD Error TSVD

slide-53
SLIDE 53

Computed extrapolated TSVD solution x23.

slide-54
SLIDE 54

Computed TSVD solution x23.

slide-55
SLIDE 55

Danke!