Solving ill-posed nonlinear systems with noisy data: a regularizing - - PowerPoint PPT Presentation

solving ill posed nonlinear systems with noisy data a
SMART_READER_LITE
LIVE PREVIEW

Solving ill-posed nonlinear systems with noisy data: a regularizing - - PowerPoint PPT Presentation

Solving ill-posed nonlinear systems with noisy data: a regularizing trust-region approach Elisa Riccietti Universit` a degli Studi di Firenze Dipartimento di Matematica e Informatica Ulisse Dini Joint work with Stefania Bellavia,


slide-1
SLIDE 1

Solving ill-posed nonlinear systems with noisy data: a regularizing trust-region approach

Elisa Riccietti Universit` a degli Studi di Firenze Dipartimento di Matematica e Informatica ’Ulisse Dini’ Joint work with Stefania Bellavia, Benedetta Morini Opening Meeting for the Research Project GNCS 2016 PING - Inverse Problems in Geophysics Florence, April 6, 2016.

slide-2
SLIDE 2

Discrete nonlinear ill-posed problems and regularizing methods

Ill-posed problems

Let us consider the following inverse problem: given F : Rn → Rm with m ≥ n, nonlinear, continuously differentiable and y ∈ Rm, find x ∈ Rn such that F(x) = y. Definition The problem is well-posed if: 1 ∀y ∈ Rm ∃x ∈ Rn such that F(x) = y (existence), 2 F is an injective function (uniqueness), 3 F −1 is a continuous function (stability). The problem is ill-posed if one or more of the previous properties do not hold.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 2 / 47

slide-3
SLIDE 3

Discrete nonlinear ill-posed problems and regularizing methods

Ill-posed problems

Let us consider problems of the form F(x) = y for x ∈ (Rn, · 2) and y ∈ (Rm, · 2), arising from the discretization of a system modeling an ill-posed problem, such that: it exists a solution x†, but is not unique, stability does not hold. In a realistic situation the data y are affected by noise, we have at disposal only y δ such that: y − y δ ≤ δ for some positive δ . We can handle only a noisy problem: F(x) = y δ.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 3 / 47

slide-4
SLIDE 4

Discrete nonlinear ill-posed problems and regularizing methods

Need for regularization

As stability does not hold, the solutions of the original problem do not depend continuously on the data. = ⇒ The solutions of the noisy problem may not be meaningful approximations of the original problem solutions.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 4 / 47

slide-5
SLIDE 5

Discrete nonlinear ill-posed problems and regularizing methods

Need for regularization

As stability does not hold, the solutions of the original problem do not depend continuously on the data. = ⇒ The solutions of the noisy problem may not be meaningful approximations of the original problem solutions. For ill-posed problems there are no finite bounds on the inverse of the Jacobian of F around a solution of the original problem.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 4 / 47

slide-6
SLIDE 6

Discrete nonlinear ill-posed problems and regularizing methods

Need for regularization

As stability does not hold, the solutions of the original problem do not depend continuously on the data. = ⇒ The solutions of the noisy problem may not be meaningful approximations of the original problem solutions. For ill-posed problems there are no finite bounds on the inverse of the Jacobian of F around a solution of the original problem. Classical methods used for well-posed systems are not suitable in this contest. ⇓

Need for regularization.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 4 / 47

slide-7
SLIDE 7

Discrete nonlinear ill-posed problems and regularizing methods

Outline

Introduction to iterative regularization methods. Description of Levenberg-Marquardt method and of its regularizing variant. Description of a new regularizing trust-region approach, obtained by a suitable choice of the trust region radius . Regularization and convergence properties of the new approach. Numerical tests: we compare the new trust-region approach to the regularizing Levenberg-Marquardt and standard trust-region methods. Open issues and future developments.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 5 / 47

slide-8
SLIDE 8

Discrete nonlinear ill-posed problems and regularizing methods

Iterative regularization methods

Hypothesis: it exists x† solution of F(x) = y. Iterative regularization methods generate a sequence {xδ

k}. If the process

is stopped at iteration k∗(δ) the method is supposed to guarantee the following properties: xδ

k∗(δ) is an approximation of x†;

{xδ

k∗(δ)} tends to x† if δ tends to zero;

local convergence to x† in the noise-free case.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 6 / 47

slide-9
SLIDE 9

Discrete nonlinear ill-posed problems and regularizing methods

Existing methods

Landweber (gradient-type method)[ Hanke, Neubauer, Scherzer, 1995,Kaltenbacher, Neubauer, Scherzer, 2008 ] Truncated Newton - Conjugate Gradients [Hanke,1997, Rieder, 2005] Iterative Regularizing Gauss-Newton [Bakushinsky, 1992, Blaschke, Neubauer, Scherzer, 1997] Levenberg-Marquardt [Hanke,1997,2010,Vogel 1990, Kaltenbacher, Neubauer, Scherzer, 2008] These methods are analyzed only under local assumptions, the definition

  • f globally convergent approaches is still an open task.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 7 / 47

slide-10
SLIDE 10

Levenberg-Marquardt methods for ill-posed problems

Levenberg-Marquardt method

Given xδ

k ∈ Rn and λk > 0, we denote with J ∈ Rm×n the Jacobian

matrix of F. The step pk ∈ Rn is the minimizer of mLM

k

(p) = 1 2F(xδ

k) − y δ + J(xδ k)p2 + 1

2λkp2; pk is the solution of (Bk + λkI)pk = −gk with Bk = J(xδ

k)T J(xδ k), gk = J(xδ k)T (F(xδ k) − y δ);

The step is then used to compute the new iterate xδ

k+1 = xδ k + pk.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 8 / 47

slide-11
SLIDE 11

Levenberg-Marquardt methods for ill-posed problems

Regularizing Levenberg-Marquardt method

The parameter λk > 0 is chosen as the solution of: F(xδ

k) − y δ + J(xδ k)p = qF(xδ k) − y δ

with q ∈ (0, 1); With noisy data the process is stopped at iteration k∗(δ) such that xδ

k∗(δ) satisfies the discrepancy principle:

F(xδ

k∗(δ)) − y δ ≤ τδ < F(xδ k) − y δ

for 0 ≤ k < k∗(δ) and τ > 1 suitable parameter.

[Hanke, 1997,2010]

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 9 / 47

slide-12
SLIDE 12

Levenberg-Marquardt methods for ill-posed problems

Local analysis

Hypothesis for the local analysis: Given the starting guess x0, it exist positive ρ and c such that the system F(x) = y is solvable in Bρ(x0); for x, ˜ x ∈ B2ρ(x0) F(x) − F(˜ x) − J(x)(x − ˜ x) ≤ cx − ˜ xF(x) − F(˜ x). [Hanke, 1997,2010] Due to the ill-posedness of the problem it is not possible to assume that a finite bound on the inverse of the Jacobian matrix exists.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 10 / 47

slide-13
SLIDE 13

Levenberg-Marquardt methods for ill-posed problems

Regularizing properties of the Levenberg-Marquardt method

Choosing λk as the solution of F(xδ

k) − y δ + J(xδ k)p = qF(xδ k) − y δ

and stopping the process when the discrepancy principle F(xδ

k∗(δ)) − y δ ≤ τδ < F(xδ k) − y δ

is satisfied, Hanke proves that: With exact data (δ = 0): local convergence to x† , With noisy data (δ > 0): if τ > 1

q, choosing x0 close to x† the

discrepancy principle is satisfied after a finite number of iterations k∗(δ) and {xδ

k∗(δ)} converges to a solution of F(x) = y if δ tends to

zero.

This is a regularizing method

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 11 / 47

slide-14
SLIDE 14

Regularizing properties of trust-region methods

Trust-region methods

Given xδ

k ∈ Rn, the step pk ∈ Rn is the minimizer of

min

p mTR k (p) = 1

2F(xδ

k) − y δ + J(xδ k)p2,

s.t. p ≤ ∆k, with ∆k > 0 trust-region radius. Set Φ(x) = 1

2F(x) − y δ2, and compute

πk(pk) = Φ(xk) − Φ(xk + pk) mTR

k (0) − mTR k (pk) .

Given η ∈ (0, 1): If πk < η then set ∆k+1 < ∆k and xk+1 = xk. If πk ≥ η then set ∆k+1 ≥ ∆k and xk+1 = xk + pk.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 12 / 47

slide-15
SLIDE 15

Regularizing properties of trust-region methods

Trust-region methods

It is possible to prove that pk solves (Bk + λkI)pk = −gk for some λk ≥ 0 such that λk(pk − ∆k) = 0, where we have set Bk = J(xδ

k)TJ(xδ k) and gk = J(xδ k)T (F(xδ k) − y δ).

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 13 / 47

slide-16
SLIDE 16

Regularizing properties of trust-region methods

Trust-region methods

From λk(pk − ∆k) = 0 it follows that: If the minimum norm solution p∗ of Bkp = −gk satisfies p∗ ≤ ∆k then λk = 0 and pk = p(0);

  • therwise λk = 0, pk = ∆k and pk = p(λk) is a

Levenberg-Marquardt step. ⇓ The standard trust-region does not ensure regularizing properties. Trust-region should be active to have a regularizing method: pk = ∆k.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 14 / 47

slide-17
SLIDE 17

Regularizing properties of trust-region methods

Regularizing trust-region

Levenberg-Marquardt and trust-region methods are strictly connected, due to the form of the step. As Hanke did, can we introduce a trust-region method with regularizing properties and still globally convergent?

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 15 / 47

slide-18
SLIDE 18

Regularizing properties of trust-region methods

Goals

We modify the standard trust-region to have: monotone decay of the function Φ(x) = 1 2F(x) − y δ2, the q-condition to hold: F(xδ

k) − y δ + J(xδ k)p ≥ qF(xδ k) − y δ.

The q-condition is a relaxed reformulation of F(xδ

k) − y δ + J(xδ k)p = qF(xδ k) − y δ.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 16 / 47

slide-19
SLIDE 19

Regularizing properties of trust-region methods

Regularizing trust-region

We now describe the new trust-region approach that thanks to a suitable trust-region radius update ensures: the q-condition to hold, the same regularizing properties of Levenberg-Marquardt method.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 17 / 47

slide-20
SLIDE 20

Regularizing properties of trust-region methods

Trust-region radius choice

Lemma Let pk the solution of trust-region problem. If ∆k ≤ 1 − q Bk gk then pk satisfies the q-condition. Consequence: ∆k’s choice ∆k ∈

  • Cmingk, min
  • Cmax, 1 − q

Bk gk

  • ,

with Cmin, Cmax suitable constant, Bk = J(xδ

k)T J(xδ k) e

gk = J(xδ

k)T (F(xδ k) − y δ).

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 18 / 47

slide-21
SLIDE 21

Regularizing properties of trust-region methods

Algorithm : k-th iteration of regularizing trust-region Given xδ

k , η ∈ (0, 1), γ ∈ (0, 1), 0 < Cmin < Cmax.

Exact data: y, q ∈ (0, 1). Noisy data: y δ, q ∈ (0, 1), τ > 1/q.

  • 1. Compute Bk = J(xδ

k )TJ(xδ k ) and gk = J(xδ k )T(F(xδ k ) − y δ).

  • 2. Choose ∆k ∈
  • Cmingk, min
  • Cmax, 1 − q

Bk

  • gk
  • 3. Repeat

3.1 Compute the solution pk of trust-region problem. 3.2 Compute πk(pk) = Φ(xδ

k ) − Φ(xδ k + pk)

mTR

k (0) − mTR k (pk)

with Φ(x) = 1

2F(x) − y δ2, mTR k (p) = 1 2F(xδ k ) + J(xδ k )p2.

3.3 If πk(pk) < η,set ∆k = γ∆k. Until πk(pk) ≥ η.

  • 4. Set xδ

k+1 = xδ k + pk. Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 19 / 47

slide-22
SLIDE 22

Regularizing properties of trust-region methods

Local analysis

Hypothesis 1: the same as for Levenberg-Marquardt method. We assume that for index ¯ k it exist positive ρ and c such that 1 the system F(x) = y is solvable in Bρ(xδ

¯ k);

2 for x, ˜ x ∈ B2ρ(xδ

¯ k)

F(x) − F(˜ x) − J(x)(x − ˜ x) ≤ cx − ˜ xF(x) − F(˜ x). Hypothesis 2: It exists positive KJ such that J(x) ≤ KJ for all x ∈ L = {x ∈ Rn s.t. Φ(x) ≤ Φ(x0)}.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 20 / 47

slide-23
SLIDE 23

Regularizing properties of trust-region methods

Results for δ = 0

Lemma The method generates a sequence {xk} such that for k ≥ ¯ k trust-region is active, i.e. λk > 0; xk belongs to B2ρ(x¯

k) and to Bρ(x†);

xk+1 − x† < xk − x†; it exists ¯ λ > 0 such that λk ≤ ¯ λ. Theorem The sequence {xk} converges to a solution x∗ of F(x) = y such that x∗ − x† ≤ ρ . It holds limk→∞ gk = 0 so the trust-region radius tends to zero.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 21 / 47

slide-24
SLIDE 24

Regularizing properties of trust-region methods

Results for δ > 0

Lemma Let ¯ k < k∗(δ). The method generates a sequence {xδ

k} such that for

¯ k ≤ k < k∗(δ) the trust-region is active, i.e. λk > 0; xδ

k belongs to B2ρ(xδ ¯ k) and to Bρ(x†);

k+1 − x† < xδ k − x†;

it exists ¯ λ > 0 such that λk ≤ ¯ λ. Theorem The discrepancy principle is satisfied after a finite number of iterations k∗(δ) and the sequence {xδ

k∗(δ)} converges to a solution of F(x) = y if δ

tends to zero.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 22 / 47

slide-25
SLIDE 25

Regularizing properties of trust-region methods

This is a regularizing method.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 23 / 47

slide-26
SLIDE 26

Numerical tests

Test problems

Four nonlinear ill-posed systems arising from the discretization of nonlinear first-kind Fredholm integral equation are considered, they model gravimetric and geophysics problems: 1 k(t, s, x(s))ds = y(t), t ∈ [0, 1], P1,P2, [Vogel, 1990], P3,P4 [Kaltenbacher,2007]; Their kernel is of the form k(t, s, x(s)) = log

  • (t − s)2 + H2

(t − s)2 + (H − x(s))2

  • ;

k(t, s, x(s)) = 1

  • 1 + (t − s)2 + x(s)2 ;

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 24 / 47

slide-27
SLIDE 27

Numerical tests

Test problems: discretization

We chose n = m, interval [0, 1] was discretized using n=64 equidistant grid points ti = (i − 1)h, h = 1/(n − 1), i = 1, . . . , n; x(s) was approximated by piecewise linear functions Φj(s) on the grid sj = tj, j = 1, . . . , n; x(s) ∼ ˆ xn(s) = n

j=1 Φj(s)xj

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 25 / 47

slide-28
SLIDE 28

Numerical tests

Test problems: discretization

The integrals 1

0 k(ti, s, ˆ

x(s))ds, i = 1, . . . , n were approximated by the composite trapezoidal rule on the points sj j = 1, . . . , n. The resulting nonlinear system is

n

  • i=1

wjk(ti, sj, ˆ x(sj)) = y(ti) j = 1, . . . , n. with w1 = wn = 1

2, wi = 1 for all i = 1, n.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 26 / 47

slide-29
SLIDE 29

Numerical tests

Choice of parameters λk

Parameters λk were computed to have an active trust-region: p(λ) = ∆k. We used Newton method to solve this reformulation of the condition: ψ(λ) = 1 p(λ) − 1 ∆k = 0. that is more suitable to the application of Newton method. Each Newton iteration requires Cholesky factorization of Bk + λkI.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 27 / 47

slide-30
SLIDE 30

Numerical tests

Regularizing trust-region implementation

Trust-region radius update: ∆k = µkF(xδ

k) − y δ,

µk =        1 6µk−1 if qk−1 < q 2µk−1 if qk−1 > νq µk−1

  • therwise

with qk = F(xδ

k )−yδ+J(xδ k )pk

||F(xδ

k )−yδ||

, and ν = 1.1. ∆k is less expensive to compute if compared to 1−q

Bkgk but

preserves convergence to zero if δ = 0. In the update the fulfillment of q-condition is considered.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 28 / 47

slide-31
SLIDE 31

Numerical tests

Regularizing properties

10 20 30 40 50 60 0.4 0.5 0.6 0.7 0.8 0.9 1 k qk

· = Values of qk = ||F(xδ

k )−yδ+J(xδ k )pk||

||F(xδ

k )−yδ||

, solid line: q = 1.1/τ. The q-condition is satisfied in most of the iterations even if not esplicitly imposed.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 29 / 47

slide-32
SLIDE 32

Numerical tests

Regularizing properties of the method.

10

−8

10

−7

10

−6

10

−5

10

−4

10

−3

10

−2

10

−4

10

−3

10

−2

10

−1

δ ||xk

*

δ −x+||

Logarithmic plot of the error ||xδ

k∗(δ) − x†|| as a function of the noise level

δ.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 30 / 47

slide-33
SLIDE 33

Numerical tests

Comparison between regularizing TR-LM, δ = 10−2.

Problem Regularizing TR Regularizing LM x0 it nf cf it nf cf P1 0 e 20 21 6 17 18 4 −0.5 e 29 30 6 22 23 4 −1 e 35 36 5 24 25 4 −2 e 40 41 5 25 26 4 P2 0 e 30 31 5 * * * 0.5 e 25 26 5 * * * 1 e 29 30 5 22 23 5 2 e 37 39 5 25 26 5 P3 x0(1.25) 15 16 4 12 13 4 x0(1.5) 17 18 4 14 15 4 x0(1.75) 19 20 4 15 16 4 x0(2) 22 23 4 16 17 4 P4 x0(1, 1) 17 18 5 10 11 4 x0(0.5, 0) 20 21 4 * * * x0(1.5, 1) 22 23 4 15 16 4 x0(1.5, 0) 26 27 4 * * * it=iterations, nf=function evaluations, cf=mean number

  • f Cholesky

factorizations. ∗=failure, reached maximum number

  • f iterations or

convergence to a solution of the noisy problem e = (1, . . . , 1)T , P3: (x0(α))j = (−4α + 4)s2

j + (4α − 4)sj + 1, P4: x0(β, χ) = β − χsj, sj grid

points, j = 1, . . . , n.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 31 / 47

slide-34
SLIDE 34

Numerical tests

Comparison between regularizing TR and LM

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

  • 0.15
  • 0.1
  • 0.05

0.05 0.1

regularizing trust-region

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

  • 0.15
  • 0.1
  • 0.05

0.05 0.1

regularizing Levenberg-Marquardt

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

  • 0.5

0.5 1 1.5 2 2.5 3 3.5 4

regularizing trust-region

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

  • 0.5

0.5 1 1.5 2 2.5 3 3.5 4

regularizing Levenberg-Marquardt

Left: regularizing TR, Right: regularizing LM , Solid line: solution of the original problem.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 32 / 47

slide-35
SLIDE 35

Numerical tests

Comparison between regularizing TR e LM

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

  • 0.35
  • 0.3
  • 0.25
  • 0.2
  • 0.15
  • 0.1
  • 0.05

0.05

regularizing trust-region

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

  • 3000
  • 2000
  • 1000

1000 2000 3000

regularizing Levenberg-Marquardt

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

  • 2
  • 1

1 2 3 4

regularizing trust-region

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

  • 2000
  • 1500
  • 1000
  • 500

500 1000 1500 2000

regularizing Levenberg-Marquardt

Left: regularizing TR , Right: regularizing LM , Solid line: solution of the original problem.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 33 / 47

slide-36
SLIDE 36

Numerical tests

The q-condition

The condition imposed by Hanke is strongly dependent on the choice of the value of free parameter q. Values of q = 0.67, 0.70, 0.73, 0.87.

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 ×104

  • 2
  • 1.5
  • 1
  • 0.5

0.5 1 1.5 2 plot of the true and the computed solution, q=0.67

LM q=0.67

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

  • 0.5

0.5 1 1.5 plot of the true and the computed solution, q=0.70

LM, q=0.70

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

  • 2000
  • 1500
  • 1000
  • 500

500 1000 1500 2000 plot of the true and the computed solution, q=0.73

LM, q=0.73

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

  • 0.5

0.5 1 1.5 plot of the true and the computed solution, q=0.87

LM, q=0.87

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 34 / 47

slide-37
SLIDE 37

Numerical tests

Comparison between regularizing and standard trust-region

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

  • 0.25
  • 0.2
  • 0.15
  • 0.1
  • 0.05

0.05 0.1 0.15 0.2

regularizing trust-region

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

  • 0.25
  • 0.2
  • 0.15
  • 0.1
  • 0.05

0.05 0.1 0.15 0.2

standard trust-region

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

  • 0.35
  • 0.3
  • 0.25
  • 0.2
  • 0.15
  • 0.1
  • 0.05

0.05

regularizing trust-region

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

  • 0.8
  • 0.6
  • 0.4
  • 0.2

0.2 0.4 0.6

standard trust-region

Left: regularizing TR, Right: standard TR , Solid line: solution of the original problem.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 35 / 47

slide-38
SLIDE 38

Future developments

Future developments: nonlinear least squares problems

Consider the following least squares problem: given F : Rn → Rm with m ≥ n, nonlinear, continuously differentiable and y ∈ Rm, solve min

x f (x) = 1

2F(x) − y2. Non-zero residual problem: let x∗ be a solution of the problem and assume that F(x∗) − y > 0. Newton: given the current iterate xk, at each iteration the step pk is computed as: H(xk)pk = −J(xk)T(F(xk) − y) where H(xk) = J(xk)T J(xk) + S(xk), S(xk) =

m

  • i=1

(Fi(x) − yi)∇2Fi(x).

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 36 / 47

slide-39
SLIDE 39

Future developments

Future developments: nonlinear least squares problems

Gauss Newton: given the current iterate xk, at each iteration the step pk is computed as: J(xk)T J(xk)pk = −J(xk)T(F(xk) − y). The Gauss Newton method converges if S(x∗) < λ∗ with λ∗ the smallest eigenvalue of J(x∗)T J(x∗). This hypothesis is rather restrictive when dealing with ill-posed problems (the Jacobian matrix should be invertible). We want to design a trust region approach to solve least squares problems, converging under less restrictive hypotheses on λ∗ compared to the Gauss- Newton method.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 37 / 47

slide-40
SLIDE 40

Future developments

Future developments: nonlinear least squares problems

We want our trust-region method to be able to deal also with ill-posed noisy least-squares problems: min

x f (x) = 1

2F(x) − y2, and only noisy data y δ are at disposal: y − y δ ≤ δ. Non-zero residual problem: if x∗ is a solution of the problem we assume that F(x∗) − y > 0.

THANK YOU FOR YOUR ATTENTION!

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 38 / 47

slide-41
SLIDE 41

Open issues and future developments

Open issues: Convergence to the infinite dimensional solution.

Let X, Y be Hilbert spaces, F∞ : X → Y, y∞ ∈ Y. The nonlinear system is the discretization of a infinite dimensional problem: find x∞ ∈ X such that F∞(x∞) = y∞. We are interested in the convergence of the discrete solution ˆ xn(s) = n

j=1 Φj(s)xj to a solution of the infinite dimensional

problem as n → ∞. Theorem The sequence { ˆ xn} has a weakly convergent subsequence { ˆ xk}. Theorem The sequence {F∞( ˆ xk) − y∞} converges to zero as k tends to infinite, i.e. the weak limit x∗ of sequence { ˆ xk} is a solution of the original problem, F∞(x∗) = y∞.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 39 / 47

slide-42
SLIDE 42

Open issues and future developments

Open issues: peaks

Problem: when solving the nonlinear system obtained computing the integral by the trapezoidal rule, the approximated solution shows peaks at the end points of the interval. Peaks are higher and higher as the starting guess moves away from the solution and the noise increases. When solving the nonlinear system obtained computing the integral by the rectangular rule, the approximated solution does not show peaks at the end points of the interval.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 40 / 47

slide-43
SLIDE 43

Open issues and future developments

Computed solution

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 trapezoidal rule 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0.2 0.25 0.3 0.35 0.4 0.45 0.5 0.55 rectangular rule

Computed solution, x0 = 1e, δ = 1.e − 2. Left: trapezoidal rule, Right: rectangular rule, Solid line: solution of the original problem.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 41 / 47

slide-44
SLIDE 44

Open issues and future developments

20 40 60 80 100 120 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 k plot of the error

eI eB

5 10 15 20 25 30 35 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 k plot of the error

eI eB

10 20 30 40 50 60 70 80 0.1 0.2 0.3 0.4 0.5 0.6 0.7 k plot of the error

eI eB

5 10 15 20 25 30 35 0.1 0.2 0.3 0.4 0.5 0.6 0.7 k plot of the error

eI eB

eI = error computed on the points inside the interval, eB=border error. Upper part: trapezoidal rule, left: δ = 0, right: δ = 1.e − 2. Lower part: rectangular rule, left: δ = 0, right: δ = 1.e − 2.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 42 / 47

slide-45
SLIDE 45

Open issues and future developments

Comparison of the nonlinear systems

Trapezoidal rule: the resulting nonlinear system is

1 2k(ti, s1, x1) + 1k(ti, s2, x2) + · · ·+ 1k(ti, sn−1, xn−1) + 1 2k(ti, sn, xn) = y(ti),

i = 1, . . . , n. Rectangular rule: the resulting nonlinear system is

1k(ti, s1, x1) + 1k(ti, s2, x2) + · · · + 1k(ti, sn−1, xn−1) + 1k(ti, sn, xn) = y(ti),

i = 1, . . . , n.

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 43 / 47

slide-46
SLIDE 46

Open issues and future developments

Linear system: trapezoidal rule

We solve (JTJ + λI)p(λ) = −JT(F − y δ). Let n = 5.

J =      

1 2 ∂1k(t1, s1, x1)

1∂2k(t1, s2, x2) . 1∂4k(t1, s4, x4)

1 2∂5k(t1, s5, x5)

. . . . . . . . . . . . . .

1 2 ∂1k(t5, s1, x1)

1∂2k(t5, s2, x2) . 1∂4k(t5, s4, x4)

1 2∂5k(t5, s5, x5)

     

We denote ki,j = k(ti, sj, xj) i, j = 1, . . . , n.

JT J =      

1 4

5

i=1 ∂1ki,1∂1ki,1 1 2

5

i=1 ∂1ki,1∂2ki,2

.

1 2

5

i=1 ∂1ki,1∂4ki,4 1 4

5

i=1 ∂1ki,1∂5ki,5 1 2

5

i=1 ∂2ki,2∂1ki,1

1 5

i=1 ∂2ki,2∂2ki,2

. 1 5

i=1 ∂2ki,2∂4ki,4 1 2

5

i=1 ∂2ki,2∂5ki,5

. . . . .

1 2

5

i=1 ∂4ki,5∂1ki,1

1 5

i=1 ∂4ki,5∂2k1,2

. 1 5

i=1 ∂4ki,5∂4ki,4 1 2

5

i=1 ∂4ki,5∂5ki,5 1 4

5

i=1 ∂4ki,5∂1ki,1 1 2

5

i=1 ∂5ki,5∂2k1,2

.

1 2

5

i=1 ∂5ki,5∂4ki,4 1 4

5

i=1 ∂5ki,5∂5ki,5

      .

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 44 / 47

slide-47
SLIDE 47

Open issues and future developments

Linear system: rectangular rule

We solve (JTJ + λI)p(λ) = −JT(F − y δ). Let n = 5.

J =      1∂1k(t1, s1, x1) 1∂2k(t1, s2, x2) . 1∂4k(t1, s4, x4) 1∂5k(t1, s5, x5) . . . . . . . . . . . . . . 1∂1k(t5, s1, x1) 1∂2k(t5, s2, x2) . 1∂4k(t5, s4, x4) 1∂5k(t5, s5, x5)     

We denote ki,j = k(ti, sj, xj) i, j = 1, . . . , n.

JT J =       1 5

i=1 ∂1ki,1∂1ki,1

1 5

i=1 ∂1ki,1∂2ki,2

. 1 5

i=1 ∂1ki,1∂4ki,4

1 5

i=1 ∂1ki,1∂5ki,5

1 5

i=1 ∂2ki,2∂1ki,1

1 5

i=1 ∂2ki,2∂2ki,2

. 1 5

i=1 ∂2ki,2∂4ki,4

1 5

i=1 ∂2ki,2∂5ki,5

. . . . . 1 5

i=1 ∂4ki,5∂1ki,1

1 5

i=1 ∂4ki,5∂2k1,2

. 1 5

i=1 ∂4ki,5∂4ki,4

1 5

i=1 ∂4ki,5∂5ki,5

1 5

i=1 ∂4ki,5∂1ki,1

1 5

i=1 ∂5ki,5∂2k1,2

. 1 5

i=1 ∂5ki,5∂4ki,4

1 5

i=1 ∂5ki,5∂5ki,5

      .

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 45 / 47

slide-48
SLIDE 48

Open issues and future developments

SVD decomposition: trapezoidal rule

Let consider matrix JTJ SVD decomposition. JTJ = UΣUT cond(JT J) = 106, λ = 15.7, cond(JT J + λI) = 1.2 100 σ = diag(Σ) =       3.8 100 8.5 10−2 2.3 10−3 7.1 10−5 1.6 10−6       , p =       −7.6 10−2 −1.7 10−1 −1.8 10−1 −1.7 10−1 −7.6 10−2       U =       −0.24 −0.44 0.58 0.56 0.32 −0.54 −0.56 0.04 −0.44 −0.46 −0.56 3.5 10−8 −0.56 −7.3 10−8 0.61 −0.54 0.56 0.04 0.44 −0.46 −0.24 0.44 0.58 −0.56 0.32      

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 46 / 47

slide-49
SLIDE 49

Open issues and future developments

SVD decomposition: rectangular rule

Let consider matrix JTJ SVD decomposition. JTJ = UΣUT cond(JT J) = 106, λ = 17.4, cond(JT J + λI) = 1.3 100 σ = diag(Σ) =       5.1 100 1.8 10−1 5.8 10−3 1.3 10−4 1.8 10−6       , p =       −1.8 10−1 −2.0 10−1 −2.1 10−1 −2.0 10−1 −1.8 10−1       U =       −0.41 −0.60 0.55 −0.38 −0.17 −0.46 −0.38 −0.19 0.60 0.5 −0.48 −4.1 10−8 −0.57 −1.4 10−6 −0.66 −0.46 0.38 −0.19 −0.60 0.50 −0.41 0.60 0.55 0.38 −0.17      

Elisa Riccietti () Adaptive Trust-Region Regularization. Florence, April 6 2016 47 / 47