Some challenges of four-dimensional data assimilation problems Juan - - PowerPoint PPT Presentation

some challenges of four dimensional data assimilation
SMART_READER_LITE
LIVE PREVIEW

Some challenges of four-dimensional data assimilation problems Juan - - PowerPoint PPT Presentation

Some challenges of four-dimensional data assimilation problems Juan Carlos De los Reyes Centro de Modelizacin Matemtica (MODEMAT) Escuela Politcnica Nacional de Ecuador New trends in PDE constrained optimization Linz, 2019 1 / 30


slide-1
SLIDE 1

Some challenges of four-dimensional data assimilation problems

Juan Carlos De los Reyes

Centro de Modelización Matemática (MODEMAT) Escuela Politécnica Nacional de Ecuador New trends in PDE constrained optimization Linz, 2019

1 / 30

slide-2
SLIDE 2

Contenidos

1

Motivation

2

Analysis of variational data assimilation

3

Optimal placement problem

4

Numerical results

5

Summary of topics

slide-3
SLIDE 3

Outline

1

Motivation

2

Analysis of variational data assimilation

3

Optimal placement problem

4

Numerical results

5

Summary of topics

slide-4
SLIDE 4

Motivation

Data assimilation in Numerical Weather Prediction

Data assimilation methods aim at finding a good initial condition of the athmospheric system in order to get better weather forecasts;

3 / 30

slide-5
SLIDE 5

Motivation

Data assimilation in Numerical Weather Prediction

Data assimilation methods aim at finding a good initial condition of the athmospheric system in order to get better weather forecasts; Information can be obtained mainly from ground stations, radionsonds or satellite images;

3 / 30

slide-6
SLIDE 6

Motivation

Data assimilation in Numerical Weather Prediction

Data assimilation methods aim at finding a good initial condition of the athmospheric system in order to get better weather forecasts; Information can be obtained mainly from ground stations, radionsonds or satellite images; Reconstruction results depend strongly on the number of observations, which can be very limited in some cases.

3 / 30

slide-7
SLIDE 7

Equations of the athmosphere

Basic model ∂u ∂t = −u∂u ∂x − v∂u ∂y − w∂u ∂z + uv tan(φ) a − uw a − 1 ρ ∂p ∂x − 2Ω(w cos(φ) − v sin(φ)) + Frx ∂v ∂t = −u∂v ∂x − v∂v ∂y − w∂v ∂z + u2 tan(φ) a − uw a − 1 ρ ∂p ∂y − 2Ωu sin(φ) + Fry ∂w ∂t = −u∂w ∂x − v∂w ∂y − w∂w ∂z + u2 + v2 a − 1 ρ ∂p ∂z + 2Ωu cos(φ) − g + Frz ∂T ∂t = −u∂T ∂x − v∂T ∂y + (γ − γd)w + 1 cp dH dt ∂ρ ∂t = −u∂ρ ∂x − v∂ρ ∂y − w∂ρ ∂z − ρ ∂u ∂x + ∂v ∂y + ∂w ∂z

  • ∂q

∂t = −u∂q ∂x − v∂q ∂y − w∂q ∂z + Qv + Boundary conditions + Initial conditions

4 / 30

slide-8
SLIDE 8

Equations of the athmosphere

Basic model ∂u ∂t = −u∂u ∂x − v∂u ∂y − w∂u ∂z + uv tan(φ) a − uw a − 1 ρ ∂p ∂x − 2Ω(w cos(φ) − v sin(φ)) + Frx ∂v ∂t = −u∂v ∂x − v∂v ∂y − w∂v ∂z + u2 tan(φ) a − uw a − 1 ρ ∂p ∂y − 2Ωu sin(φ) + Fry ∂w ∂t = −u∂w ∂x − v∂w ∂y − w∂w ∂z + u2 + v2 a − 1 ρ ∂p ∂z + 2Ωu cos(φ) − g + Frz ∂T ∂t = −u∂T ∂x − v∂T ∂y + (γ − γd)w + 1 cp dH dt ∂ρ ∂t = −u∂ρ ∂x − v∂ρ ∂y − w∂ρ ∂z − ρ ∂u ∂x + ∂v ∂y + ∂w ∂z

  • ∂q

∂t = −u∂q ∂x − v∂q ∂y − w∂q ∂z + Qv + Boundary conditions + Initial conditions

4 / 30

slide-9
SLIDE 9

Approaches to data assimilation

Approaches based on static covariance matrices

Optimal interpolation 3D-Var

Recent approaches based on dynamic covariance matrices

4D-Var Extended Kalman filters Ensemble methods

Eugenia Kalnay Atmospheric Modeling, Data Assimilation and Predictability Cambridge Univ. Press, 2002.

5 / 30

slide-10
SLIDE 10

Optimal interpolation

Kalman filter

ub : background information vector ua : “analysis” (estimation of the state) z : observation vector z = Hu + v, where H is an observation operator and v is the observation error all variables are assumed to be Gaussian

Kalman filter

ua = ub + K(z − Hub), where K = BHT(HBHT + R)−1 is the Kalman gain corresponding to the linear unbiased estimator of minimim variance.

6 / 30

slide-11
SLIDE 11

Intuition of OI

7 / 30

slide-12
SLIDE 12

3D-Var

MAP estimation

The data assimilation problem may be treated via a Bayesian approach to find the maximum-a-posteriori (MAP) estimator.

8 / 30

slide-13
SLIDE 13

3D-Var

MAP estimation

The data assimilation problem may be treated via a Bayesian approach to find the maximum-a-posteriori (MAP) estimator. Using the Gaussian probability density functions p(z|u) ∝ exp

  • −1

2(z − Hu)TR−1(z − Hu)

  • p(u) ∝ exp
  • −1

2(ub − u)TB−1(ub − u)

  • 8 / 30
slide-14
SLIDE 14

3D-Var

MAP estimation

The data assimilation problem may be treated via a Bayesian approach to find the maximum-a-posteriori (MAP) estimator. Using the Gaussian probability density functions p(z|u) ∝ exp

  • −1

2(z − Hu)TR−1(z − Hu)

  • p(u) ∝ exp
  • −1

2(ub − u)TB−1(ub − u)

  • Using Bayes formula, p(u|z) = p(z|u)p(u)

p(z)

, the MAP estimator corresponds to the solution of the 3D-Var problem min 1 2(z − Hu)TR−1(z − Hu) + 1 2(ub − u)TB−1(ub − u).

8 / 30

slide-15
SLIDE 15

Intuition of 4D-Var

9 / 30

slide-16
SLIDE 16

4D-Var

Le Dimet and Talagrand (1986)

Finite dimensional problem

min

u

J(y, u) = 1 2

l

  • i=0

[H(y(ti)) − zo(ti)]T R−1

i

[H(y(ti)) − zo(ti)] +1 2 [u − ub]T B−1 [u − ub] subject to: y(tl) = M(y(t0)) (Dynamical system) y(t0) = u (Initial condition).

10 / 30

slide-17
SLIDE 17

4D-Var

Le Dimet and Talagrand (1986)

Finite dimensional problem

min

u

J(y, u) = 1 2

l

  • i=0

[H(y(ti)) − zo(ti)]T R−1

i

[H(y(ti)) − zo(ti)] +1 2 [u − ub]T B−1 [u − ub] subject to: y(tl) = M(y(t0)) (Dynamical system) y(t0) = u (Initial condition).

Features

The dynamic problem incorporates all observations in a given time window; The nonlinear dynamics may be taken into account; The operational use is still a computational challenge.

10 / 30

slide-18
SLIDE 18

4D-Var

Le Dimet and Talagrand (1986)

Finite dimensional problem

min

u

J(y, u) = 1 2

l

  • i=0

[H(y(ti)) − zo(ti)]T R−1

i

[H(y(ti)) − zo(ti)] +1 2 [u − ub]T B−1 [u − ub] subject to: y(tl) = M(y(t0)) (Dynamical system) y(t0) = u (Initial condition). What about the infinite-dimensional problem?

10 / 30

slide-19
SLIDE 19

Outline

1

Motivation

2

Analysis of variational data assimilation

3

Optimal placement problem

4

Numerical results

5

Summary of topics

slide-20
SLIDE 20

Semilinear problem

min

u

J(y, u) = 1 2

  • k,i

[y(xk, ti) − z0(xk, ti)]2 + 1 2u − ub2

B−1 ∂y ∂t + Ay + g(y) =

in Q = Ω×]0, T[ subject to: y =

  • n Σ = Γ×]0, T[

y(x, 0) = u in Ω,

Difficulties

Higher regularity of the initial condition is required to get some sort of continuity of the state. Pointwise misfits in the cost leads to right hand sides in M(Q) for the adjoint equation. Ill-posedness!

11 / 30

slide-21
SLIDE 21

Semilinear problem

min

u

J(y, u) = 1 2

  • k,i

[y(xk, ti) − z0(xk, ti)]2 + 1 2u − ub2

B−1 ∂y ∂t + Ay + g(y) =

in Q = Ω×]0, T[ subject to: y =

  • n Σ = Γ×]0, T[

y(x, 0) = u in Ω,

Difficulties

Higher regularity of the initial condition is required to get some sort of continuity of the state. Pointwise misfits in the cost leads to right hand sides in M(Q) for the adjoint equation. Ill-posedness!

11 / 30

slide-22
SLIDE 22

Semilinear problem

min

u

J(y, u) = 1 2

T

  • k,i

wkσiρi(t)[y(xk, t) − zo(xk, t)]2 dt + 1 2u − ub2

B−1 + ϑ

2 ∇(u − ub)2

L2(Ω) ∂y ∂t + Ay + g(y) =

in Q = Ω×]0, T[ subject to: y =

  • n Σ = Γ×]0, T[

y(x, 0) = u in Ω,

Difficulties

Higher regularity of the initial condition is required to get some sort of continuity of the state. Pointwise misfits in the cost leads to right hand sides in M(Q) for the adjoint equation. Ill-posedness! w and σ are binary vectors, and ρi(t) support functions

11 / 30

slide-23
SLIDE 23

Well-posedness

Assumption on the nonlinearity

g = g(x, t, y) : Q × R → R satisfies the Carathéodory conditions and is uniformly bounded at the origin, i.e., |g(x, t, 0)| ≤ K, for some K > 0, g is monotone increasing with respect to y for almost every (x, t) ∈ Q, g is twice continuously differentiable with respect to y and |gy(x, t, y)| + |gyy(x, t, y)| ≤ K, for some K > 0, for almost all (x, t) ∈ Q and any y ∈ R.

Theorem

If u ∈ H1

0(Ω) and the nonlinear term verifies the assumption, then the semilinear

equation has a unique solution y ∈ H2,1(Q) ֒ → L2(0, T; C(Ω)). Moreover, yH2,1(Q) ≤ c(1 + uH1

0(Ω)), for some c > 0. 12 / 30

slide-24
SLIDE 24

Differentiability

Theorem

The control-to-state mapping S : H1

0(Ω) → H2,1(Q), u → S(u) = y, is Gâteaux

  • differentiable. Its derivative, in direction h ∈ H1

0(Ω), is given by η ∈ H2,1(Q)

solution of:   

∂η ∂t + Aη + g′(y)η =

in Q η =

  • n Σ

η(x, 0) = h in Ω,

Adjoint equation

−∂p ∂t + A∗p + g′(y)p = µ in Q p =

  • n Σ

p(x, T) = in Ω, with µ ∈ L2(0, T; M(Ω)) has a unique solution p ∈ L2(0, T; W1,r

0 (Ω)), with

r ∈ [1,

m m−1[. Casas-Clason-Kunisch 2013, Meyer-Susu 2017

13 / 30

slide-25
SLIDE 25

Optimality system

State equation (in strong form):

∂¯ y ∂t + A¯

y + g(¯ y) = in Q, ¯ y =

  • n Σ,

¯ y(x, 0) = ¯ u in Ω. Adjoint equation (in very weak form): −∂¯ p ∂t + A∗¯ p + g′(¯ y)¯ p =

  • k,i

wkσiρi(t) [¯ y(x, t) − zo(x, t)] ⊗ δ(x − xk) in Q, ¯ p =

  • n Σ,

¯ p(x, T) = in Ω. Gradient equation (in weak form): −ϑ∆(¯ u − ub) + B−1(¯ u − ub) + ¯ p(0) = in Ω, ¯ u =

  • n Γ.

14 / 30

slide-26
SLIDE 26

Extensions

Interplay between function spaces and type of observations;

15 / 30

slide-27
SLIDE 27

Extensions

Interplay between function spaces and type of observations; Solutions for the dynamics of the atmosphere are not necessarily continuous;

15 / 30

slide-28
SLIDE 28

Extensions

Interplay between function spaces and type of observations; Solutions for the dynamics of the atmosphere are not necessarily continuous; Variational DA models typically contain a classical Tikhonov regularization u − ub2

B−1, which is not appropriate to reconstruct sharp fronts

15 / 30

slide-29
SLIDE 29

Extensions

Interplay between function spaces and type of observations; Solutions for the dynamics of the atmosphere are not necessarily continuous; Variational DA models typically contain a classical Tikhonov regularization u − ub2

B−1, which is not appropriate to reconstruct sharp fronts

Recently Freitag and co. (2013, 2015) proposed an alternative total variation (TV) regularization to recover solutions with sharp fronts.

15 / 30

slide-30
SLIDE 30

Extensions

Interplay between function spaces and type of observations; Solutions for the dynamics of the atmosphere are not necessarily continuous; Variational DA models typically contain a classical Tikhonov regularization u − ub2

B−1, which is not appropriate to reconstruct sharp fronts

Recently Freitag and co. (2013, 2015) proposed an alternative total variation (TV) regularization to recover solutions with sharp fronts. A drawback of TV is, however, the staircase effect, which may lead to undesired artifacts in the reconstruction.

15 / 30

slide-31
SLIDE 31

Extensions

Interplay between function spaces and type of observations; Solutions for the dynamics of the atmosphere are not necessarily continuous; Variational DA models typically contain a classical Tikhonov regularization u − ub2

B−1, which is not appropriate to reconstruct sharp fronts

Recently Freitag and co. (2013, 2015) proposed an alternative total variation (TV) regularization to recover solutions with sharp fronts. A drawback of TV is, however, the staircase effect, which may lead to undesired artifacts in the reconstruction. We studied second order total generalized variation regularization De los

Reyes, Loayza (2019)

15 / 30

slide-32
SLIDE 32

Extensions

Interplay between function spaces and type of observations; Solutions for the dynamics of the atmosphere are not necessarily continuous; Variational DA models typically contain a classical Tikhonov regularization u − ub2

B−1, which is not appropriate to reconstruct sharp fronts

Recently Freitag and co. (2013, 2015) proposed an alternative total variation (TV) regularization to recover solutions with sharp fronts. A drawback of TV is, however, the staircase effect, which may lead to undesired artifacts in the reconstruction. We studied second order total generalized variation regularization De los

Reyes, Loayza (2019)

Is this of operational use?

15 / 30

slide-33
SLIDE 33

Outline

1

Motivation

2

Analysis of variational data assimilation

3

Optimal placement problem

4

Numerical results

5

Summary of topics

slide-34
SLIDE 34

Observation stations

16 / 30

slide-35
SLIDE 35

Observation stations

More observations are required to further improve forecasts

16 / 30

slide-36
SLIDE 36

Observation stations

Problem

One has to make the best possible decision about where to locate the next

  • bservation stations to obtain better reconstructions of the initial condition.

16 / 30

slide-37
SLIDE 37

State-of-the-art

Optimal filtering Bensoussan and collab. (1971, 1972), Burns and collab. (1994, 1995,

1998, 2009, 2011), Rautenberg and collab. (2015, 2016)

Observability approaches Privat, Trelat, Zuazua (2013, 2015, 2017) Supervised learning Haber, Horesh, Tenorio (2008, 2010), De los Reyes, Schönlieb

and collab. (2013, 2016, 2017), Kunisch and collab. (2013, 2018)

Bayesian optimal experimental design Alexanderian et al. (2014, 2015), Herzog et

  • al. (2015, 2017)

17 / 30

slide-38
SLIDE 38

State-of-the-art

Optimal filtering Bensoussan and collab. (1971, 1972), Burns and collab. (1994, 1995,

1998, 2009, 2011), Rautenberg and collab. (2015, 2016)

Observability approaches Privat, Trelat, Zuazua (2013, 2015, 2017) Supervised learning Haber, Horesh, Tenorio (2008, 2010), De los Reyes, Schönlieb

and collab. (2013, 2016, 2017), Kunisch and collab. (2013, 2018)

Bayesian optimal experimental design Alexanderian et al. (2014, 2015), Herzog et

  • al. (2015, 2017)

Features and goals

Bilevel learning approach: flexibility about quality measures and more natural framework for nonlinear lower level problems. Analysis of the resulting problem in function space: optimality conditions and multiplier regularity. Design of efficient solution algorithms.

17 / 30

slide-39
SLIDE 39

Bilevel learning problem

Mixed integer-infinite dimensional optimization min

w,σ∈{0,1}

  • Q

N

  • j=1

L(yj, y†

j ) dxdt + β

N

  • j=1

l(uj, u†

j ) dx + βw

  • k

wk + βσ

  • i

σi subject to (∀j = 1, . . . , N) :                              min

uj

1 2

T

  • k,i
  • wkσiρi(t)[yj(xk, t) − zoj(xk, t)]2

dt +1 2uj − ubjB−1 + ϑ

2 ∇(uj − ubj)2 L2(Ω)

subject to: ∂yj ∂t + Ayj + g(yj) = in Q yj =

  • n Σ

yj(0) = uj in Ω.

18 / 30

slide-40
SLIDE 40

Bilevel learning problem

Mixed integer-infinite dimensional optimization min

w,σ∈{0,1}

  • Q

N

  • j=1

L(yj, y†

j ) dxdt + β

N

  • j=1

l(uj, u†

j ) dx + βw

  • k

wk + βσ

  • i

σi βw > 0 corresponds to the sparsity penalty term for w and βσ > 0 the one for σ; The training set

  • u†

j , y† j

  • , ∀j = 1, . . . , N is built from improved reconstructions of the initial

condition and the observed state; We aim at learning the vector of placements w such that the quality measure of the training set is minimized in average

  • u†

j , y† j

  • .

18 / 30

slide-41
SLIDE 41

Dealing with sparsity

Relaxing the integer constraints, 0 ≤ w, σ ≤ 1, we simply get linearity with respect to w and σ: min

0≤w,σ≤1

  • Q

N

  • j=1

L(yj, y†

j ) dxdt + β

N

  • j=1

l(uj, u†

j ) dx + βw

  • k

wk + βσ

  • i

σi

19 / 30

slide-42
SLIDE 42

Dealing with sparsity

Relaxing the integer constraints, 0 ≤ w, σ ≤ 1, we simply get linearity with respect to w and σ: min

0≤w,σ≤1

  • Q

N

  • j=1

L(yj, y†

j ) dxdt + β

N

  • j=1

l(uj, u†

j ) dx + βw

  • k

wk + βσ

  • i

σi Although sparsity is obtained as a result of the ℓ1-norm penalty, the number

  • f values different from 0 or 1 makes the solution difficult to interprete

19 / 30

slide-43
SLIDE 43

Dealing with sparsity

Relaxing the integer constraints, 0 ≤ w, σ ≤ 1, we simply get linearity with respect to w and σ: min

0≤w,σ≤1

  • Q

N

  • j=1

L(yj, y†

j ) dxdt + β

N

  • j=1

l(uj, u†

j ) dx + βw

  • k

wk + βσ

  • i

σi Although sparsity is obtained as a result of the ℓ1-norm penalty, the number

  • f values different from 0 or 1 makes the solution difficult to interprete

We consider also an approximation of the lp norm, with p ∈ (0, 1). For instance φ(x) =     

x ǫ

if 0 ≤ x < ǫ

2

pǫ(x) if ǫ

2 < x ≤ 2ǫ

1 if 2ǫ ≤ x ≤ 1

Alexandarian et al. 2014

19 / 30

slide-44
SLIDE 44

Control of a singular system with measures

Replacing the lower level problems by their necessary optimality condition:

min

0≤w,σ≤1 J(y, p, u, w) =

  • Q

N

  • j=1

L(yj, y†

j ) dxdt + β

N

  • j=1

l(uj, u†

j ) dx + βw

  • k

wk + βσ

  • i

σi subject to: ∂yj ∂t + Ayj + g(yj) = yj|Γ = yj(0) = uj −∂pj ∂t + A∗pj + g′(yj)pj =

  • k,i

wkσiρi(t) [¯ y(x, t) − zo(x, t)] ⊗ δ(x − xk) pj|Γ = pj(T) = −ϑ∆(uj − ub) + B−1(uj − ub) = −pj(0) ¯ u|Γ = 0.

20 / 30

slide-45
SLIDE 45

Control of a singular system with measures

Replacing the lower level problems by their necessary optimality condition:

min

0≤w,σ≤1 J(y, p, u, w) =

  • Q

N

  • j=1

L(yj, y†

j ) dxdt + β

N

  • j=1

l(uj, u†

j ) dx + βw

  • k

wk + βσ

  • i

σi subject to: ∂yj ∂t + Ayj + g(yj) = yj|Γ = yj(0) = uj −∂pj ∂t + A∗pj + g′(yj)pj =

  • k,i

wkσiρi(t) [¯ y(x, t) − zo(x, t)] ⊗ δ(x − xk) pj|Γ = pj(T) = −ϑ∆(uj − ub) + B−1(uj − ub) = −pj(0) ¯ u|Γ = 0. Casas, Clason, Kunisch, Neitzel, Pieper, Vexler, ...

20 / 30

slide-46
SLIDE 46

Control of a singular system with measures

Replacing the lower level problems by their necessary optimality condition:

min

0≤w,σ≤1 J(y, p, u, w) =

  • Q

N

  • j=1

L(yj, y†

j ) dxdt + β

N

  • j=1

l(uj, u†

j ) dx + βw

  • k

wk + βσ

  • i

σi subject to: ∂yj ∂t + Ayj + g(yj) = yj|Γ = yj(0) = uj −∂pj ∂t + A∗pj + g′(yj)pj =

  • k,i

wkσiρi(t) [¯ y(x, t) − zo(x, t)] ⊗ δ(x − xk) pj|Γ = pj(T) = −ϑ∆(uj − ub) + B−1(uj − ub) = −pj(0) ¯ u|Γ = 0.

Difficulty: No unique solution of the optimality system!

20 / 30

slide-47
SLIDE 47

Adapted penalty approach

For (¯ y, ¯ p, ¯ u) ∈ H2,1(Q) × L2(W1,r

0 (Ω)) × H1 0(Ω) solution of the bilevel problem, find

(yγ, pγ, wγ, sγ, vγ) ∈ H2,1(Q) × L2(W1,r

0 (Ω)) × Vad × L2(Q) × L2(Q) that solves:

min

w,s,v Jγ(y, p, w, s, v) = J(y, p, w) + γ

2

  • Q

(s − g(y))2 + γ 2

  • Q

(v − g′(y)p)2 + 1 2p − ¯ p2

L2(Q) + 1

2

  • Q

(s − g(¯ y))2 + 1 2

  • Q

(v − g′(¯ y)¯ p)2 subject to: ∂y ∂t + Ay + s = in Q y =

  • n Σ

y(0) = −G−1p(0) in Ω −∂p ∂t + A∗p + v =

  • k,i

wkσiρi(t) [y(x, t) − zo(x, t)] ⊗ δ(x − xk) in Q p =

  • n Σ

p(T) = in Ω. Gu, τH−1,H1

0 :=

(u − ub)B−1τ + ϑ

∇(u − ub).∇τ = −

p(0)τ, ∀τ ∈ H1

0(Ω).

21 / 30

slide-48
SLIDE 48

Consistency

The penalized problem has at least one solution (yγ, pγ, wγ, sγ, vγ) ∈ H2,1(Q) × L2(W1,r

0 (Ω)) × Vad × L2(Q) × L2(Q) and

corresponding Lagrange multipliers (ηγ, ζγ) ∈ L2(Q) × H2,1(Q). Moreover, the sequence {(yγ, pγ, wγ, sγ, vγ)} converges strongly to the solution (¯ y, ¯ p, ¯ w, g(¯ y), g′(¯ y)¯ p). Structure of penalized cost and properties of state and adjoint eq. to get boundedness of variables; Compact embedding H2,1(Q) ֒ →֒ → Lµ(Q), with µ ≤ 10, and continuous embedding W1,r(Ω) ֒ → Lq(Ω) for q ≤

mr m−r;

Existence of multipliers using the linearity of the constraints. Using Hölder properties of Bochner spaces and properties of the PDE, we get uniform bounds on the regularized variables.

22 / 30

slide-49
SLIDE 49

Consistency

The penalized problem has at least one solution (yγ, pγ, wγ, sγ, vγ) ∈ H2,1(Q) × L2(W1,r

0 (Ω)) × Vad × L2(Q) × L2(Q) and

corresponding Lagrange multipliers (ηγ, ζγ) ∈ L2(Q) × H2,1(Q). Moreover, the sequence {(yγ, pγ, wγ, sγ, vγ)} converges strongly to the solution (¯ y, ¯ p, ¯ w, g(¯ y), g′(¯ y)¯ p). Structure of penalized cost and properties of state and adjoint eq. to get boundedness of variables; Compact embedding H2,1(Q) ֒ →֒ → Lµ(Q), with µ ≤ 10, and continuous embedding W1,r(Ω) ֒ → Lq(Ω) for q ≤

mr m−r;

Existence of multipliers using the linearity of the constraints. Using Hölder properties of Bochner spaces and properties of the PDE, we get uniform bounds on the regularized variables.

Lagrange multipliers

There exists an adjoint state (ηj, ζj, τj) ∈ L2(Q) × H2,1(Q) × H1

0(Ω), for all

j = 1, . . . , N, and KKT multipliers λa, λb ∈ Rns × RnT.

22 / 30

slide-50
SLIDE 50

Adjoint system of the bilevel problem

−∂ηj ∂t + A∗ηj + g′(yj)ηj + g′′(yj)pjζj =

  • k,i

wkσiρi(t)ζj(x, t) ⊗ δ(x − xk) − ∇yjL(yj) in Q ηj =0

  • n Σ

ηj(T) =0 in Ω. ∂ζj ∂t + Aζj + g′(yj)ζj =0 in Q ζj =0

  • n Σ

ζj(0) = − τj in Ω −ϑ∆τj + B−1τj =ηj(0) − β∇ujl(uj) in Ω τj =0

  • n Γ,

very weakly, strongly and weakly, respectively.

23 / 30

slide-51
SLIDE 51

Karush-Kuhn-Tucker condition

Gradient system: βw −

N

  • j=1

T

  • i

σiρi(t)ζj(xk, t) (yj(xk, t) − zoj(xk, t)) dt = λa

k − λb k,

∀k, βσ −

N

  • j=1

T

  • k

wkρi(t)ζj(xk, t) (yj(xk, t) − zoj(xk, t)) dt = λa

ns+i − λb ns+i,

∀i. Complementarity system: λa

r ≥ 0, λb r ≥ 0,

for all r = 1, . . . , ns + nT λa

k ¯

wk = λb

k(¯

wk − 1) = 0, for all k = 1, . . . , ns λa

ns+i¯

σi = λb

ns+i(¯

σi − 1) = 0, for all i = 1, . . . , nT 0 ≤ ¯ wk ≤ 1, for all k = 1, . . . , ns 0 ≤ ¯ σi ≤ 1, for all i = 1, . . . , nT.

24 / 30

slide-52
SLIDE 52

Outline

1

Motivation

2

Analysis of variational data assimilation

3

Optimal placement problem

4

Numerical results

5

Summary of topics

slide-53
SLIDE 53

Projected BFGS method

Let S be an index set and RS the matrix defined by RS = δij, if i ∈ S or j ∈ S; 0, if not. .

25 / 30

slide-54
SLIDE 54

Projected BFGS method

Let S be an index set and RS the matrix defined by RS = δij, if i ∈ S or j ∈ S; 0, if not. We denote by y# = RIǫk(wk)(y).

25 / 30

slide-55
SLIDE 55

Projected BFGS method

Let S be an index set and RS the matrix defined by RS = δij, if i ∈ S or j ∈ S; 0, if not. We denote by y# = RIǫk(wk)(y). Hk corresponds to the iteration matrix of the projected BFGS method Hk+1 = RIǫk(wk)HkRIǫk(wk) − RIǫk(wk) HksksT

k Hk

sT

k Hksk

RIǫk(wk) + y#

k (y# k )T

sT

k y# k

,

25 / 30

slide-56
SLIDE 56

Projected BFGS method

Let S be an index set and RS the matrix defined by RS = δij, if i ∈ S or j ∈ S; 0, if not. We denote by y# = RIǫk(wk)(y). Hk corresponds to the iteration matrix of the projected BFGS method Hk+1 = RIǫk(wk)HkRIǫk(wk) − RIǫk(wk) HksksT

k Hk

sT

k Hksk

RIǫk(wk) + y#

k (y# k )T

sT

k y# k

, We consider the recursive update of the inverse Kelly 1999: Bk+1 =

  • I − s#

k (y# k )T

(y#

k )Ts# k

  • RIBkRI
  • I − y#

k (s# k )T

(y#

k )Ts# k

  • + s#

k (s# k )T

(y#

k )Ts# k

,

25 / 30

slide-57
SLIDE 57

Implementation details

We solve in parallel N independent data assimilation optimality systems and N independent adjoint systems. The information is then integrated via the gradient formula The projected BFGS is used for the update of the placement vectors w and σ. The method is initialized with the identity matrix. Projected line-search rule with backtracking of the form

1 2k∇f(w0), k = 0, 1, . . .

The nonlinearity considered as example is g(y) =

y

y2+ε2

26 / 30

slide-58
SLIDE 58

Experiment 1

Placement is allowed in any grid point Observations are taken in every time step Goal: verify the descent of the cost functional along the iterations Goal: observe how the solution structure changes with respect to γ

27 / 30

slide-59
SLIDE 59

Experiment 1

Placement is allowed in any grid point Observations are taken in every time step Goal: verify the descent of the cost functional along the iterations Goal: observe how the solution structure changes with respect to γ

Figure: Descent of the cost function

27 / 30

slide-60
SLIDE 60

Experiment 1

Placement is allowed in any grid point Observations are taken in every time step Goal: verify the descent of the cost functional along the iterations Goal: observe how the solution structure changes with respect to γ

βw # zeros in w # ones in w 1 × 10−5 400 0.0001 88 312 0.0005 116 284 0.0020 176 224 0.0050 228 172 0.0072 268 132 0.0073 302 98 0.0074 334 66 0.0079 361 39 0.0080 400

27 / 30

slide-61
SLIDE 61

Experiment 1

Placement is allowed in any grid point Observations are taken in every time step Goal: verify the descent of the cost functional along the iterations Goal: observe how the solution structure changes with respect to γ

27 / 30

slide-62
SLIDE 62

Experiment 2

Placement is allowed only at 8 candidate locations x1 = (0.2, 0.2) x4 = (0.8, 0.0) x7 = (0.4, 0.9) x2 = (0.5, 0.4) x5 = (0.8, 1.0) x8 = (0.3, 0.8) x3 = (0.7, 0.3) x6 = (0.8, 0.6) Observations are taken in every time step Goal: observe how the solution structure changes with respect to the penalization parameters

28 / 30

slide-63
SLIDE 63

Experiment 2

Placement is allowed only at 8 candidate locations x1 = (0.2, 0.2) x4 = (0.8, 0.0) x7 = (0.4, 0.9) x2 = (0.5, 0.4) x5 = (0.8, 1.0) x8 = (0.3, 0.8) x3 = (0.7, 0.3) x6 = (0.8, 0.6) Observations are taken in every time step Goal: observe how the solution structure changes with respect to the penalization parameters

28 / 30

slide-64
SLIDE 64

Outline

1

Motivation

2

Analysis of variational data assimilation

3

Optimal placement problem

4

Numerical results

5

Summary of topics

slide-65
SLIDE 65

Summary

4D data assimilation problems in infinite dimensions: control spaces, type

  • f observations, nonlinear dynamics;

Bilevel learning (data-driven) approaches for estimating parameters in PDE-constrained optimization problems; Optimal placement of observations/sensors in inverse problems; Sparse solutions for PDE-constrained optimization problems; Singular control problems with measures as controls; Efficient numerical strategies for solving bilevel PDE-constrained

  • ptimization problems.

29 / 30

slide-66
SLIDE 66

IFIP TC7 Conference

August 31 - September 4, 2020 Save the date!

30 / 30