EnKF and filter divergence David Kelly Andrew Stuart Kody Law - - PowerPoint PPT Presentation

enkf and filter divergence
SMART_READER_LITE
LIVE PREVIEW

EnKF and filter divergence David Kelly Andrew Stuart Kody Law - - PowerPoint PPT Presentation

EnKF and filter divergence David Kelly Andrew Stuart Kody Law Courant Institute New York University New York, NY dtbkelly.com December 11, 2014 Applied and computational mathematics seminar, NIST . David Kelly (NYU) EnKF December 11, 2014


slide-1
SLIDE 1

EnKF and filter divergence

David Kelly Andrew Stuart Kody Law

Courant Institute New York University New York, NY dtbkelly.com

December 11, 2014 Applied and computational mathematics seminar, NIST.

David Kelly (NYU) EnKF December 11, 2014 1 / 28

slide-2
SLIDE 2

Talk outline

  • 1. What is EnKF?
  • 2. What is known about EnKF?
  • 3. How can we use stochastic analysis to better

understand EnKF?

David Kelly (NYU) EnKF December 11, 2014 2 / 28

slide-3
SLIDE 3

The filtering problem

We have a deterministic model dv dt = F(v) with v0 ∼ N(m0, C0) . We will denote v(t) = Ψt(v0). Think of this as very high dimensional and nonlinear. We want to estimate vj = v(jh) for some h > 0 and j = 0, 1, . . . , J given the observations yj = Hvj + ξj for ξj iid N(0, Γ).

David Kelly (NYU) EnKF December 11, 2014 3 / 28

slide-4
SLIDE 4

The filtering problem

We have a deterministic model dv dt = F(v) with v0 ∼ N(m0, C0) . We will denote v(t) = Ψt(v0). Think of this as very high dimensional and nonlinear. We want to estimate vj = v(jh) for some h > 0 and j = 0, 1, . . . , J given the observations yj = Hvj + ξj for ξj iid N(0, Γ).

David Kelly (NYU) EnKF December 11, 2014 3 / 28

slide-5
SLIDE 5

We can write down the conditional density using Bayes’ formula ... But for high dimensional nonlinear systems it’s horrible.

David Kelly (NYU) EnKF December 11, 2014 4 / 28

slide-6
SLIDE 6

We can write down the conditional density using Bayes’ formula ... But for high dimensional nonlinear systems it’s horrible.

David Kelly (NYU) EnKF December 11, 2014 4 / 28

slide-7
SLIDE 7

Bayes’ formula filtering update

Let Y j = {y0, y1, . . . , yj}. We want to compute the conditional density P(vj+1|Y j+1), using P(vj|Y j) and yj+1. By Bayes’ formula, we have P(vj+1|Y j+1) = P(vj+1|Y j, yj+1) ∝ P(yj+1|vj+1)P(vj+1|Y j) But we need to compute the integral P(vj+1|Y j) =

  • P(vj+1|Y j, vj)P(vj|Y j)dvj .

For high dimensional nonlinear systems, this is computationally infeasible.

David Kelly (NYU) EnKF December 11, 2014 5 / 28

slide-8
SLIDE 8

Bayes’ formula filtering update

Let Y j = {y0, y1, . . . , yj}. We want to compute the conditional density P(vj+1|Y j+1), using P(vj|Y j) and yj+1. By Bayes’ formula, we have P(vj+1|Y j+1) = P(vj+1|Y j, yj+1) ∝ P(yj+1|vj+1)P(vj+1|Y j) But we need to compute the integral P(vj+1|Y j) =

  • P(vj+1|Y j, vj)P(vj|Y j)dvj .

For high dimensional nonlinear systems, this is computationally infeasible.

David Kelly (NYU) EnKF December 11, 2014 5 / 28

slide-9
SLIDE 9

Bayes’ formula filtering update

Let Y j = {y0, y1, . . . , yj}. We want to compute the conditional density P(vj+1|Y j+1), using P(vj|Y j) and yj+1. By Bayes’ formula, we have P(vj+1|Y j+1) = P(vj+1|Y j, yj+1) ∝ P(yj+1|vj+1)P(vj+1|Y j) But we need to compute the integral P(vj+1|Y j) =

  • P(vj+1|Y j, vj)P(vj|Y j)dvj .

For high dimensional nonlinear systems, this is computationally infeasible.

David Kelly (NYU) EnKF December 11, 2014 5 / 28

slide-10
SLIDE 10

The Ensemble Kalman Filter (EnKF) is a lower dimensional algorithm. (Evensen ’94) EnKF generates an ensemble of approximate samples from the posterior.

David Kelly (NYU) EnKF December 11, 2014 6 / 28

slide-11
SLIDE 11

The Ensemble Kalman Filter (EnKF) is a lower dimensional algorithm. (Evensen ’94) EnKF generates an ensemble of approximate samples from the posterior.

David Kelly (NYU) EnKF December 11, 2014 6 / 28

slide-12
SLIDE 12

For linear models, one can draw samples, using the Randomized Maximum Likelihood method.

David Kelly (NYU) EnKF December 11, 2014 7 / 28

slide-13
SLIDE 13

RML method

Let u ∼ N( m, C) and η ∼ N(0, Γ). We make an observation y = Hu + η . We want the conditional distribution of u|y. This is called an inverse problem. RML takes a sample { u(1), . . . , u(K)} ∼ N( m, C) and turns them into a sample {u(1), . . . , u(K)} ∼ u|y

David Kelly (NYU) EnKF December 11, 2014 8 / 28

slide-14
SLIDE 14

RML method

Let u ∼ N( m, C) and η ∼ N(0, Γ). We make an observation y = Hu + η . We want the conditional distribution of u|y. This is called an inverse problem. RML takes a sample { u(1), . . . , u(K)} ∼ N( m, C) and turns them into a sample {u(1), . . . , u(K)} ∼ u|y

David Kelly (NYU) EnKF December 11, 2014 8 / 28

slide-15
SLIDE 15

RML method: How does it work?

Along with the prior sample { u(1), . . . , u(K)}, we create artificial

  • bservations {y(1), . . . , y(K)} where

y(k) = y + η(k) where η(k) ∼ N(0, Γ) i.i.d Then define u(k) using the Bayes formula update, with ( u(k), y(k)) u(k) = u(k) + G( u)(y(k) − H u(k)) . Where the “Kalman Gain” G( u) is computing using the covariance of the prior u. The set {u(1), . . . , u(K)} are exact samples from u|y.

David Kelly (NYU) EnKF December 11, 2014 9 / 28

slide-16
SLIDE 16

RML method: How does it work?

Along with the prior sample { u(1), . . . , u(K)}, we create artificial

  • bservations {y(1), . . . , y(K)} where

y(k) = y + η(k) where η(k) ∼ N(0, Γ) i.i.d Then define u(k) using the Bayes formula update, with ( u(k), y(k)) u(k) = u(k) + G( u)(y(k) − H u(k)) . Where the “Kalman Gain” G( u) is computing using the covariance of the prior u. The set {u(1), . . . , u(K)} are exact samples from u|y.

David Kelly (NYU) EnKF December 11, 2014 9 / 28

slide-17
SLIDE 17

EnKF uses the same method, but with an approximation of the covariance in the Kalman gain.

David Kelly (NYU) EnKF December 11, 2014 10 / 28

slide-18
SLIDE 18

The set-up for EnKF

Suppose we are given the ensemble {u(1)

j

, . . . , u(K)

j

} at time j. For each ensemble member, we create an artificial observation y(k)

j+1 = yj+1 + ξ(k) j+1

, ξ(k)

j+1 iid N(0, Γ).

We update each particle using the Kalman update u(k)

j+1 = Ψh(u(k) j

) + G(uj)

  • y(k)

j+1 − HΨh(u(k) j

)

  • ,

where G(uj) is the Kalman gain computed using the forecasted ensemble covariance

  • Cj+1 = 1

K

K

  • k=1

(Ψh(u(k)

j

) − Ψh(uj))T(Ψh(u(k)

j

) − Ψh(uj)) .

David Kelly (NYU) EnKF December 11, 2014 11 / 28

slide-19
SLIDE 19

The set-up for EnKF

Suppose we are given the ensemble {u(1)

j

, . . . , u(K)

j

} at time j. For each ensemble member, we create an artificial observation y(k)

j+1 = yj+1 + ξ(k) j+1

, ξ(k)

j+1 iid N(0, Γ).

We update each particle using the Kalman update u(k)

j+1 = Ψh(u(k) j

) + G(uj)

  • y(k)

j+1 − HΨh(u(k) j

)

  • ,

where G(uj) is the Kalman gain computed using the forecasted ensemble covariance

  • Cj+1 = 1

K

K

  • k=1

(Ψh(u(k)

j

) − Ψh(uj))T(Ψh(u(k)

j

) − Ψh(uj)) .

David Kelly (NYU) EnKF December 11, 2014 11 / 28

slide-20
SLIDE 20

The set-up for EnKF

Suppose we are given the ensemble {u(1)

j

, . . . , u(K)

j

} at time j. For each ensemble member, we create an artificial observation y(k)

j+1 = yj+1 + ξ(k) j+1

, ξ(k)

j+1 iid N(0, Γ).

We update each particle using the Kalman update u(k)

j+1 = Ψh(u(k) j

) + G(uj)

  • y(k)

j+1 − HΨh(u(k) j

)

  • ,

where G(uj) is the Kalman gain computed using the forecasted ensemble covariance

  • Cj+1 = 1

K

K

  • k=1

(Ψh(u(k)

j

) − Ψh(uj))T(Ψh(u(k)

j

) − Ψh(uj)) .

David Kelly (NYU) EnKF December 11, 2014 11 / 28

slide-21
SLIDE 21

The set-up for EnKF

Suppose we are given the ensemble {u(1)

j

, . . . , u(K)

j

} at time j. For each ensemble member, we create an artificial observation y(k)

j+1 = yj+1 + ξ(k) j+1

, ξ(k)

j+1 iid N(0, Γ).

We update each particle using the Kalman update u(k)

j+1 = Ψh(u(k) j

) + G(uj)

  • y(k)

j+1 − HΨh(u(k) j

)

  • ,

where G(uj) is the Kalman gain computed using the forecasted ensemble covariance

  • Cj+1 = 1

K

K

  • k=1

(Ψh(u(k)

j

) − Ψh(uj))T(Ψh(u(k)

j

) − Ψh(uj)) .

David Kelly (NYU) EnKF December 11, 2014 11 / 28

slide-22
SLIDE 22

The set-up for EnKF

Suppose we are given the ensemble {u(1)

j

, . . . , u(K)

j

} at time j. For each ensemble member, we create an artificial observation y(k)

j+1 = yj+1 + ξ(k) j+1

, ξ(k)

j+1 iid N(0, Γ).

We update each particle using the Kalman update u(k)

j+1 = Ψh(u(k) j

) + G(uj)

  • y(k)

j+1 − HΨh(u(k) j

)

  • ,

where G(uj) is the Kalman gain computed using the forecasted ensemble covariance

  • Cj+1 = 1

K

K

  • k=1

(Ψh(u(k)

j

) − Ψh(uj))T(Ψh(u(k)

j

) − Ψh(uj)) .

David Kelly (NYU) EnKF December 11, 2014 11 / 28

slide-23
SLIDE 23

What do we know about EnKF? Not much. Theorem : For linear forecast models, ENKF → KF as N → ∞

(Le Gland et al / Mandel et al. 09’).

David Kelly (NYU) EnKF December 11, 2014 12 / 28

slide-24
SLIDE 24

What do we know about EnKF? Not much. Theorem : For linear forecast models, ENKF → KF as N → ∞

(Le Gland et al / Mandel et al. 09’).

David Kelly (NYU) EnKF December 11, 2014 12 / 28

slide-25
SLIDE 25

What do we know about EnKF? Not much. Theorem : For linear forecast models, ENKF → KF as N → ∞

(Le Gland et al / Mandel et al. 09’).

David Kelly (NYU) EnKF December 11, 2014 12 / 28

slide-26
SLIDE 26

Ideally, we would like a theorem about long time behaviour of the filter for a finite ensemble size.

David Kelly (NYU) EnKF December 11, 2014 13 / 28

slide-27
SLIDE 27

Filter divergence

In certain situations, it has been observed (⋆) that the ensemble can blow-up (ie. reach machine-infinity) in finite time, even when the model has nice bounded solutions. This is known as catastrophic filter divergence. We would like to investigate whether this has a dynamical justification

  • r if it is simply a numerical artefact.

⋆ Harlim, Majda (2010), Gottwald (2011), Gottwald, Majda (2013).

David Kelly (NYU) EnKF December 11, 2014 14 / 28

slide-28
SLIDE 28

Filter divergence

In certain situations, it has been observed (⋆) that the ensemble can blow-up (ie. reach machine-infinity) in finite time, even when the model has nice bounded solutions. This is known as catastrophic filter divergence. We would like to investigate whether this has a dynamical justification

  • r if it is simply a numerical artefact.

⋆ Harlim, Majda (2010), Gottwald (2011), Gottwald, Majda (2013).

David Kelly (NYU) EnKF December 11, 2014 14 / 28

slide-29
SLIDE 29

Filter divergence

In certain situations, it has been observed (⋆) that the ensemble can blow-up (ie. reach machine-infinity) in finite time, even when the model has nice bounded solutions. This is known as catastrophic filter divergence. We would like to investigate whether this has a dynamical justification

  • r if it is simply a numerical artefact.

⋆ Harlim, Majda (2010), Gottwald (2011), Gottwald, Majda (2013).

David Kelly (NYU) EnKF December 11, 2014 14 / 28

slide-30
SLIDE 30

Assumptions on the dynamics

We make a dissipativity assumption on the model. Namely that dv dt + Av + B(v, v) = f with A linear elliptic and B bilinear, satisfying certain estimates and symmetries. This guarantees uniformly bounded solutions.

  • Eg. 2d-Navier-Stokes, Lorenz-63, Lorenz-96.

David Kelly (NYU) EnKF December 11, 2014 15 / 28

slide-31
SLIDE 31

Discrete time results

For a fixed observation frequency h > 0 we can prove

Theorem (AS,DK,KL)

If H = Γ = Id then there exists constant β > 0 such that E|u(k)

j

|2 ≤ e2βjhE|u(k)

0 |2 + 2Kγ2

e2βjh − 1 e2βh − 1

  • Rmk. This becomes useless as h → 0

David Kelly (NYU) EnKF December 11, 2014 16 / 28

slide-32
SLIDE 32

Discrete time results

For a fixed observation frequency h > 0 we can prove

Theorem (AS,DK,KL)

If H = Γ = Id then there exists constant β > 0 such that E|u(k)

j

|2 ≤ e2βjhE|u(k)

0 |2 + 2Kγ2

e2βjh − 1 e2βh − 1

  • Rmk. This becomes useless as h → 0

David Kelly (NYU) EnKF December 11, 2014 16 / 28

slide-33
SLIDE 33

Discrete time results

For a fixed observation frequency h > 0 we can prove

Theorem (AS,DK,KL)

If H = Γ = Id then there exists constant β > 0 such that E|u(k)

j

|2 ≤ e2βjhE|u(k)

0 |2 + 2Kγ2

e2βjh − 1 e2βh − 1

  • Rmk. This becomes useless as h → 0

David Kelly (NYU) EnKF December 11, 2014 16 / 28

slide-34
SLIDE 34

Discrete time results with variance inflation

Suppose we replace

  • Cj+1 → α2I +

Cj+1 at each update step. This is known as additive variance inflation.

Theorem (AS,DK,KL)

If H = Id and Γ = γ2Id then there exists constant β > 0 such that E|e(k)

j

|2 ≤ θjE|e(k)

0 |2 + 2Kγ2

1 − θj 1 − θ

  • where θ = γ2e2βh

α2+γ2 .

In particular, if we pick α large enough (so that θ < 1) then lim

j→∞ E|e(k) j

|2 ≤ 2Kγ2 1 − θ

David Kelly (NYU) EnKF December 11, 2014 17 / 28

slide-35
SLIDE 35

Discrete time results with variance inflation

Suppose we replace

  • Cj+1 → α2I +

Cj+1 at each update step. This is known as additive variance inflation.

Theorem (AS,DK,KL)

If H = Id and Γ = γ2Id then there exists constant β > 0 such that E|e(k)

j

|2 ≤ θjE|e(k)

0 |2 + 2Kγ2

1 − θj 1 − θ

  • where θ = γ2e2βh

α2+γ2 .

In particular, if we pick α large enough (so that θ < 1) then lim

j→∞ E|e(k) j

|2 ≤ 2Kγ2 1 − θ

David Kelly (NYU) EnKF December 11, 2014 17 / 28

slide-36
SLIDE 36

Discrete time results with variance inflation

Suppose we replace

  • Cj+1 → α2I +

Cj+1 at each update step. This is known as additive variance inflation.

Theorem (AS,DK,KL)

If H = Id and Γ = γ2Id then there exists constant β > 0 such that E|e(k)

j

|2 ≤ θjE|e(k)

0 |2 + 2Kγ2

1 − θj 1 − θ

  • where θ = γ2e2βh

α2+γ2 .

In particular, if we pick α large enough (so that θ < 1) then lim

j→∞ E|e(k) j

|2 ≤ 2Kγ2 1 − θ

David Kelly (NYU) EnKF December 11, 2014 17 / 28

slide-37
SLIDE 37

For observations with h ≪ 1, we need another approach.

David Kelly (NYU) EnKF December 11, 2014 18 / 28

slide-38
SLIDE 38

The EnKF equations look like a discretization

Recall the ensemble update equation u(k)

j+1 = Ψh(u(k) j

) + G(uj)

  • y(k)

j+1 − HΨh(u(k) j

)

  • = Ψh(u(k)

j

) + Cj+1HT(HT Cj+1H + Γ)−1 y(k)

j+1 − HΨh(u(k) j

)

  • Subtract u(k)

j

from both sides and divide by h u(k)

j+1 − u(k) j

h = Ψh(u(k)

j

) − u(k)

j

h + Cj+1HT(hHT Cj+1H + hΓ)−1 y(k)

j+1 − HΨh(u(k) j

)

  • Clearly we need to rescale the noise (ie. Γ).

David Kelly (NYU) EnKF December 11, 2014 19 / 28

slide-39
SLIDE 39

The EnKF equations look like a discretization

Recall the ensemble update equation u(k)

j+1 = Ψh(u(k) j

) + G(uj)

  • y(k)

j+1 − HΨh(u(k) j

)

  • = Ψh(u(k)

j

) + Cj+1HT(HT Cj+1H + Γ)−1 y(k)

j+1 − HΨh(u(k) j

)

  • Subtract u(k)

j

from both sides and divide by h u(k)

j+1 − u(k) j

h = Ψh(u(k)

j

) − u(k)

j

h + Cj+1HT(hHT Cj+1H + hΓ)−1 y(k)

j+1 − HΨh(u(k) j

)

  • Clearly we need to rescale the noise (ie. Γ).

David Kelly (NYU) EnKF December 11, 2014 19 / 28

slide-40
SLIDE 40

The EnKF equations look like a discretization

Recall the ensemble update equation u(k)

j+1 = Ψh(u(k) j

) + G(uj)

  • y(k)

j+1 − HΨh(u(k) j

)

  • = Ψh(u(k)

j

) + Cj+1HT(HT Cj+1H + Γ)−1 y(k)

j+1 − HΨh(u(k) j

)

  • Subtract u(k)

j

from both sides and divide by h u(k)

j+1 − u(k) j

h = Ψh(u(k)

j

) − u(k)

j

h + Cj+1HT(hHT Cj+1H + hΓ)−1 y(k)

j+1 − HΨh(u(k) j

)

  • Clearly we need to rescale the noise (ie. Γ).

David Kelly (NYU) EnKF December 11, 2014 19 / 28

slide-41
SLIDE 41

The EnKF equations look like a discretization

Recall the ensemble update equation u(k)

j+1 = Ψh(u(k) j

) + G(uj)

  • y(k)

j+1 − HΨh(u(k) j

)

  • = Ψh(u(k)

j

) + Cj+1HT(HT Cj+1H + Γ)−1 y(k)

j+1 − HΨh(u(k) j

)

  • Subtract u(k)

j

from both sides and divide by h u(k)

j+1 − u(k) j

h = Ψh(u(k)

j

) − u(k)

j

h + Cj+1HT(hHT Cj+1H + hΓ)−1 y(k)

j+1 − HΨh(u(k) j

)

  • Clearly we need to rescale the noise (ie. Γ).

David Kelly (NYU) EnKF December 11, 2014 19 / 28

slide-42
SLIDE 42

Continuous-time limit

If we set Γ = h−1Γ0 and substitute y(k)

j+1, we obtain

u(k)

j+1 − u(k) j

h = Ψh(u(k)

j

) − u(k)

j

h + Cj+1HT(hHT Cj+1H + Γ0)−1

  • Hv + h−1/2Γ1/2

ξj+1 + h−1/2Γ1/2 ξ(k)

j+1 − HΨh(u(k) j

)

  • But we know that

Ψh(u(k)

j

) = u(k)

j

+ O(h) and

  • Cj+1 = 1

K

K

  • k=1

(Ψh(u(k)

j

) − Ψh(uj))T(Ψh(u(k)

j

) − Ψh(uj)) = 1 K

K

  • k=1

(u(k)

j

− uj)T(u(k)

j

− uj) + O(h) = C(uj) + O(h)

David Kelly (NYU) EnKF December 11, 2014 20 / 28

slide-43
SLIDE 43

Continuous-time limit

If we set Γ = h−1Γ0 and substitute y(k)

j+1, we obtain

u(k)

j+1 − u(k) j

h = Ψh(u(k)

j

) − u(k)

j

h + Cj+1HT(hHT Cj+1H + Γ0)−1

  • Hv + h−1/2Γ1/2

ξj+1 + h−1/2Γ1/2 ξ(k)

j+1 − HΨh(u(k) j

)

  • But we know that

Ψh(u(k)

j

) = u(k)

j

+ O(h) and

  • Cj+1 = 1

K

K

  • k=1

(Ψh(u(k)

j

) − Ψh(uj))T(Ψh(u(k)

j

) − Ψh(uj)) = 1 K

K

  • k=1

(u(k)

j

− uj)T(u(k)

j

− uj) + O(h) = C(uj) + O(h)

David Kelly (NYU) EnKF December 11, 2014 20 / 28

slide-44
SLIDE 44

Continuous-time limit

We end up with u(k)

j+1 − u(k) j

h = Ψh(u(k)

j

) − u(k)

j

h − C(uj)HTΓ−1

0 H(u(k) j

− vj) + C(uj)HTΓ−1

  • h−1/2ξj+1 + h−1/2ξ(k)

j+1

  • + O(h)

This looks like a numerical scheme for Itˆ

  • S(P)DE

du(k) dt = F(u(k)) − C(u)HTΓ−1

0 H(u(k) − v)

(•) + C(u)HTΓ−1/2

  • dB

dt + dW (k) dt

  • .

David Kelly (NYU) EnKF December 11, 2014 21 / 28

slide-45
SLIDE 45

Continuous-time limit

We end up with u(k)

j+1 − u(k) j

h = Ψh(u(k)

j

) − u(k)

j

h − C(uj)HTΓ−1

0 H(u(k) j

− vj) + C(uj)HTΓ−1

  • h−1/2ξj+1 + h−1/2ξ(k)

j+1

  • + O(h)

This looks like a numerical scheme for Itˆ

  • S(P)DE

du(k) dt = F(u(k)) − C(u)HTΓ−1

0 H(u(k) − v)

(•) + C(u)HTΓ−1/2

  • dB

dt + dW (k) dt

  • .

David Kelly (NYU) EnKF December 11, 2014 21 / 28

slide-46
SLIDE 46

Nudging

du(k) dt = F(u(k)) − C(u)HTΓ−1

0 H(u(k) − v)

(•) + C(u)HTΓ−1/2

  • dB

dt + dW (k) dt

  • .

1 - Extra dissipation term only sees differences in observed space 2 - Extra dissipation only occurs in the space spanned by ensemble

David Kelly (NYU) EnKF December 11, 2014 22 / 28

slide-47
SLIDE 47

Nudging

du(k) dt = F(u(k)) − C(u)HTΓ−1

0 H(u(k) − v)

(•) + C(u)HTΓ−1/2

  • dB

dt + dW (k) dt

  • .

1 - Extra dissipation term only sees differences in observed space 2 - Extra dissipation only occurs in the space spanned by ensemble

David Kelly (NYU) EnKF December 11, 2014 22 / 28

slide-48
SLIDE 48

Nudging

du(k) dt = F(u(k)) − C(u)HTΓ−1

0 H(u(k) − v)

(•) + C(u)HTΓ−1/2

  • dB

dt + dW (k) dt

  • .

1 - Extra dissipation term only sees differences in observed space 2 - Extra dissipation only occurs in the space spanned by ensemble

David Kelly (NYU) EnKF December 11, 2014 22 / 28

slide-49
SLIDE 49

Kalman-Bucy limit

If F were linear and we write m(t) = 1

K

K

k=1 u(k)(t) then

dm dt = F(m) − C(u)HTΓ−1

0 H(m − v)

+ C(u)HTΓ−1/2 dB dt + O(K −1/2) . This is the equation for the Kalman-Bucy filter, with empirical covariance C(u). The remainder O(K −1/2) can be thought of as a sampling error.

David Kelly (NYU) EnKF December 11, 2014 23 / 28

slide-50
SLIDE 50

Kalman-Bucy limit

If F were linear and we write m(t) = 1

K

K

k=1 u(k)(t) then

dm dt = F(m) − C(u)HTΓ−1

0 H(m − v)

+ C(u)HTΓ−1/2 dB dt + O(K −1/2) . This is the equation for the Kalman-Bucy filter, with empirical covariance C(u). The remainder O(K −1/2) can be thought of as a sampling error.

David Kelly (NYU) EnKF December 11, 2014 23 / 28

slide-51
SLIDE 51

Continuous-time results

Theorem (AS,DK)

Suppose that{u(k)}K

k=1 satisfy (•) with H = Γ = Id. Let

e(k) = u(k) − v . Then there exists constant β > 0 such that 1 K

K

  • k=1

E|e(k)(t)|2 ≤ 1 K

K

  • k=1

E|e(k)(0)|2

  • exp (βt) .

David Kelly (NYU) EnKF December 11, 2014 24 / 28

slide-52
SLIDE 52

Why do we need H = Γ = Id ?

In the equation du(k) dt = F(u(k)) − C(u)HTΓ−1

0 H(u(k) − v)

+ C(u)HTΓ−1/2

  • dW (k)

dt + dB dt

  • .

The energy pumped in by the noise must be balanced by contraction of (u(k) − v). So the operator C(u)HTΓ−1

0 H

must be positive-definite. Both C(u) and HTΓ−1

0 H are pos-def, but this doesn’t guarantee the same

for the product!

David Kelly (NYU) EnKF December 11, 2014 25 / 28

slide-53
SLIDE 53

Why do we need H = Γ = Id ?

In the equation du(k) dt = F(u(k)) − C(u)HTΓ−1

0 H(u(k) − v)

+ C(u)HTΓ−1/2

  • dW (k)

dt + dB dt

  • .

The energy pumped in by the noise must be balanced by contraction of (u(k) − v). So the operator C(u)HTΓ−1

0 H

must be positive-definite. Both C(u) and HTΓ−1

0 H are pos-def, but this doesn’t guarantee the same

for the product!

David Kelly (NYU) EnKF December 11, 2014 25 / 28

slide-54
SLIDE 54

Why do we need H = Γ = Id ?

In the equation du(k) dt = F(u(k)) − C(u)HTΓ−1

0 H(u(k) − v)

+ C(u)HTΓ−1/2

  • dW (k)

dt + dB dt

  • .

The energy pumped in by the noise must be balanced by contraction of (u(k) − v). So the operator C(u)HTΓ−1

0 H

must be positive-definite. Both C(u) and HTΓ−1

0 H are pos-def, but this doesn’t guarantee the same

for the product!

David Kelly (NYU) EnKF December 11, 2014 25 / 28

slide-55
SLIDE 55

Testing stability on the fly

Suppose we can actually measure the spectrum of the operator C(u)HTΓ−1

0 H

whilst the algorithm is running. If we know that it is pos-def, then the filter must not be blowing up. If we knew that C(u)HTΓ−1

0 H ≥ λ(t) > 0 .

Then we can say even more (eg. stability).

David Kelly (NYU) EnKF December 11, 2014 26 / 28

slide-56
SLIDE 56

Testing stability on the fly

Suppose we can actually measure the spectrum of the operator C(u)HTΓ−1

0 H

whilst the algorithm is running. If we know that it is pos-def, then the filter must not be blowing up. If we knew that C(u)HTΓ−1

0 H ≥ λ(t) > 0 .

Then we can say even more (eg. stability).

David Kelly (NYU) EnKF December 11, 2014 26 / 28

slide-57
SLIDE 57

Summary + Future Work

(1) Writing down an SDE/SPDE allows us to see the important quantities in the algorithm. (2) Does not “prove” that catastrophic filter divergence is a numerical phenomenon, but is a decent starting point. (1) Improve the condition on H. (2) If we can measure the important quantities, then we can test the performance during the algorithm. (3) Make use of controllability and observability.

David Kelly (NYU) EnKF December 11, 2014 27 / 28

slide-58
SLIDE 58

Summary + Future Work

(1) Writing down an SDE/SPDE allows us to see the important quantities in the algorithm. (2) Does not “prove” that catastrophic filter divergence is a numerical phenomenon, but is a decent starting point. (1) Improve the condition on H. (2) If we can measure the important quantities, then we can test the performance during the algorithm. (3) Make use of controllability and observability.

David Kelly (NYU) EnKF December 11, 2014 27 / 28

slide-59
SLIDE 59

Thank you! Well-posedness and accuracy of the ensemble Kalman filter in discrete and continuous time.

  • D. Kelly, K.Law, A. Stuart.

Nonlinearity 2014. www.dtbkelly.com

David Kelly (NYU) EnKF December 11, 2014 28 / 28