Analysis and design covariance inflation methods from functional - - PowerPoint PPT Presentation

analysis and design covariance inflation methods from
SMART_READER_LITE
LIVE PREVIEW

Analysis and design covariance inflation methods from functional - - PowerPoint PPT Presentation

Analysis and design covariance inflation methods from functional viewpoint Le Duc 13 , Kazuo Saito 23 , and Daisuke Hotta 3 1 Japan Agency for Marine-Earth Sciences and Technology 2 AORI, University of Tokyo 3 Meteorology Research Institute of


slide-1
SLIDE 1

Analysis and design covariance inflation methods from functional viewpoint

Le Duc13, Kazuo Saito23, and Daisuke Hotta3

1Japan Agency for Marine-Earth Sciences and Technology 2AORI, University of Tokyo 3Meteorology Research Institute of Japan

slide-2
SLIDE 2

Scope

Additive inflation and physics-based model uncertainty schemes are out of the scope of this study. Here, we focus on the covariance inflation (CI) methods applied at each data assimilation cycle that

  • nly

use forecast perturbations and observations to increase ensemble spreads.

slide-3
SLIDE 3

Covariance inflation methods

Prior multiplicative inflation (MI): belongs to prior CI

Posterior MI (inspired by prior MI): belongs to posterior CI

Relaxation schemes RTPP and RTPS: belongs to posterior

  • CI. The nature of these two methods is quite puzzling and

it seems that there are not any similar methods in prior CI.

= (1 − α) + α = + (1 − ) After this talk, I hope you will know more than 10 other CI methods!

slide-4
SLIDE 4

Questions on covariance inflation methods

Inflating is somehow equivalent to inflating , but posterior CI cannot say explicitly how is inflated. Because of this, posterior CI leaves mean analysis intact and only inflate analysis perturbations , leading

  • inconsistency. Can we find equivalently prior CI for each

posterior CI method? How can we explain the relaxation forms of RTPP and RTPS in posterior CI? Is there any other posterior CI method? If they exist, do they have relaxation forms? Can they have their counter- parts in prior CI?

slide-5
SLIDE 5

RTPP vs. RTPS

Given the similar relaxation forms of RTPP and RTPS, one may expect that perturbations produced by both methods should not have large differences. But this is far from the true. Whereas RTPS tends to generate analysis perturbations with small spreads, thus leading to under-dispersive ensemble forecasts, RTPP tends to be less under- dispersive with analysis spreads comparable to analysis errors (Bowler et al., 2017). No adequate interpretation has been given yet as to why we have such different behavior, and we will address this issue.

slide-6
SLIDE 6
  • 1. Background:

Decomposition form

  • f data assimilation
slide-7
SLIDE 7

Kalman filter theory

Then: = = + +

= − +

  • = +

= + where: = − = − = Assume we have a decomposition (ensemble, modulated ensemble, hybrid, …):

slide-8
SLIDE 8

More compact form

By setting: We have: =

  • =

= = +

  • =

= +

  • Here we replace the equation for with the equation for

using the positive symmetric square root :

slide-9
SLIDE 9

Singular value decomposition

Then the scaled background error covariance // is given by

  • =
  • Here ≤ min (, ) is the rank of

, where , are the dimensions

  • f
  • bservation

and ensemble space, respectively.

  • =
  • =
slide-10
SLIDE 10

Impact of assimilation

That means the component-wise scaling factors are

  • and

, respectively.

  • =
  • =
  • =
  • =
  • =
  • =
slide-11
SLIDE 11

Does this make sense?

If we assume that = and observations are of the same kind with observation error variance , we will have

=

  • /

where

  • is the background error variance

associated with the mode . And we recover DA for a scalar variable:

  • =
  • =
  • =
  • =
  • =

/( + )

  • =

(

+

slide-12
SLIDE 12

Perturbation response

This defines response of analysis singular values as a function of background singular values:

  • =
  • =
  • = =

1 + ⁄ Two important properties:

  • 1. Order-preserving:

the

  • rders of singular values

are preserved in assimilation, i.e. if ≥ then ≥

  • 2. Upper-bound:

analysis singular values are bounded from above by 1.

slide-13
SLIDE 13
  • 2. Inflation function
slide-14
SLIDE 14

Impact of ensemble transform matrix in ensemble space

Model space: Or if and represent background and analysis perturbations in ensemble space: Ensemble space: = = .

= .

  • That means for each jth row vector:

= = ,

  • In other words:

, = , ∀ ≤

slide-15
SLIDE 15

Idea of inflation functions

Not any function can be an inflation function. It has to satisfy the following conditions.

  • =
  • =

f

  • =
  • = f
  • f
slide-16
SLIDE 16

Definition

  • 1. The functional condition: analysis perturbations have

to increase at all modes.

  • 2. The no-observation condition: when no observations

are available, inflated analysis perturbations have to be identical to background perturbations.

  • 3. The order-preserving condition: the response function

= f () has to be an increasing function. Given any spectrum , ∈ [0,1] of , we associate with this spectrum a continuous function, which we coin the term an inflation function, f: [0,1] → [0, ∞) that maps to its inflated value f and satisfies the following three conditions:

slide-17
SLIDE 17

The functional condition

This requires: = ,

  • = f

,

  • < , > ≥ < , > ∀ ≤

f ≥ ∀ ≤ Extending to the whole domain, we require: Equivalently: f ≥ ∀ ∈ [0,1]

slide-18
SLIDE 18

The no-observation condition

This condition requires f 1 = 1 when all = 1 . To ensure continuity

  • f

inflation impact

  • n

analysis perturbations when spectra change continuously, we require f 1 ⟶ 1 whenever all ⟶ 1. = ,

  • = f

,

  • This condition involves a function that inflation functions

converge to when the spectrum , converges to the spectrum of the identity matrix.

slide-19
SLIDE 19

The order-preserving condition

This condition requires the inflated analysis singular values f() have to form an increasing function when considered as a function of . = ,

  • = f

,

slide-20
SLIDE 20

Examples

Here is a fixed inflation factor. No assimilation: f = 1 No inflation: f = Prior multiplicative inflation: f = 1 + 1 − 1

Posterior multiplicative inflation: f =

slide-21
SLIDE 21

Derivation of the inflation function corresponding to prior MI

If we inflate: → This is equivalent to: →

  • Then from:

This can be

  • btained

by changing

  • to

= 1/

1/ +

. In other words:

= +

= +

  • ⁄ =

⁄ +

We have: f = 1 + 1 − 1

slide-22
SLIDE 22

Examples

It is easy to check the three conditions: inflation functions lie above the line f = and go through the point (1,1), response functions are increasing functions.

slide-23
SLIDE 23

Prior MI vs. Posterior MI

For the first time, it shows clearly that the prior and posterior MI are two different methods. Both do not satisfy the no-observation condition.

slide-24
SLIDE 24
  • 3. Theory summary
slide-25
SLIDE 25

What I skip: prior inflation functions

Prior inflation functions help to explain for the unusual form of the inflation function of the prior MI method.

  • =
  • =

g 1 + g

  • =

g 1 + g

  • =
  • g
slide-26
SLIDE 26

Flow chart

slide-27
SLIDE 27

Implementation

Calculate the analysis increment at the jth point: Or to avoid calculation with − vectors in the null space of , we use the fact that .

= ∑

, .

  • to

rewrite:

  • =

.

  • =

< , . >

  • .

= f

, .

  • .

= . − 1 − f ) , .

slide-28
SLIDE 28
  • 4. Linear inflation

function

slide-29
SLIDE 29

Linear inflation function

If a, do not depend on the th point: f = + .

= f

, .

  • = , .
  • + , .
  • = .

+ .

  • = +
slide-30
SLIDE 30

The functional and order-preserving conditions

The functional and order-preserving conditions impose two conditions for linear inflation functions: f = + ≥ 0 + ≥ 1

slide-31
SLIDE 31

Constant inflation functions

If is fixed, the no-observation condition entails f 1 = = 1, which is f . Therefore, to obtain non-trivial solutions, we need to parameterize as a function of certain variables. f = These variables should characterize the spectrum , so that the no-observation condition when all = 1 can be described. There are various candidates in this case. We can list here the trace of : ∑ or ̅ = ∑ /, the determinant of : ∏ , the spectrum norm

  • f : , or the Frobenius norm of :

.

slide-32
SLIDE 32

Parameterization of constant inflation functions

We will mainly work with two quantities: / and ̅, which are both equal to 1 when no observations are assimilated. f = Here we choose ̅ to parameterize . Thus, we replace the function f = by the new form f = α̅ + where α, are two fixed hyper-parameters and a linear function, the simplest one, is chosen for the form of . The no-observation condition now yields the relation f 1 = α + = 1. And we have: f = α̅ + 1 − α = 1 − α(1 − ̅ Do you know what it is?

slide-33
SLIDE 33

Diagonal ETKF

When = 1, we recover the diagonal ETKF However, it is better to think that the generalized case defines the family of Diagonal ETKF. f = α̅ + 1 − α = 1 − α(1 − ̅ = (̅ + 1 − ) = ̅

slide-34
SLIDE 34

Constant inflation functions

Note we relax the functional condition by a weaker one: Norms of analysis perturbations have to increase

.

slide-35
SLIDE 35

Multiplicative inflation functions

If is fixed, the no-observation implies = 1, which is the function f . Therefore, like in constant inflation functions, cannot be fixed and has to be a function of certain variables. f = Here we first choose / to parameterize . Again, a linear function is chosen for the form of and we have f = (α

  • + )

with two fixed hyper- parameters α, .

slide-36
SLIDE 36

Parameterization of multiplicative inflation functions

The no-observation condition now yields the relation f 1 = α + = 1. And we have: f = (α

  • + )

f = α

  • + 1 − α =

1 + α(

  • − 1

In other words: = 1 + α(

  • − 1

Can you guess what it is?

slide-37
SLIDE 37

RTPS

From this, the norm of is given by = 1 + α(

  • − 1
  • = (
  • + 1 − )

= + (1 − ) = (1 − α) + α And we recover the well-known RTPS method

slide-38
SLIDE 38

Multiplicative inflation functions

slide-39
SLIDE 39

More surprising results

RTPS is not the unique realization of multiplicative inflation functions: f = 1 + α(

  • − 1

However, now it is more difficult to interpret the meaning of the resulting inflation functions. f = 1 + α(1 ̅ − 1 f = 1 + α(1 − f = 1 + α(1 − ̅

slide-40
SLIDE 40

Lorenz 40-variable experiments with different parameterizations of MIFs

This raises the question whether the use of / , and therefore the RTPS, is truly important or not. Clearly, it is not necessary to use / if we want to obtain the minimum RMSE.

slide-41
SLIDE 41

NHM-LETKF experiments with MIFs

Clearly, MIFLM with α = 2.3 had the same skill as RTPS with α = 0.9. The forecast performances would have been better if we had increased the values of α. f = 1 + α(1 − ̅

slide-42
SLIDE 42

Fixed linear inflation functions

Applying the no-observation condition, we must have f 1 = + = 1. If is a fixed parameter, this entails that is also a fixed parameter given by = 1 − . Therefore, we obtain: f = + f = + (1 − ) In other words: Can you guess what it is? = + (1 − ) RTPP!

slide-43
SLIDE 43

Fixed linear inflation functions

slide-44
SLIDE 44

RTPS vs. RTPP

The figures point out that RTPP are distinctively different from RTPS despite the apparent resemblance in relaxation

  • forms. Furthermore, this explains why RTPS tends to

produce under-dispersive analysis ensemble since RTPS sets an upper bound for magnitudes of the most active processes (large ) in inflated analysis perturbations.

slide-45
SLIDE 45

Parameter-varying linear inflation functions

The family of fixed linear inflations functions is quite restrictive, covering only a very small part of potential linear inflation functions. A way to extend the repertoire

  • f linear inflation functions is to use the fact that if f and

f are two inflation functions and the range of f belongs to [0,1] then the composite map f ∘ f is also an inflation function. This suggests that generalized linear inflation functions should have the following form: f = + = (α̅ + ) + (α̅ + ),

slide-46
SLIDE 46

Parameterization of generalized linear inflation functions

The no-observation condition dictates the first constraint f 1 = α + + α + = 1. The second constraint is determined by the converged form of f when all converge to 1. For generalized linear inflation functions, it is reasonable to choose f as its converged form. This means we impose the constraint ⟶ 0 when ̅ ⟶ 1, or equivalently α + =

  • 0. Finally, we have

f = + = (α̅ + ) + (α̅ + ), f = 1 + α(1 − ̅) + α(1 − ̅)

slide-47
SLIDE 47

Parameter-varying linear inflation functions

slide-48
SLIDE 48

Lorenz 40-variable experiments with PVLIFs

slide-49
SLIDE 49

NHM-LETKF experiments with PVLIFs

The best performance was obtained with α + α = 3.3, better forecasts might have been obtained if α and α + α were set to higher values.

slide-50
SLIDE 50

Different kinds of linear inflation functions

slide-51
SLIDE 51
  • 5. Quadratic inflation

function

slide-52
SLIDE 52

Quadratic inflation function

Compared with the case of linear inflation functions, we have an additional term ∑

  • , .
  • .

f = + +

.

= f

, .

  • =

, .

  • + .

+ .

  • , .
  • = .

− , .

  • = .

− .

  • Then we have:

.

= ( + ). − . + .

  • = ( + ) − +
slide-53
SLIDE 53

Quadratic inflation function with linear term omitted

For full quadratic inflation functions, the additional step compared to linear inflation functions is with . To

  • btain

we need to run the same program that produces the analysis times. However, the fact that we still need to run another program to generate analysis perturbations make this less practical. Therefore, it is better to set equal to zero so that analysis perturbations are not needed to be computed explicitly.

= ( + ) − +

f = +

= ( + ) −

slide-54
SLIDE 54

The functional and order-preserving conditions

The functional and order-preserving conditions impose four conditions for quadratic inflation functions without first-order terms: ≥ 0 + ≥ 1 f = + If ≥ 1/2, then c ≥ 1/4 c ⁄ ≤ 8

slide-55
SLIDE 55

Fixed quadratic inflation functions

If a and c are fixed, the no-observation condition yields f 1 = a + = 1. Thus, similar to the family of fixed linear inflation functions we have the family of fixed quadratic inflation functions: In other words: Can you guess what it is? f = + f = + (1 − ), = −

slide-56
SLIDE 56

Deterministic EnKF

The four conditions together imply the inequality ≤ 0.5. When = 0.5, the parabola + (1 − ) is tangent to the line f = at = 1. For this critical value of , we recover the well-known DEnKF (Sakov and Oke, 2008). It is quite surprising to see that the functional approach to inflation problem leads to a new interpretation for the DEnKF scheme. = − = − 1 2

slide-57
SLIDE 57

Parameter-varying quadratic inflation functions

Again, the family of fixed quadratic inflation functions is quite restrictive, and extension to parameter-varying quadratic inflation functions is straightforward. Now we choose the quadratic inflation function associated with DEnKF as the converged function when all converge to 1. By applying the same mathematical arguments we have the family of generalized quadratic inflation functions with two hyper-parameters α, α f = + = (α̅ + ) + (α̅ + ). f = 0.5 + α(1 − ̅ + 0.5 + α(1 − ̅

slide-58
SLIDE 58

Parameter-varying quadratic inflation functions

slide-59
SLIDE 59

Lorenz 40-variable experiments with PVQIFs

slide-60
SLIDE 60
  • 6. High-order

polynomial inflation function

slide-61
SLIDE 61

Cubic inflation function

Using the same arguments with quadratic inflation functions, it can be shown that the cubic form implies We can reduce the computational cost by setting = 0, but we still need to compute and . This shows that inflation using high-order polynomials consume more computational

  • resource. This is one of the reasons that makes high-order

polynomials not appealing in practice. However, the more important reason is that high-order polynomials do not necessarily yield better inflation than linear

  • r

quadratic polynomials.

f = + + + .

= ( + ). − . + ( + ). − .

  • = ( + ) − + ( + ) −
slide-62
SLIDE 62
  • 7. Summary
slide-63
SLIDE 63

The topics that I skip

  • Transformation

between prior inflation functions and inflation functions.

  • Non-polynomial inflation functions.
  • Explanation why the constraint α < 1 in RTPS is
  • irrelevant. Better analyses and forecasts can be
  • btained if we increase α over 1.
  • Adaptive inflation functions.
slide-64
SLIDE 64

Summary

  • Inflation functions can establish a robust foundation for the

CI methods.

  • The theory covers many existing CI methods in practice,

including multiplicative inflation, the RTPS, RTPP and DEnKF.

  • The theory proposes many new CI methods.
  • Constant inflation functions show existence of an unknown

family of EnKF: diagonal ETKF.

  • Linear inflation functions help to explain many properties
  • f RTPP and RTPS that we observe in practice.
  • Quadratic inflation functions with the first-order terms
  • mitted suggest a new form to calculate inflated analysis

perturbations based on the Kalman gain without a need for ensemble transform matrices.

slide-65
SLIDE 65

Existing CI methods under viewpoint

  • f inflation functions