combining data assimilation and machine learning to
play

Combining Data Assimilation and Machine Learning to emulate a - PowerPoint PPT Presentation

Combining Data Assimilation and Machine Learning to emulate a numerical model Julien Brajard, Alberto Carrassi, Marc Bocquet, Laurent Bertino 05 June 2019 NERSC, LOCEAN-IPSL-Sorbonne Universit, CEREA 1 Motivation Chloropyhll-a (Model) July


  1. Combining Data Assimilation and Machine Learning to emulate a numerical model Julien Brajard, Alberto Carrassi, Marc Bocquet, Laurent Bertino 05 June 2019 NERSC, LOCEAN-IPSL-Sorbonne Université, CEREA 1

  2. Motivation Chloropyhll-a (Model) July 26, 2018 TOPAZ4-ECOSMO forecast • Unresolved process • Unknown parameters Chloropyhll-a (Observation) July 26, 2018 MODIS Aqua • Sparse • Noisy 2

  3. Motivation Chloropyhll-a (Model) July 26, 2018 TOPAZ4-ECOSMO forecast • Unresolved process • Unknown parameters Chloropyhll-a (Observation) July 26, 2018 MODIS Aqua • Sparse • Noisy 2

  4. Motivation Chloropyhll-a (Model) July 26, 2018 TOPAZ4-ECOSMO forecast • Unresolved process • Unknown parameters Chloropyhll-a (Observation) July 26, 2018 MODIS Aqua • Sparse • Noisy 2

  5. Typology of problems and approaches State State + Model pa- rameters State + co- efficients of the ODE State + Emulator Emulator Data Assimilation DA+ML Machine Learning 3

  6. Typology of problems and approaches Emulator Machine Learning DA+ML Data Assimilation State (sparse, noisy) Imperfect observations 3 Emulator of the ODE efficients State + State + co- Model pa- rameters State + l e d o m t c e f r e p

  7. Typology of problems and approaches Emulator Machine Learning DA+ML Data Assimilation State (sparse, noisy) Imperfect observations 3 Emulator of the ODE efficients State + State + co- Model pa- rameters State + l e d o m t c e f r e p

  8. Typology of problems and approaches Imperfect observations Machine Learning DA+ML Data Assimilation State observations Perfect (sparse, noisy) 3 Emulator Emulator State + of the ODE efficients State + Model pa- rameters State + co- (no ODE) l e d o m l e d t c o e m f r e o p n

  9. Typology of problems and approaches Imperfect observations Machine Learning DA+ML Data Assimilation State observations Perfect (sparse, noisy) 3 Emulator Emulator State + of the ODE efficients State + Model pa- rameters State + co- (no ODE) l e d o m l e d t c o e m f r e o p n

  10. Typology of problems and approaches Imperfect observations Machine Learning DA+ML Data Assimilation State observations Perfect (sparse, noisy) 3 Emulator State + co- State + of the ODE efficients State + Model pa- rameters Emulator (no ODE) l e d l e o d m o m t l c e e d t f c o r e e m f p r e m o p n i

  11. Typology of problems and approaches Imperfect observations Machine Learning DA+ML Data Assimilation State observations Perfect (sparse, noisy) 3 Emulator State + co- State + of the ODE efficients State + Model pa- rameters Emulator (no ODE) l e d (general ODE form) l e o l e d m d o o m t l c m e e d t f c o o r e e n m f p r e m o p n i

  12. Typology of problems and approaches Imperfect observations Machine Learning DA+ML Data Assimilation State observations Perfect (sparse, noisy) 3 Emulator State + co- State + of the ODE efficients State + Model pa- rameters Emulator (no ODE) l e d (general ODE form) l e o l e d m d o o m t l c m e e d t f c o o r e e n m f p r e m o p n i

  13. Typology of problems and approaches Imperfect observations This talk Machine Learning DA+ML Data Assimilation State observations Perfect (sparse, noisy) 3 Emulator rameters efficients State + co- Emulator Model pa- State + of the ODE State + (no ODE) l e d (general ODE form) l e o l e d m d o o m t l m c e e d t f c o o r e e n m f p r e m o p n i

  14. Our Objective: Producing an accurate and reliable emulator of a numerical model given sparse and noisy observations 3

  15. Specification of the problem Underlying dynamical model: x d t t k 1 t k x k 1 x k Resolvent: x d t d x is a noise Data k k k y obs underlying dynamical process: k Multidimensional time series y obs 4 (1 ≤ k ≤ K ) observed from an = H k ( x k ) + ϵ obs • H k is the known observation operator: R N x → R p • ϵ obs

  16. Specification of the problem Data x d t t k 1 t k x k 1 x k Resolvent: d x Underlying dynamical model: is a noise k k k y obs underlying dynamical process: k Multidimensional time series y obs 4 (1 ≤ k ≤ K ) observed from an = H k ( x k ) + ϵ obs • H k is the known observation operator: R N x → R p • ϵ obs d t = φ ( x )

  17. Specification of the problem k t k Resolvent: d x Underlying dynamical model: is a noise Data k k y obs underlying dynamical process: k Multidimensional time series y obs 4 (1 ≤ k ≤ K ) observed from an = H k ( x k ) + ϵ obs • H k is the known observation operator: R N x → R p • ϵ obs d t = φ ( x ) t k + 1 ∫ x k + 1 = x k + φ ( x ) d t ,

  18. 2. Emulation of the resolvent combining DA and ML [Brajard et al., 2019] : W x k W is a neural network parametrized by W and k is a m stochastic noise. m where k Two complementary goals 1. Inferring the ODE using only DA algorithm [Bocquet et al., 2019] : 1 x k d x 5 d t = φ A ( x ) , φ A ( x ) = Ar ( x ) , where r ( x ) ∈ R N p is specified and A ∈ R N x × N p is to be determined.

  19. Two complementary goals 1. Inferring the ODE using only DA algorithm [Bocquet et al., 2019] : d x stochastic noise. 5 d t = φ A ( x ) , φ A ( x ) = Ar ( x ) , where r ( x ) ∈ R N p is specified and A ∈ R N x × N p is to be determined. 2. Emulation of the resolvent combining DA and ML [Brajard et al., 2019] : x k + 1 = G W ( x k ) + ϵ m k , where G W is a neural network parametrized by W and ϵ m k is a

  20. First goal: Inferring the ODE using DA

  21. 2 N x 9 . First goal: ODE representation for the surrogate model N p 10 6 Intractable in high-dimension! Typically, N x 2 . 1 N x 1 2 1 N x 6 Ordinary differential equations (ODEs) representation of the for one-dimensional spatial systems and up to bilinear order: where d x surrogate dynamics d t = φ A ( x ) , φ A ( x ) = Ar ( x ) , • A ∈ R N x × N p is a matrix of coefficients to be determined. • r ( x ) is a vector of nonlinear regressors of size N p . For instance, [ ] r ( x ) = 1 , { x n } 0 ≤ n < N x , { x n x m } 0 ≤ n ≤ m < N x .

  22. First goal: ODE representation for the surrogate model where 2 Ordinary differential equations (ODEs) representation of the for one-dimensional spatial systems and up to bilinear order: 6 surrogate dynamics d x d t = φ A ( x ) , φ A ( x ) = Ar ( x ) , • A ∈ R N x × N p is a matrix of coefficients to be determined. • r ( x ) is a vector of nonlinear regressors of size N p . For instance, [ ] r ( x ) = 1 , { x n } 0 ≤ n < N x , { x n x m } 0 ≤ n ≤ m < N x . ( N x + 1 ) N p = = 1 2 ( N x + 1 )( N x + 2 ) . → Intractable in high-dimension! Typically, N x = O ( 10 6 − 9 ) . −

  23. Moreover, we can additionally assume translational invariance. In Reducing the number of regressors where that case A becomes a vector of size N a . Homogeneity Locality 7 arrangement of grid points around a given node. Physical locality of the physics: all multivariate monomials in the ODEs have variables x n that belong to a stencil, i.e. a local In 1D and with a stencil of size 2 L + 1, the size of the dense A is 2 L + 2 ∑ N x × N a N a = l = 3 2 ( L + 1 )( L + 2 ) . l = L + 1

  24. Reducing the number of regressors where that case A becomes a vector of size N a . Homogeneity Locality 7 Physical locality of the physics: all multivariate monomials in the arrangement of grid points around a given node. ODEs have variables x n that belong to a stencil, i.e. a local In 1D and with a stencil of size 2 L + 1, the size of the dense A is 2 L + 2 ∑ N x × N a N a = l = 3 2 ( L + 1 )( L + 2 ) . l = L + 1 Moreover, we can additionally assume translational invariance. In

  25. H k x k F A x k p x 0 A where F A is the resolvent of the model between t k and t k Allows to handle partial and noisy observations. Typical machine learning cost function with H k I k in the limit R k F A y k p y 0 A Bayesian analysis of the problem k t . 0 : A 2 1 Q K k 1 y k 1 2 Q 1 k 1 1 2 Bayesian view on state and model estimation: Data assimilation cost function assuming Gaussian error statistics and Markovian dynamics: A x 0 K 1 2 K k 0 y k 2 R 1 k 1 2 K k 1 x k 8 p ( A , x 0 : K | y 0 : K ) = p ( y 0 : K | x 0 : K , A ) p ( x 0 : K | A ) p ( A ) . p ( y 0 : K )

  26. Typical machine learning cost function with H k I k in the limit R k F A y k p y 0 A k 0 : A 1 2 K Bayesian analysis of the problem 1 Bayesian view on state and model estimation: 1 2 Q 1 k y k k 8 K Data assimilation cost function assuming Gaussian error statistics and Markovian dynamics: K 2 k 2 p ( A , x 0 : K | y 0 : K ) = p ( y 0 : K | x 0 : K , A ) p ( x 0 : K | A ) p ( A ) . p ( y 0 : K ) ∑ ∑ J ( A , x 0 : K ) = 1 ∥ y k − H k ( x k ) ∥ 2 + 1 ∥ x k − F A ( x k − 1 ) ∥ 2 − ln p ( x 0 , A ) , R − 1 Q − 1 k = 0 k = 1 where F A is the resolvent of the model between t k and t k + ∆ t . − → Allows to handle partial and noisy observations.

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend