Uncertainty Quantification Framework for Modeling Prediction - - PowerPoint PPT Presentation

uncertainty quantification framework for modeling
SMART_READER_LITE
LIVE PREVIEW

Uncertainty Quantification Framework for Modeling Prediction - - PowerPoint PPT Presentation

Columbia University October 10, 2014 Uncertainty Quantification Framework for Modeling Prediction Michael Frenklach Collaborators: Andy Packard W. A. Lester, Jr. (DFT) P. Westmoreland (ALS) N. Slavinskaya (SynGas) R.


slide-1
SLIDE 1

Uncertainty Quantification Framework for Modeling Prediction

Michael Frenklach

Columbia University October 10, 2014 Supported by: NSF, AFOSR, DOE‐NNSA (PSAAP II) Collaborators:

‒ Andy Packard

‒ W. A. Lester, Jr. (DFT) ‒ P. Westmoreland (ALS) ‒ N. Slavinskaya (SynGas) ‒ R. Feely, P. Seiler, T. Russi, D. Yeates, X. You, F. Lei,

  • D. Edwards, D. Zubarev, W. Speight
slide-2
SLIDE 2
  • Introduction: UQ-predictive modeling
  • Bound-To-Bound Data Collaboration
  • Introductory case: Energetics of water clusters
  • Full-blown case: Combustion of natural gas
slide-3
SLIDE 3
slide-4
SLIDE 4

“Model predicts data” ?

“Model predicts reasonably well the experimental behavior” “…excellent agreement between model and data.”

“Model falls short in predicting experimental data” “The model predictions match reasonably well the experimental data” “The model well predicts the data”

“Model matches the experimental data” “The prediction matches very well with experimental data” “Simulation agrees well with the data” “Good agreement was found between the model and the data”

slide-5
SLIDE 5
slide-6
SLIDE 6

WWW 2012 Lyon, France

We develop a temporal modeling framework adapted from physics and signal processing … The results … indicate that by using

  • ur framework …. we can achieve

significant improvements in prediction compare to baseline models …

slide-7
SLIDE 7
  • Predictive  UQ‐Predictive
  • Physics‐based models with the focus on data
  • Validation is part of the process
  • Dimensionality reduction is part of the process
  • Practicality  use of surrogate models (Emulators)
  • Data/Models
  • Access, sharing, documentation, …
  • Reproducibility
slide-8
SLIDE 8

model

(dif eq, nonlinear)

slide-9
SLIDE 9

model

(dif eq, nonlinear)

slide-10
SLIDE 10
slide-11
SLIDE 11
slide-12
SLIDE 12
slide-13
SLIDE 13

x2 x3 x1,min  x1  x1,max

prior knowledge on parameters experimental uncertainties B2B‐DC

y2 y3 L1  y1  U1

H D

─ an optimization‐based framework for combining models and data to ascertain the collective information content

slide-14
SLIDE 14

Phys Rev Lett 112: 253003 (2014)

slide-15
SLIDE 15
  • empirical: force‐field – guessed potential, empirically fitted; …
  • semi‐empirical HF – quantum “core” with some terms replaced by

parameters fitted to data (AM1, RM1, PM3, PM6, ZINDO, …)

  • DFT with fitted parameters: meta‐GGA (Truhlar, M05,M06,M11,…),

double‐hybrid DFT (Grimme), ...

  • “static” outcome: the optimized model

needs (constant) retuning

  • the optimum is not unique!
  • partial loss of information (two‐step process)

   

DFT HF DFT MP2 XC X X C C X C C X

1 1 E E E E E          

training data parameters blind prediction

model‐based UQ

slide-16
SLIDE 16

   

GGA HF GGA MP2 XC X X C C

1 1 E E E E E           Model:

use E intervals computed for dimer, trimer, tetramer, and pentamer to predict E interval of hexamer

Data Solve for:

slide-17
SLIDE 17

   

GGA HF GGA MP2 XC X X C C

1 1 E E E E E          

VIP

 

high‐level theory result (or experiment) Combined Feasible Set

Hexamer prediction (kcal/mol):

Source Min Max Range Over Feasible Set 267.7 269.7 2.0 Segarra‐Marti et al. 265.9 270.0 4.1
slide-18
SLIDE 18
  • mixture of mostly methane with other light gases
  • lowest emissions among fossil fuels; no soot;

smallest carbon footprint

  • various, expanding sources (biofuels, artificial synthesis,…)
  • plenty and cheap; booming US (and world) economy
  • technology issues/needs
  • varying compositions – hard to categorize empirically
  • prediction needs: emissions, combustion efficiency, …
slide-19
SLIDE 19

Methane Combustion: CH4 + 2 O2  CO2 + 2 H2O

300+ reactions, 50+ species

Foundation

  • A physically‐based model
  • The network is complex, but the governing

equations (rate laws) are known

  • Uncertainty exists, but much is known where

the uncertainty lies (rate parameters)

  • Numerical simulations with parameters fixed to

certain values may be performed “reliably”

  • There is an accumulating experimental

portfolio on the system

  • The model is reduced in size for applications

theory experiments numerical simulations

slide-20
SLIDE 20

flow‐reactor measurements theoretical rate constants laboratory flame measurements PREDICTION: ignition delay in HCCI engine

slide-21
SLIDE 21

Dataset unit U = ( U, L, M ) Dataset unit

U2 = (U2, L2, M2)

Dataset unit

U3 = (U3, L3, M3)

Dataset unit

U4 = (U4, L4, M4)

Dataset unit

U5 = (U5, L5, M5)

Dataset unit

Ue = ( Ue, Le, Me )

y1 y2 y3 x1 x2 x3

Dataset {Ue}

Model parameters

Feasible set of x, F

If empty, inconsistent, otherwise, consistent

Feasible set of x, F

If empty, inconsistent, otherwise, consistent

Experiment, E L  y  U Model, M(x) Dataset imposes constraints

 

,

e e e

L M U e    x

xe forms n‐dimensional hypercube

 

,min ,max

:

n i i i

x x x     x  

H D

slide-22
SLIDE 22

yupper ylower yexpt

M(x1,x2)

H

prior knowledge bounds on x1 feasible set

F

slide-23
SLIDE 23

a set of individual uncertainties does not represent the true compound uncertainty a realistic feasible set:

slide-24
SLIDE 24
  • build surrogate models for

individual responses y (rather than for overall objective)

  • construct global objective from

individual responses (higher fidelity)

model response

ODE Model

x1 x2  T P C  y parameters

 

 

2 2 2 1 2 1 1 1 2 1,2 1,1 2, 2 2

x x x y x x a a a a a a x x             surrogate model

 

2

min

computed

  • bserved

x all responses

y y w    

slide-25
SLIDE 25
  • build surrogate models for individual

responses (rather than for overall objective)

  • construct global objective from

individual responses (higher fidelity)

model response

ODE Model

x1 x2  T P C  y parameters

 

 

2 2 2 1 2 1 1 1 2 1,2 1,1 2, 2 2

x x x y x x a a a a a a x x             surrogate model

 

2

min

computed

  • bserved

x all responses

y y w    

0.2 0.4 0.6 0.8 1 45 2 17 11 3 9 58 1 29 33 47 4 73 82 5 6 98 …

|sensitivity| (×uncertainty)

active variables

dimensionality reduction

slide-26
SLIDE 26
  • build surrogate models for individual

responses (rather than for overall objective)

  • construct global objective from

individual responses (higher fidelity)

model response

ODE Model

x1 x2  T P C  y parameters

 

2

min

computed

  • bserved

x all responses

y y w    

dimensionality reduction

flame speed P2x1, x2, x7, x23, …

  • • •

ignition delay P2x1, x4, x5, x17, …

  • • •

species conc P2x3, x4, x7, x12, …

  • • •

dimensionality of individual response dimensionality of optimization

slide-27
SLIDE 27

Uncertainty is constrained by:

  • prior knowledge of parameters,

xH, the “H cube”

  • observed data/models,

M(x)D, the “D cube”

Prediction model: f(x) ‒establish possible range of f(x), constrained by

 

 

: max

x U L M x

r f r r x

 

  

 

 

 

: min

x U L M x

p f p p x

 

  

 

 

 

 

 

min max

x x M x M x

f x f x

   

       

    Hard‐to‐solve

  • ptimization

Computable bounds, easily verified as valid

Hard‐to‐solve

  • ptimization

Computable bounds, easily verified as valid

inner inner
  • uter
  • uter

If f and M are quadratic, then the min and max problems  SDP and p’s and r’s bounds are

  • computable
  • easily verified as valid
  • same for their global sensitivities
slide-28
SLIDE 28

 all parameters are within prior bounds, 

A dataset is consistent if the Feasible Set is nonempty; i.e., there exists a parameter vector that satisfies:

1,min 1 1,max 2,min 2 2,max

x x x x x x    

 

e e e

L M U   x

 all model predictions are within experimental bounds

x1 x2 x3 y1 y2 y3

H D

 numerical measure of consistency

     

1 1 ,

max

e e e

D H L M U e

C

 

     

x x

slide-29
SLIDE 29

Consistent Feasible Set Inconsistent Feasible Set

  • J. Phys. Chem. A 108: 9573 (2004)
slide-30
SLIDE 30

Wiesner et al. 1996 27 active variables Lemon et al. 2003 34 active variables inconsistent consistent consistency measure

+0.24 0.03

  • J. Phys. Chem. A 110: 6803 (2006)
slide-31
SLIDE 31

prediction para model mo eter del m             prediction parameter interval uncertainty             ex p p re er dict imen interval uncert io ainty t n            

slide-32
SLIDE 32

1 : 2 experiment uncertainty prediction interva M M M M L U L l U                                

max min max min

1 : 2 parameter uncertaint prediction interv M M M M x a x y x x l                                

slide-33
SLIDE 33
  • f

to

i j

Y Y predicti uncertainty in uncertainty in ob serving ng

Sensitivity

Yi Yj

slide-34
SLIDE 34
  • 0.1
0.0 0.1 0.2 0.3 H + O2 → O + OH OH + CO → H + CO2 CH3 + CH3 → H + C2H5 H + C2H4 + (M) → C2H5 + (M) H + CH3 + (M) → CH4 + (M) OH + CH3 → CH2* + H2O C2H5 + O2 → HO2 + C2H4 H + O2 + H2O → HO2 + H2O HCO + H2O → H + CO + H2O H + HCO → H2 + CO HO2 + CH3 → OH + CH3O O + H2 → H + OH H + OH + M → H2O + M H + HO2 → OH + OH

sensitivity

“w.r.t. value” vs “w.r.t. uncertainty” laminar flame speed in a stoichiometric atmospheric C2H6‐air mixture

  • J. Phys. Chem. A 112: 2579 (2008)

Lower

0.2 0.4 0.6 1 11 21 31 41 51 61 71

lower

0.001 0.002 1 11 21 31 41 51 61 71 81 91101

sensitivity of methane dataset consistency

to uncertainty in experimental observations to uncertainty in model parameters

upper bound lower bound upper bound lower bound

slide-35
SLIDE 35

Initial prediction from prior info Final prediction

Mp(x1,x2)

slide-36
SLIDE 36

   

p p

M M   x x

         

p p p p

max m : n , i

e e e e e

M L M M d U M M e          

x x

x x x x x x subject to 

feasible set

is the range of values Mp takes over the set of feasible values of parameters

prediction interval

slide-37
SLIDE 37

35th Combust. Symp. 2014

PREDICTION FEATURE PREDICTION INTERVAL O Peak Value

[2.7, 4.3] × 10‐2

O Peak Location

[1.9, 2.2] cm

OH Peak Value

[3.0, 3.6] × 10‐2

OH Peak Location

[1.60, 1.67] cm

C2H3 Peak Value

[0.09, 1.15] × 10‐4

C2H3 Peak Location

[0.6, 3.9 ] × 10‐2 cm

slide-38
SLIDE 38

Including Experimental Observations Only prior knowledge

prior knowledge feasible set

1

I 

 Posterior Range Prior Range

Information Gain:

  • Int. J. Chem. Kinet. 36: 57 (2004)
0.2 0.4 0.6 0.8 1 ig.1a ig.2 ig.6b ig.t2 ig.st1b ig.st3b ig.st4b ch3.c1b ch3.t1b ch3.t2 ch3.t3 ch3.t4 ch3.stc7
  • h.1b
  • h.3a
co.t1a co.c1b
  • h.3c
co.t1c co.c1d
  • h.st8
co.sc8 bco.t2 bco.t4 bco.t6 bch2o.t1 bch2o.t3 f1 f3 f6 stf8 sch.c11 sch.c13 nfr1 nfr3 nf7 nf12 nfr5
slide-39
SLIDE 39

   

arg min             

cur experim cur parame en t p ts t ers

  •  

 

: T C C                 

parameters expr c emen ur ts

subject to

Given a budget T, determine the best strategy for reducing the uncertainty in model prediction

1

a

C          

current

cost uncertainty, 

16 18 20 22

5 10 15 20 50 100 150 200 50 100 200 total cost range predicted for C2H6 flame speed (cm/s)
  • J. Phys. Chem. A 112: 2579 (2008)
slide-40
SLIDE 40

Slavinskaya, et al., 2013, 2014

flame speeds ignition delays

slide-41
SLIDE 41

Factor Compute gradients of f(x) at points of  Perform SVD of F; this gives S Sample r‐subspace of H to build surrogate design

 

 

T

f g S  x x

C2H2 flat flame (ALS) 12 responses, 51 active variables

Yeates et al. 7th US Combust. 2011

While a M(x) formally depends on all n active variables, in reality it mostly vary in r « n linear combinations of the variables. For creating a surrogate of M(x) we would like to do the design in the r‐dimensional space.

slide-42
SLIDE 42
  • Even in case of rigorous Bayesian, use of a prescribed prior (e.g., Gaussian) underestimates

the uncertainty in prediction (Phillip Stark, “Constraints versus Priors”, 2012) AND we unlikely to have Gaussian priors!

  • Approximations, even seemingly “harmless”,

may lead to substantial differences in prediction

  • f uncertainty (Russi et al, Chem. Phys. Lett. 2010)
  • Optimization‐based methods, transferring

uncertainty in two‐steps — from data to parameters and then from parameters to prediction — necessarily

  • verestimate the predicted uncertainty

Baulch et al. 2005

slide-43
SLIDE 43

An ongoing collaborative study with Jerome Sacks, National Institute of Statistical Sciences Rui Paulo, ISEG Technical University of Lisbon Gonzalo Garcia‐Donato, Universidad de Castilla‐La Mancha, Spain

B2B‐DC prediction for this blind target is [ 1.89 2.12] Bayesian

Example: GRI‐Mech 3.0

  • 102 active variables
  • 76 experimental targets
  • predicting one “blind”

Example: H2/O2

  • 21 active variables
  • 12 experimental targets
  • predicting one “blind”

Bayesian B2B sampling F

slide-44
SLIDE 44
  • current inability of truly predictive modeling

– conflicting data in/among sources – poor documentation of data/models – no uncertainty reporting or analysis – not much focus on integration of data

  • resistance to data sharing

– no personal incentives – no easy‐to‐use technology

  • no recognition of the problem
slide-45
SLIDE 45
  • is mathematically rigorous, numerically efficient, and UQ‐rich

approach to analysis of practical systems

  • is data‐centric, handles heterogeneous data, and is easily scalable to

a large number of data sets

  • is scalable to a large number of parameters through Solution

Mapping features, combined with the Active Space Discovery

  • establishes a clear measure of consistency among data and models,

and identifies the cause of inconsistency if detected

  • “measures” information content of an experiment

‒ assess an impact of a given or planned experiment (“what if”) ‒ design new experiments/theory that impact the most

  • reduces uncertainties of known and predicts correctly uncertainty of

unknown

slide-46
SLIDE 46
slide-47
SLIDE 47
slide-48
SLIDE 48
  • What causes/skews model predictiveness?
  • Are there new experiments to be performed, old

repeated, theoretical studies to be carried out?

  • What impact could a planned experiment have?
  • What is the information content of the data?
  • What would it take to bring a given model to a

desired level of accuracy?

slide-49
SLIDE 49

from algorithm-centric view to data-centric view

  • utput

input

code

data data

slide-50
SLIDE 50

 

2 i i

data model  

k2 k1

deviations

slide-51
SLIDE 51 (top) Cool et al., Rev. Sci. Instrum. 76, 094102 (2005) (bottom) Courtesy of Sandia CRF ‐ http://www.sandia.gov/ERN/images/CRF‐Science.jpg

Integrated signal

slide-52
SLIDE 52

Concentration profile

Data Analysis y = f(x,c) Calibration, c Model Assumptions Integrated signal

slide-53
SLIDE 53

Concentration profile

Data Analysis y = f(x,c) Calibration, c Model Assumptions Integrated signal

slide-54
SLIDE 54

Data Analysis y = f(x,c) Calibration, c Model Integrated signal Instrumental Model

slide-55
SLIDE 55

O, OH, C2H3

– Peak Value – Peak Location

PREDICTION FEATURE PREDICTION INTERVAL O Peak Value

[2.73, 4.29] × 10‐2

O Peak Location

[1.87, 2.20] cm

OH Peak Value

[2.97, 3.59] × 10‐2

OH Peak Location

[1.60, 1.67] cm

C2H3 Peak Value

[0.09, 1.15] × 10‐4

C2H3 Peak Location

[0.60, 3.90] × 10‐2 cm

slide-56
SLIDE 56

hypothesis collection

  • f data

model prediction hypothesis collection

  • f data

model

  • experiments
  • theory
slide-57
SLIDE 57

hypothesis collection

  • f data

model prediction hypothesis collection

  • f data

model

SENSITIVITY ANALYSIS:

What conditions will maximize sens to k?

slide-58
SLIDE 58

hypothesis collection

  • f data

model prediction hypothesis collection

  • f data

model

UNCERTAINTY QUANTIFICATION:

What experiment will be most informative?

slide-59
SLIDE 59

?    

flame code

input conditions reaction model validation data thermo and transport data prediction

slide-60
SLIDE 60

Volume of sphere Volume of cube

dimension

slide-61
SLIDE 61

 

p

M x

 

p

M x

 

p

M x

slide-62
SLIDE 62

Data analysis performed

in isolation

leads to loss of information

slide-63
SLIDE 63

hypothesis|data data|hypothesis hypothesis

 

slide-64
SLIDE 64

prior likelihood posterior

hypothesis|data data|hypothesis hypothesis

  model / analysis

slide-65
SLIDE 65

prior likelihood posterior

hypothesis|data data|hypothesis hypothesis

  model / analysis

slide-66
SLIDE 66
  • Quadratic surrogates enable mathematically rigorous, numerically efficient,

and UQ‐rich approach of B2B‐DC to practical systems

  • Quadratics work in practice because model parameters are limited

‒ by physical constraints; e.g., 0 < k < collision limit ‒ by reaction theory / chemical analogy ‒ by prior experimental / theoretical studies ‒ and can be linearized; e.g., by the log transformation

  • And if they do not work, then

‒ rational quadratics (“native” with B2B framework) ‒ a two‐level surrogate approach

 first, use machine learning to build “high‐order surrogates”, e.g., Gaussian Process, Kriging, ‐SVM, Polynomial Chaos  then, build/use on‐demand piece‐wise quadratics from the high‐order surrogates

  • J. Phys. Chem. A 110: 6803 (2006)
slide-67
SLIDE 67

L  y  U y2 y3

D

y1 L ‒   M(x)  U +  y = M(x) +  L′  M(x)  U′

surrogate model fitting error

D