Quantification of Uncertainty in Extreme Scale Computations (QUEST) - - PowerPoint PPT Presentation

quantification of uncertainty in extreme scale
SMART_READER_LITE
LIVE PREVIEW

Quantification of Uncertainty in Extreme Scale Computations (QUEST) - - PowerPoint PPT Presentation

Intro QUEST Technical Progress Closure Quantification of Uncertainty in Extreme Scale Computations (QUEST) www.quest-scidac.org H. Najm 1 , B. Debusschere 1 , M. Eldred 1 , R. Ghanem 2 , O. Ghattas 3 , R. Moser 3 , E. Prudencio 3 , D. Higdon 4


slide-1
SLIDE 1

Intro QUEST Technical Progress Closure

Quantification of Uncertainty in Extreme Scale Computations (QUEST)

www.quest-scidac.org

  • H. Najm1, B. Debusschere1, M. Eldred1, R. Ghanem2,
  • O. Ghattas3, R. Moser3, E. Prudencio3, D. Higdon4,
  • J. Gattiker4, O. Knio5, Y. Marzouk6

1Sandia National Laboratories, Livermore, CA & Albuquerque, NM 2University of Southern California, Los Angeles, CA 3University of Texas, Austin, TX 4Los Alamos National Laboratory, Los Alamos, NM 5Duke University, Durham, NC 6Massachusetts Institute of Technology, Cambridge, MA

SciDAC PI Meeting, 10–12 Sep 2012, Rockville, MD

SNL Najm QUEST 1 / 28

slide-2
SLIDE 2

Intro QUEST Technical Progress Closure

Outline

1

Introduction

2

QUEST Overview

3

Technical Progress

4

Closure

SNL Najm QUEST 2 / 28

slide-3
SLIDE 3

Intro QUEST Technical Progress Closure

Introduction – Motivation

Why Uncertainty Quantification (UQ) ? Assessment of confidence in computational predictions Validation and comparison of scientific/engineering models Design optimization Use of computational predictions for decision-support Assimilation of observational data and model construction Why UQ in SciDAC ? Explore model response over range of parameter variation Enhanced understanding extracted from computations Particularly important given cost of SciDAC computations

SNL Najm QUEST 3 / 28

slide-4
SLIDE 4

Intro QUEST Technical Progress Closure

QUEST Goals

1

Advance the state of the art in UQ theory, methods, and software, addressing UQ challenges with extreme scale computational problems

High-dimensionality Nonlinearity Sparse data

2

Provide expertise, advice, and state of the art UQ algorithms and software tools to SciDAC projects

UQ software products SciDAC partnerships Outreach: UQ tutorials, summer school, web

SNL Najm QUEST 4 / 28

slide-5
SLIDE 5

Intro QUEST Technical Progress Closure

Scope

The scope of QUEST covers a range of UQ activities including: UQ problem setup Characterization of the input space Local and global sensitivity analysis Adaptive stochastic dimensionality and order reduction Forward and Inverse UQ Fault tolerant UQ methods Model comparison and validation

SNL Najm QUEST 5 / 28

slide-6
SLIDE 6

Intro QUEST Technical Progress Closure

Key Elements of our UQ strategy

Probabilistic framework

Uncertainty is represented using probability theory

Parameter Estimation, Model Calibration

Experimental measurements Regression, Bayesian Inference

Forward propagation of uncertainty

Polynomial Chaos (PC) Stochastic Galerkin methods – Intrusive/non-intrusive Stochastic Collocation methods

Model comparison, selection, and validation Model averaging Experimental design and uncertainty management

SNL Najm QUEST 6 / 28

slide-7
SLIDE 7

Intro QUEST Technical Progress Closure

Team Expertise and Capabilities

Institution Expertise Tools SNL Forward and inverse UQ methods, DAKOTA design under uncertainty UQTK USC Intrusive UQ methods probabilistic modeling Duke Sparse adaptive forward UQ methods UT Large scale inverse problems QUESO validation, inverse UQ LANL Gaussian process modeling, inverse UQ GPMSA MIT Calibration, adaptive sampling, inverse UQ, experimental design

SNL Najm QUEST 7 / 28

slide-8
SLIDE 8

Intro QUEST Technical Progress Closure

QUEST UQ Software tools

DAKOTA

Optimization and calibration Non-intrusive UQ Global Sensitivity Analysis > 10K registered downloads

QUESO

Bayesian Inference Parallel MultiChain MCMC Bayesian Model Analysis Model Calibration

GPMSA

Bayesian Inference Gaussian Process Emulation Model Calibration Model discrepancy analysis

UQTk

Intrusive PC UQ Non-intrusive sampling Customized sparse PCE Random fields

SNL Najm QUEST 8 / 28

slide-9
SLIDE 9

Intro QUEST Technical Progress Closure

QUEST Partnerships

DOE

Project Title

Lead PI QUEST NNSA

Parallel Dislocation Simulator

  • T. Arsenlis

Najm

(ParaDiS)

LLNL SNL FES

Center for Edge Plasma Physics

C.S. Chang Moser

Simulation (EPSI)

Princeton UT FES

Plasma Surface Interactions: Bridging

  • B. Wirth

Higdon

from the Surface to the Micron Frontier

ORNL LANL BER

Predicting Ice Sheet & Climate Evolution

P . Jones Eldred, Ghattas

at Extreme Scales (PISCEES)

LANL SNL, UT BER

Multiscale Methods for Accurate, Efficient

  • B. Collins

Debusschere

& Scale-Aware Earth System Modeling

LBNL SNL NP

Nuclear Computational Low Energy

  • J. Carlson

Higdon

Initiative (NUCLEI)

LANL LANL HEP

Computation-Driven Discovery

  • S. Habib

Higdon

for the Dark Universe

ANL LANL HEP

Community Project for Accelerator

P . Spentzouris Prudencio

Science & Simulation (ComPASS)

FNAL UT

SNL Najm QUEST 9 / 28

slide-10
SLIDE 10

Intro QUEST Technical Progress Closure

Outreach Activities

Website

www.quest-scidac.org Production version will be publicly accessible soon

UQ Tutorials in workshops/conferences

SAMSI UQ workshop, Raleigh, NC; Sep 7-9, 2011 SIAM Conference on UQ, Raleigh, NC; Apr 2-5, 2012

UQ Summer School

USC, LA; Aug 22-24, 2012

UQ Tools Tutorial

Hands-on practice with UQ software tools SNL, Livermore, CA; Oct 22-23, 2012. Announcements went out in late July http://cadmus.usc.edu/Quest-Tutorial – Some openings still available

SNL Najm QUEST 10/ 28

slide-11
SLIDE 11

Intro QUEST Technical Progress Closure SNL LANL UT Duke MIT USC

SNL: Software: DAKOTA – dakota.sandia.gov

  • M. Eldred, J. Jakeman

Development of interfaces: QUESO–DAKOTA–GPMSA

Ongoing DAKOTA interfaces to both C++ GPMSA implementation using QUESO components

Stochastic collocation

Nodal or hierarchical interpolation on structured grids Interpolants may be local or global – value-based or gradient-enhanced Automated refinement – uniform, dimension-adaptive, or locally-adaptive Hierarchical surplus error estimates for values and gradients applied to QoI (e.g., response covariance)

Compressive sensing: basis pursuit and basis denoising

SNL Najm QUEST 11/ 28

slide-12
SLIDE 12

Intro QUEST Technical Progress Closure SNL LANL UT Duke MIT USC

DAKOTA: Application in Nuclear Reactor Modeling

  • M. Eldred

Work with CASL energy innovation hub PCE/SC with uniform/adaptive refinement vs LHS

n = 4, smooth, mild anisotropy

10

1

10

2

10

3

10

4

10

5

10

−4

10

−3

10

−2

10

−1

Evaluations Change in σ for MEmean LHS PCE uniform SC uniform PCE adaptive SC adaptive

n = 10, discontinuous, high anisotropy

10

2

10

3

10

4

10

5

10

−2

10

−1

10 Evaluations Change in σ for MEmean LHS PCE uniform SC uniform PCE adaptive SC adaptive

SNL Najm QUEST 12/ 28

slide-13
SLIDE 13

Intro QUEST Technical Progress Closure SNL LANL UT Duke MIT USC

SNL: Software: UQTk – www.sandia.gov/UQToolkit

  • B. Debusschere, C. Safta, K. Sargsyan

Version 1.0 published under the GNU LGPL

Intrusive PC functionality

New release targeted for Fall 2012

Intrusive and non-intrusive utilities User-specified multi-index capabilities

Flexible efficient sparse tensor representations Effective for high-dimensional systems

Random fields:

Covariance matrix estimation (many samples) Karhunen-Loève expansions (KLEs)

Matlab version

Example/benchmark problems, tutorial materials

SNL Najm QUEST 13/ 28

slide-14
SLIDE 14

Intro QUEST Technical Progress Closure SNL LANL UT Duke MIT USC

SNL: Algorithms: Gradients & Sparsity

  • M. Eldred, J. Jakeman, K. Sargsyan, C. Safta, B. Debusschere, H. Najm

Hierarchical interpolation with generalized sparse grids

Gradient-enhancement Error indicators leverage both value and gradient surpluses

Building Sparse PC representations

Compressed Sensing (CS) – ℓ1 regularization – cross validation, tolerances for model choice Bayesian Compressed Sensing (BCS) – Laplace priors BCS/CS comparisons on Genz functions – 5-10d – Similar convergence with no. of samples – Slightly higher accuracy with CS – BCS: O(100)× reduction in no. of PCE terms

discovery of sparse signals:

SNL Najm QUEST 14/ 28

slide-15
SLIDE 15

Intro QUEST Technical Progress Closure SNL LANL UT Duke MIT USC

SNL: Algorithms: Missing Data

  • H. Najm, B. Debusschere, C. Safta, K. Sargsyan, K. Chowdhary

Context Missing/failed measurements or computational samples Partial specification of uncertain information

Error bars vs. joint PDF

Processed data products Imputation methods Existing data ⇒ probabilistic prediction of missing data Data Free Inference (DFI) algorithm Given information ⇒ probabilistic models of missing data – Application in chemical ignition – Extension to processed data products

SNL Najm QUEST 15/ 28

slide-16
SLIDE 16

Intro QUEST Technical Progress Closure SNL LANL UT Duke MIT USC

LANL: GPMSA & BART Developments

  • D. Higdon, J. Gattiker

New release of GPMSA for sensitivity analysis and computer model calibration using Bayesian methods

Tutorial material Range of sample problems – sensitivity, calibration, & multivariate output

Prototype parallel implementation of the Bayesian additive regression tree (BART) for HPC.

linear scaling up to ∼50p tests with higher proc counts in progress

!" #" $" %" "&" "&# "&% "&' "&( )*+,-./01/2.03-440.4 ! .*)56+- (70*.4−!)

SNL Najm QUEST 16/ 28

slide-17
SLIDE 17

Intro QUEST Technical Progress Closure SNL LANL UT Duke MIT USC

UT-Austin: Scalable Parallel Algorithms for Extreme-Scale Stochastic Inverse Problems

  • T. Bui-Thanh, O. Ghattas, J. Martin, G. Stadler (also funded by AFOSR and NSF)

Stochastic Inverse Probs:

PDEs & high-dim parameter spaces (from discretized fields) Current methods are prohibitive Challenges: appropriate choice of prior consistent discretizations (guarantee convergence to infinite-dim problem) scalable parallel MCMC algorithms

Recent accomplishments:

Consistent discretizations via appropriate mass matrix weightings Prior defined by inverse of elliptic

  • perator; carried out by multigrid

Low rank approximation of Hessian enables sampling of Gaussianized posterior in dimension-independent number of forward solves Scaling to 1M parameters and 100K processor cores

SNL Najm QUEST 17/ 28

slide-18
SLIDE 18

Intro QUEST Technical Progress Closure SNL LANL UT Duke MIT USC

Example: Extreme-scale Seismic Inversion

  • T. Bui-Thanh, O. Ghattas, J. Martin, G. Stadler

linearized 3D global seismic inversion 1.07M earth model parameters 630M wave propagation unknowns 100K cores on Jaguar (ORNL) 2000× reduction in effective problem dimension due to low rank approx Top row: Prior samples Bottom row: Posterior samples Difference between rows indicates information gained from (and uncertainty reduced due to) data Gordon Bell Prize finalist, SC12

SNL Najm QUEST 18/ 28

slide-19
SLIDE 19

Intro QUEST Technical Progress Closure SNL LANL UT Duke MIT USC

UT-Austin: Software: QUESO

  • K. C. Estacio-Hiroms, E. E. Prudencio, K. W. Schulz (also funded by NNSA)

Improvement of QUESO-DAKOTA usability

Periodic output of samples Output of extra information Informative output summary

Implementation of GPMSA models and algorithms

QUESO capabilities will be usable through DAKOTA

Preparation of tutorial material

Bayesian inversion, and forward propagation of uncertainty Object-oriented mapping of mathematical concepts Solution of statistical inverse problems with DRAM MCMC Solution of statistical forward problems with Monte Carlo Use of parallel computing for statistical analysis References to Bayesian analysis, MCMC, Monte Carlo, C/C++, MPI

SNL Najm QUEST 19/ 28

slide-20
SLIDE 20

Intro QUEST Technical Progress Closure SNL LANL UT Duke MIT USC

Duke: Stochastic Preconditioning

  • O. Knio, A. Alexanderian, O. Le Maître

Developed a multiscale Bayesian preconditioning approach Demonstrated capability to simultaneously

address stiffness and noise represent noisy outputs w/sparse, low-order, PCEs

Order of magnitude reduction in # of samples / replicas

! "! #! $! %! &! '! ! #! %! '! (! "!! )*+,-./012/-*134-5/,3647/+ 372/ ˆ µ(t, ·) ! !"# $ $"# % %"# ! !"% !"& !"' !"( $ )*+,-./0-+1/23+4-*2536-7 τ ˆ µY (t, ·)

SNL Najm QUEST 20/ 28

slide-21
SLIDE 21

Intro QUEST Technical Progress Closure SNL LANL UT Duke MIT USC

Duke-MIT: Sparse Adaptive Sampling

  • J. Winokur, P

. Conrad, O. Knio, Y. Marzouk

Developed a sparse adaptive pseudospectral sampling algorithm

accommodates arbitrary admissible stencils including a maximal polynomial basis – without internal aliasing

Analysis of algorithm performance based on existing Ocean General Circulation Model (OGCM) databases Demonstrated order-of-magnitude computational savings in simulations of the ocean circulation in the Pacific

SNL Najm QUEST 21/ 28

slide-22
SLIDE 22

Intro QUEST Technical Progress Closure SNL LANL UT Duke MIT USC

MIT: Large-Scale Bayesian Inference

  • T. Moselhy, Y. Marzouk

Current state of the art Markov chain Monte Carlo (MCMC) sampling is the workhorse algorithm for Bayesian inference and prediction Challenges: enormous computational effort, difficult proposal design, insufficient convergence diagnostics Inference with optimal maps New approach: find a deterministic map that pushes forward the prior measure to the posterior measure Converts inference to an optimization problem, with natural convergence diagnostics Outperforms MCMC in efficiency and accuracy on a variety

  • f inference problems, with 103 dimensions or more

SNL Najm QUEST 22/ 28

slide-23
SLIDE 23

Intro QUEST Technical Progress Closure SNL LANL UT Duke MIT USC

MIT: Large-Scale Bayesian Inference

  • T. Moselhy, Y. Marzouk

!! !" # " ! !#$% !#$& !#$! !#$" # #$" #$! '!

(

'" !! !" # " ! !#$% !#$& !#$! !#$" # #$" #$! '!

(

'" !! !" # " ! !#$% !#$& !#$! !#$" # #$" #$! '!

(

'"

(above) sequence of maps yields samples from non-Gaussian posterior in a chemical kinetic system Current work on map-based inference:

Hierarchical Bayesian models Parallel algorithms for stochastic optimization Sequential data assimilation (i.e., filtering and smoothing)

SNL Najm QUEST 23/ 28

slide-24
SLIDE 24

Intro QUEST Technical Progress Closure SNL LANL UT Duke MIT USC

MIT: Optimal Experimental Design

  • X. Huan, Y. Marzouk

How to choose observations or experimental conditions

  • ptimally?

Bayesian approach: maximize expected information gain for parameter inference, prediction, model discrimination, etc

Key computational ingredients:

Surrogates for physical model describing experiments Statistical estimators and stochastic optimization methods

Recent accomplishments: stochastic approximation and sample-average approximation for optimal Bayesian design, using estimators of mutual information gradient

0.5 1 0.5 1 0.5 1 1.5 2 x y Expected Utility 0.8 0.9 1 1.1 1.2 1.3 1.4 1.5 0.5 1 0.5 1 100 200

SNL Najm QUEST 24/ 28

slide-25
SLIDE 25

Intro QUEST Technical Progress Closure SNL LANL UT Duke MIT USC

USC: Constrained & Adaptive Constructions

  • E. Kalligiannaki, R. Tipireddy, G. Ghanem

Develop Constrained Stochastic Representations Positive random variables More general constraints on either function values or values of nonlinear functionals of the random variables Develop Bases Adapted to Quantity of Interest Scales linearly with stochastic dimension

SNL Najm QUEST 25/ 28

slide-26
SLIDE 26

Intro QUEST Technical Progress Closure SNL LANL UT Duke MIT USC

USC: Constrained Stochastic Representations

  • E. Kalligiannaki, G. Ghanem

I = {y(ω) ∈ L2(Ω, Σ(H), P) : y(w) satisfies constraints ∀ω} The projection of y ∈ L2 on I: Sample from prior PC expansion Delete realizations that do not satisfy constraints Recompute PC coefficients from remaining realizations

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time t Solution u(t)

Initial data, u(t=0) = U=0.2 + U1 ξ

U1=0 U1=0.08, Ns=104, Nit=100 U1=0.08, Ns=104, Nit=150 U1=0.08, Ns=104, Nit=200 U1=0.08, Ns=104, Nit=10 U1=0.08, Ns=104, Nit=0

Improve Convergence of Stochastic ODE Generator for constrained populations

SNL Najm QUEST 26/ 28

slide-27
SLIDE 27

Intro QUEST Technical Progress Closure SNL LANL UT Duke MIT USC

USC: Adaptation to Quantity of Interest (QoI)

  • R. Tipireddy, G. Ghanem

Expand u(ξ): polynomials in η = Aξ with proper choice of A, the measure of the solution is concentrated along leading η1 dimension A is chosen so that η1 contains all Gaussian content of QoI

! !"!# !"!$ !"!% !"!& !"' !"'# !"'$ ! ( '! '( #! #( )! )( *+,-./0121345+35657+8104+935/45:!"#(;5!"!!< =89>/>+.+4?5713,+4?5@A304+93

5 5

BA..!*+213,+93 C17A0175'*5D971.

!"#$ !"% !"%# !"%& !"%' !"%$ !"& ( )! )( #! #( %!

*+,-./010-023100-42-5!"##&(6!"))!)7 !"#

  • 89::!;40/0

31<9=1<!;40/0

SNL Najm QUEST 27/ 28

slide-28
SLIDE 28

Intro QUEST Technical Progress Closure

Closure

Work on UQ software and algorithms development

Computational efficiency Functionality, usability, scalability Adaptivity, sparsity, preconditioning Reduced-order, low-rank Convergence, stability Partial information, missing data

Robustifying algorithms for large-scale applications Software integration well along the way Outreach via web, tutorials, and summer school SciDAC partnership activities getting off the ground

SNL Najm QUEST 28/ 28