Tutorial on The Design of Computer Experiments for Optimization - - PowerPoint PPT Presentation

tutorial on the design of computer experiments for
SMART_READER_LITE
LIVE PREVIEW

Tutorial on The Design of Computer Experiments for Optimization - - PowerPoint PPT Presentation

Tutorial on The Design of Computer Experiments for Optimization Thomas Santner Department of Statistics The Ohio State University The A-C-NW Optimization Tutorials June 13, 2005 Outline 1. Three Types of Experiments # What are Computer


slide-1
SLIDE 1

Tutorial on

The Design of Computer Experiments for Optimization Thomas Santner Department of Statistics The Ohio State University The A-C-NW Optimization Tutorials June 13, 2005

slide-2
SLIDE 2

Outline

  • 1. Three Types of Experiments

#Þ What are Computer Experiments? $Þ An Example %Þ Nomenclature and a Taxonomy of Problems for Computer Exeriments

  • 5. Sequential Design of Computer Experiments for Global Optimization
  • 6. Take Home Points

Some References

 %& 

slide-3
SLIDE 3
  • 1. Three Types of Experiments
  • 1. Physical Experiments
  • Gold standard for establishing cause and effect relationships
  • Mainstay of Engineering, Agriculture, Medicine
  • Principles of randomization, blocking, choice of sample size, and

stochastic modeling of response variables all developed in response to needs of physical experiments 2. Complex physical system each of whose parts behave Simulation Experiments stochastically and interact in a known manner but whose ensemble stochastic behavior is not understood analytically

  • Used extensively in IE/OR--compare hospital emergency room setups

3. past 15 years Computer Experiments

 %% 

slide-4
SLIDE 4

#. What are Computer Experiments? Idea: Many physical processes can not be studied by conventional experiments Why?

(1) technically, the physical process is too difficult (expensive) to study experimentally (2) ethical considerations (3) number of input variables is too large

If either (1) the components of the process of interest and their interactions are adequately understood so it can simulated (with negligable MC error) or (2) the physics of the process of interest is

  • sufficiently well understood so it can be described by a mathematical model

relating the to the that affect the output inputs, response potential factors

  • Numerical methods exist for solving the mathematical model
  • The numerical methods can be implemented with computer code

Then proxy the computer code can serve as a for the physical process. As in a physical experiment, B − § B k ‘ Ò Ò

.

Code CÐ Ñ

 %$ 

slide-5
SLIDE 5

B − § B k ‘ Ò Ò

.

Code CÐ Ñ Features of Computer Experiments

  • CÐ Ñ

B is deterministic

  • Our interest is in settings where very few computer runs are possible due to
  • 1. Complex codes
  • 2. High--dimensional input B
  • Traditional principles used in designing physical experimentals

(eg randomization, blocking, ) are irrelevant. á

 %# 

slide-6
SLIDE 6

$Þ Examples (1) Design of VLSI circuits (2) Modeling weather or climate (3) Design of automobile (components) (4) Determine the performance of controlled nuclear fusion devices (5) Temporal evolution of contained and wild fires (6) Design of helicopter rotor blades (7) Biomechanics Design of prosthetic devices 

 %" 

slide-7
SLIDE 7

Example (7) Designing hip and knee implants

 $( 

slide-8
SLIDE 8

Biomechanics I Design a hip implant

Goal To determine the design of a hip implant, i.e, that minimizes Ð,ß .Ñ femoral stress shielding while providing adequate resistance to implant toggling.

 $' 

slide-9
SLIDE 9

Inputs 1. (manufacturing design) Engineering Variables

  • Prosthesis geometry (length, cross-section, width, etc)
  • Prosthesis material
  • Nominal insertion parameters

2. & Environmental Variables (Patient Surgical Variables)

  • Bone material properties, weight (and other patient variables)
  • Deviation from nominal insertion parameters

(and other surgical variables) In the above figure, the are Engineering Variables

  • bullet-tip length

, œ

  • midstem diameter

. œ the are Environmental Variables

  • trabecular bone elastic modulus

I œ

  • joint force angle

) œ

  • Implant-bone interface friction

0 œ

 $& 

slide-10
SLIDE 10

Computed Response constructed from:

  • ( , ,

, ) normalized measure of bone stress shielding W œ W , . I ß 0 œ )

  • ( , ,

, ) normalized measure of implant toggling H œ H , . I ß 0 œ ) (competing objectives!!) Formulation #1 Combine & because they represent competing W H

  • bjectives. Goal is to minimize

C , . I ß 0 œ AW , . I ß 0  Ð"  AÑH , . I ß 0 ( , , , ) ( , , , ) ( , , , ) ) ) ) where measures the relative importance of the two objectives. A Formulation #2 minimize Goal is to W , . I ß 0 ( , , , ) ) subject to H , . I ß 0 Ÿ F ( , , , ) ) where is a given bound (a constrained optimization problem) F

 $% 

slide-11
SLIDE 11

Biomechanics II A Computer Model of a Knee Simulator Three data sources

  • 1. Knee Simulator (a machine)
  • 2. Computer code that emulates the Knee Simulator

Inputs (7-10)

  • Loading pattern (Flexion angle, Axial Force, AP Force, IE Torque)
  • Knee design (stem lengths, constrained or not, etc)
  • Frequency with which the loading pattern is applied (running/walking)
  • Elastic modulus of the polyethylene in the tibial tray
  • Polyethylene Irradiated or not
  • Friction between knee and femoral component
  • Surface type (Elemental vs Analytic

in finite element code) 

  • Mesh Density in finite element code
  •  $$ 
slide-12
SLIDE 12

Mathematical Model (finite element model 12 hours/per run)

 $# 

slide-13
SLIDE 13

Output data from Knee simulator & computer code

G ait APDisp 100 80 60 40 20

  • 1
  • 2
  • 3
  • 4
  • 5
  • 6

APDisp (Case 1): True and Predicted vs Gait

APDisp True (red dashed line), APDisp Predicted (black solid line)

and

 $" 

slide-14
SLIDE 14

G ait IERot 100 80 60 40 20 6 4 2

  • 2
  • 4

I ERot (Case1): True and Predicted vs Gait

IERot True (red dashed line), IERot Predicted (black solid line)

Project Goals

  • 1. "Calibrate" computer code to mimic knee simulator

Use calibrated computer code to produce effects seen in retreived 2. knees Explain the biomechanics of prosthetic joint failure 3.

 $! 

slide-15
SLIDE 15

%. Nomenclature and a Taxonomy of Problems for Computer Exeriments Inputs B B B = ( , ) where

  • /

B œ B

  • Engineering Design variables (each choice of

is )

c

Engineering Design B/ œ Environmental Variables (field, noise) , eg, patient bone densities. Philosophy We often regard as random variables with a environmental variables distribution that represents target field conditions, i.e., F \/ µ Ð Ñ

ñ

Outputs Real-valued : CÐñÑ

  • r

Multivariate:

  • C ÐñÑß C ÐñÑß á ß C ÐñÑ

" # 5

  • r

Functional: (>ß CÐ>ß ñÑÑ

  • Multivariate data: single or multiple codes, e.g., code computes

and CÐñÑ all first partial derivatives of CÐñÑ Ñ

  • Functional Data: APD or IER gait profile

 #* 

slide-16
SLIDE 16

Special Features of (Biomechanics) Problems

  • Codes for

are often CÐ Ñ B long-running

  • Sometimes associated

are available with output, . Physical Experiments ] ÐñÑ

P

Usual philosophy is that is a measurements of the true input-output ] : B noisy relationship, which we denote . In detail, .X Ð Ñ B ] œ Ð Ñ  Ð Ñ

: X

B B B . % where the are independent measurement errors having mean zero and e f %Ð Ñ B

B

unknown variance and we regard as the i-o relationship. 5 .

# X %

Ð Ñ B true, unknown Caveats Sometimes only physical experiments are available for components

  • f the ensemble process -- nuclear reactor simulator, code that emulates auto crash
  • test. In other cases, only experiments that

reality are available--knee approximate simulator

  • When there are field variables,

has a distribution. We might typically CÐ Ñ B \

c, /

be interested in one of several summary quantities associated with the distribution of CÐ Ñ B \

  • ,

/ . For example,

(that the quantity around which the computer  œ I CÐ Ñ .

  • e

f B B \

  • ,

/

  • utput varies)

= Var (one measure of the very ability of the computer  ÐCÐ ÑÑ 5#

/

  • B

B \

  • ,
  • utput due to variation in field inputs)
slide-17
SLIDE 17

Taxonomy of Problems

  • Given computer model output at a set of inputs (training data),

Interpolation/Emulation  predict the computer simulation output at a new, untried input settings

  • Determine input settings in which to carry out the sequence of

Experimental design  simulation designs (a "good" design of a physical or computer experiment depends on the scientific objective

  • f the research)

Exploratory Designs ("space-filling")  Prediction-based Designs   Optimization-based Designs (e.g., find ) B œ CÐBÑ

argmin

  • Determine the distribution of the computer model output

Uncertainty/Output Analysis  when (some or all of) the inputs are random, i.e., determine the distribution of . CÐ \ Ñ B.,

/

Examples of randomly varying inputs are patient specific variables (patient weight or patient bone material properties) or surgeon specific variables (measuring surgical skill) Example In his Cornell PhD thesis, Kevin Ong studied the effect of Surgical, Patient, and Fluid Effects on the Stability of Uncemented Acetabular Components

  • Determine how variation in

can be apportioned to the Sensitivity Analysis 

CÐ Ñ B

different computer model inputs B ( Inputs that have relatively little effect on the output can be set to some nominal Philosophy value and additional investigation restricted to determining how the output depends on the active inputs)  #( 

slide-18
SLIDE 18
  • use physical experimental data and computer simulation runs to best estimate

Calibration  the computer code calibration variable (or to update the uncertainty regarding these parameters) Example Set Mesh Density ? Load Discretization ? etc œ ß œ ß

  • Using the calibrated simulator to give predictions (with uncertainty bounds)

Prediction  for an associated physical system.

  • In experiments with engineering design and patient-specific

Find Robust Inputs  ß enviromental variables, determine robust choices of the engineering design variables. If .

  • e

f B B \

c c

œ I CÐ Ñ

J /

, then a robust set of inputs is an engineering "design" whose output is Bc minimally sensitive to uncertainty in the distribution

  • of

JÐ Ñ \/ Many of the problems above have "natural" solutions obtained by approximating the computer model by an even "faster" predictor, a . Statistical issues choosing the best metamodel possible surrogate for the code and devising valid methods to accomplish calibration etc.

 #' 

slide-19
SLIDE 19
  • 5. Sequential Design of Computer Experiments for Global Optimization

Recall B − § B k ‘ Ò Ò

.

Code CÐ Ñ

  • Minimizing the number of function evaluations (computer runs) is critical in many

computer experiments (Many methods, eg, direct search algorithms such as Nelder Mead "simplex" algorithm, or gradient-based algorithms can require "too many" function evaluations to be useful in the computer experiment settings)

  • Some specific optimization problems

find Case 1 B B

9:> ´

CÐ Ñ argmin Suppose where Case 2 B œ ÐB ß B Ñ

  • /

( ) B œ

  • control manufacturing, engineering design variables

( , ) B œ

/

noise field enviromental variables X and

  • ("target field conditions'').

is a random variable

/ µ JÐ Ñ

CÐ Ñ B ß \

  • /

with distribution induced by Find \/. B œ B

  • 9:>

J ,

argmin . Ð Ñ

  • where .J

J

Ð Ñ I CÐ Ñ B œ B ß \

  • /

e f Suppose output is Find Case 3

  • C ÐñÑ C ÐñÑß C ÐñÑß á ß C ÐñÑ

, .

" # 5

B ´ C Ð Ñ

argmin

0 B

subject to C ÐñÑ Ÿ F Ð Ÿ 3 Ÿ 5Ñ

3 3 1

 #& 

slide-20
SLIDE 20

Goal Describe Efficient Global Optimization EGO Algorithms Ð Ñ Idea EGO is a that uses a predictor, , to explore direct search algorithm CÐ Ñ CÐ Ñ B B s CÐ Ñ B surface and also in the predictor accounts for uncertainty Part 1 The Predictor! CÐBÑ Problem Given training data ˆ ‰ ˆ ‰ B B B B

" 8 8 " >< >< >< ><

ß CÐ Ñ ß á ß ß CÐ Ñ predict C ß

  • B

B

! !

where is an untried new input Idea Regard the function as a realization, a "draw " from a random function CÐ Ñ ß B ] B .

Philosophically different than regression, MARS, and other prediction methods that assume (complicated mean structure + simple correlation structure) and versus the methodology (below) which is based the assumptions of (simple mean + complicated correlation structure).

 #% 

slide-21
SLIDE 21

Issues

  • What model for

? One that permits great flexibility in the form of ] CÐ Ñx B B

  • The simplest possible model for

is ] B ] Ð Ñ œ 0 Ð Ñ  ^ B B B ðóóñóóò

  • î

4 4 4

" " " " " large scale trends smooth deviations œ 0Ð Ñ  ^ "ð B B where , are known regression functions, 0 Ð Ñ á ß 0 Ð Ñ

" 5

B B is an unknown regression vector, and " is a "stationary Gaussian Stochastic Process" (GSP), a random function ^ B  #$ 

slide-22
SLIDE 22

Example Four draws from a zero mean, unit variance GSP from [0,1] with the ^ÐBÑ î ‘ ''Gaussian'' correlation function VÐ2Ñ œ /B:  2 ˆ ‰ )

#

for 0.5 (solid lines), 1.0 (dotted lines), and 10.0 (dashed lines) ) ) ) œ œ œ

 ## 

slide-23
SLIDE 23

 #" 

slide-24
SLIDE 24

 #! 

slide-25
SLIDE 25
  • Smoothness properties of

depend on smoothness of at the origin CÐ Ñ VÐ Ñ ß /Þ1ß B VÐ Ñ œ  l2 l 2l0 $ e f

3œ" 5 3 3 :

exp

3

  • Best linear unbiased predictor BLUP
  • r empirical BLUP

Ð Ñ 0 known (EBLUP- ) can be calculated to predict

  • at any

0 estimated CÐ Ñ B C sÐB Ñ ´ I ] ÐB Ñl

! !

e f data

  • Engineering literature often calls such a

CÐ Ñ ´ s B metamodel

  • In addition,

5#

! ! ! ! #

ÐB Ñ œ I ] ÐB Ñ  ÐB Ñ l œ ÐB Ñl s s Š ‹

  • C

C data Var( data) is a measure of our uncertainty about the predicted value of CÐ ÑÞ B  "* 

slide-26
SLIDE 26

Example 8 œ (ß "-dim

The BLUP and corresponding pointwise 95% prediction interval limits for for CÐ Ñ B 8 œ ( training data observations  ") 

slide-27
SLIDE 27

Properties of CÐ Ñ s B!

  • Simple to compute (linear in training data

i.e., ÐCÐ Ñß á ß CÐ ÑÑÑß B B

" >< >< 8

CÐ Ñ œ - 

Ñ s B B

! ! 4 4œ" 8 >< 4

  • Viewed as a function of B!

CÐ Ñ œ .  . VÐ  Ñ s B B B

! ! 4 ! 3œ" 8 >< 3

  • interpolates the training data, i.e.,

CÐ Ñ s B CÐ Ñ œ CÐ Ñ 3 œ "ß á ß 8 s B B

>< >< 3 3 for

  • Splines, neural networks and other well-known interpolators correspond to specific

choices of correlation function

  • VÐ Ñ
  • Software
  • 1. SAS Proc Mixed
  • 2. PErK (B. J. Williams)

 "( 

slide-28
SLIDE 28

Part 2 Designing Computer Experiments to Find Global Optima Expected Improvement-I Goal Find B B

9:> œ

CÐ Ñ argmin

  • Implicit keep number of

evaluations "small'' CÐ Ñ B Idea of Sequential Design Algorithm for the Computer Experiment Given training data ˆ ‰ ˆ ‰ B B B B

" 8 8 " >< >< >< ><

ß CÐ Ñ ß á ß ß CÐ Ñ choose to be that which maximizes an criterion as B B

8" ><

expected improvement the "best'' input at which to compute

  • B9:>

CÐ Ñ Improvement Let C œ CÐ Ñ

8 3 >< min

min

" Ÿ 3 Ÿ 8

B ´ CÐ Ñ 8 best (smallest)

  • calculated through the

evaluation

th

 "' 

slide-29
SLIDE 29

Let C œ CÐ Ñ

8 3 >< min

min

" Ÿ 3 Ÿ 8

B ´ CÐ Ñ 8 Þ best (smallest)

  • calculated through the

evaluation Consider new

th

potential site . Then B define MÐ Ñ œ !ß CÐ Ñ C C  CÐ Ñß CÐ Ñ  C B B B B œ

8 8 8 min min min

œ !ß C  CÐ Ñ max˜ ™

8 min

B is the improvement in using as the minimizer over current training data. B B CÐ Ñ Warning In , is BUT and (hence) are :--(. MÐ Ñ C CÐ Ñ MÐ Ñ B B B

8 min

known unknown  "& 

slide-30
SLIDE 30

Idea of the Algorithm

  • 1. Obtain a starting design, ie., set of initial inputs

at which to calculate y(•) ß ß B3 (eg Space-filling "Latin Hypercube Design")

  • 2. Evaluate

at the starting design and use the data to estimate any unknown CÐ Ñ B correlation parameters 0

  • 3. Choose

to maximize the expected improvement the current data B8"

><

given (and ), i.e., B B B B

8" 3 3 >< 8 >< >< 3œ" 8

´ I MÐ Ñl] ´ ß CÐ Ñ argmax ˜ ™ ˜ ™ ˆ ‰ (1) (the expected value of under the stochastic process model) MÐ Ñ B %. when the maximum expected improvement is "small,'' and set Stop B B

  • 9:> œ

CÐ Ñß argmin where is the EBLUP of CÐ Ñ CÐ Ñ

  • B

B

Fact

B B B B B B B

8" 8 >< 8 8

´ ÐC  CÐ ÑÑ  Ð Ñ s C  CÐ Ñ C  CÐ Ñ s s Ð Ñ Ð Ñ argmax k F 5 9 5 5 œ

  • Œ
  • Œ
  • min

min min

slide-31
SLIDE 31

B B B B B B B

8" 8 >< 8 8

´ ÐC  CÐ ÑÑ  Ð Ñ s C  CÐ Ñ C  CÐ Ñ s s Ð Ñ Ð argmax k F 5 9 5 5 Ú Þ Û ß Ü à ðóóóóóóóóóóóóóóóñóóóóóóóóóóóóóóóò ðóóóóóóóóóñóóóóóóóóóò Œ

  • Œ
  • min

min min

Ñ Ð"Ñ Ð#Ñ Ð"Ñ B B B is large at if the predicted is "much lower" than the current CÐ Ñß CÐ Ñß s champion minimum, , i.e. both factors are large in this case C8

min

Ð#Ñ B B B is large at if there is large uncertainty in our prediction of , i.e., CÐ Ñ Ð Ñ 5 is "large" and is "small" - remember that (h) is max at ¹ ¹

C CÐ Ñ s Ð Ñ

8 min

B B 5

9 2 œ !Þ  "$ 

slide-32
SLIDE 32

Example (Hartman Function) 0ÐB ß B ß B ß B ß B ß B Ñ œ 

B  :

" # $ % & ' 3 34 4 34 3œ" 4œ" % ' #

  • Ÿ
  • exp

α ! Ÿ B Ÿ " 3 œ "ß á ß '

3 34

for , ( ) are given by α  "# 

slide-33
SLIDE 33

and ( ) are given by :34 Characteristics

  • has a unique has a unique minimum value of

0Ð Ñ  3.322375 EGO Algorithm

  • 1. Initial sample 8 œ &"
  • 2. Additional observations 74 (125 total)

(stopping when max expected improvement 0001 Ÿ Result 09 Minimum identified B 

  • 9:>

3.3223  "" 

slide-34
SLIDE 34

Expected Improvement-II

  • Setup:

1. where Inputs B œ ÐB ß B Ñ

  • /

( ) B œ

  • control manufacturing, engineering design variables

( , ) B œ

/

noise field enviromental variables (model variables) 2. has known distribution with support and probabilities i.e., X/

/ß4 =?: 4œ" 8 8 4 4œ"

˜ ™ ˜ ™ B

/ /

: ß : œ T œ " Ÿ 4 Ÿ 8

4 / / /ß3 =?:

˜ ™ X B for ("target field conditions'')

  • 3. In this case

which we can summarize in the usual ways. CÐ Ñ B ß \

  • /

distribution The simplest summary of the performance of the design is B- ß .Ð Ñ I CÐ ß Ñ œ : CÐ ß Ñ B œ B \ B B

  • e

f

  • /

3

  • 3œ"

8 3 =?:

/

Ð œ Z +< CÐ ß Ñ ß á Ñ Var( ) B B \

  • e

f

  • /
  • Goal Find B

B

  • ß9:> œ

Ð Ñ argmin .

  •  "! 
slide-35
SLIDE 35
  • Goal Find B

B

  • ß9:> ´

Ð Ñ argmin .

  • Training Data B

B B B B

3

  • ß3

/ß3

  • ß3

/ß3 >< >< >< >< ><

œ ß CÐ ß Ñ " Ÿ 3 Ÿ 8 ( ) and for

  • Bad News We don't know and won't compute

Ð Ñ . . . .

8

  • ß"
  • ß8
  • ß8

>< >< >< min ´

Ð Ñß Ð Ñß á ß Ð Ñ min˜ ™ B B B because we would need for for example, CÐ ß Ñ " Ÿ 4 Ÿ 8 ß B B

  • ß"

>< /ß4 =?: /

.Ð Ñ œ : CÐ ß Ñ B B B

  • ß"
  • ß"

>< >< 4œ" 8 3 4 =?:

  • /

  • Good News Can predict

because we can predict each component . .

8

  • ß3

>< min

ß Ð Ñß B by . s s Ð Ñ œ : CÐ ß Ñ B B B

  • ß3
  • ß3

>< >< 4œ" 8 4 4 =?:

  • /

and, in addition has a prior distribution induced by the .Ð Ñ ] Ð ß Ñ B B B

  • ß3

><

  • ß3

4 =?: /ß

QÐ Ñ œ : ] Ð ß Ñ B B B

  • ß3
  • ß3

>< >< 4œ" 8 3 4 =?:

  • /

slide-36
SLIDE 36

Idealized Improvment Function Define the improvement at (generic) control variable site B- MÐ Ñ œ !ß Ð Ñ  Ð Ñß Ð Ñ  B B B B

  • œ

. . . . . .

8 8 8 min min min

œ !ß  Ð Ñ max˜ ™ . .

8 min

B- is the improvement in using the mean response at . B- [all terms are :-(, BUT all terms :-)] unknown can be predicted

:

 ) 

slide-37
SLIDE 37

Idea of the Algorithm

  • 1. Obtain a starting design, ie., set of initial inputs

at which to B B B

3

  • 3

/ß3 >< >< ><

œ Ð ß Ñ

,

calculate y(•) e.g., Space-filling "Latin Hypercube Design") Ð

  • 2. Evaluate
  • • at the

starting design and use the training data to CÐ ß Ñ Ð ß Ñ ˜ ™ B B

>< ><

  • 3

/ß3 3œ" 8 ,

estimate any unknown correlation parameters 0

  • 3. Choose

to maximize the posterior expected improvement, i.e., B-ß8"

><

B B B B

  • ß8"

3 3 >< 8 >< >< 3œ" 8

´ I MÐ Ñl] ´ C ß CÐ Ñ argmax ˜ ™ ˜ ™ ˆ ‰

  • (1)

(the expected value of under the stochastic process model MÐ Ñ B %. Choose in the enviromental variable space to minimize Var( ( )) ^ B B

/ß8"

  • ß8"

>< ><

. where ( ) is the predicted mean (over the environmental variables) at the ^ . B-ß8"

><

selected next control variable &Þ Stop when the maximum expected improvement is "small,'' and set B B

  • ß9:>
  • œ

Ð Ñß argmin . where is the EBLUP of . OW, augment the current design. . .

  • Ð Ñ

Ð Ñ B B

slide-38
SLIDE 38

 ' 

slide-39
SLIDE 39

The mean surface and the estimated based on a maximin true . . ÐB ß B Ñ ÐB ß B Ñ

" # " #

LHD of size 40. The minimizer is (0.2036, 0.2545) with global B-ß9:> . . Ð Ñ œ $#$Þ!""(% Ð ÐB ß B Ñ !Þ#&%%& B-ß9:>

" # has

minimum at (1, ) and local (0.46287, )) !Þ#&%%&  & 

slide-40
SLIDE 40

 % 

slide-41
SLIDE 41
  • 1. The initial fit is poor
  • 2. Running the algorithm with the Matern correlation function, stopping occurs after

156 total (•) function evaluations with C œ Ð!Þ#"!*'ß !Þ#$$#%Ñ B

  • ß9:>

Ð œ $#'Þ'(!!&Ñ and ( ) . B

  • ß9:>

$Þ The global minimum of (•) is within 1.15% of the true global minimum .  $ 

slide-42
SLIDE 42

  #

slide-43
SLIDE 43
  • 6. Take Home Points

". An increasing number of phenomenon that could be previously be studied only by physical experiments, can now be investigated using "computer experiments'' #. Modeling responses from computer experiment must account for the (highly) correlated nature of the output

  • ver the input space. Predictive models are used

CÐ Ñ B to interpolate the computer response at untried locations $. The design of most computer experiments is naturally sequential; we evaluate CÐ Ñ CÐ Ñ

  • at one set of inputs , learn more about
  • , and select new inputs to achieve

B some objective %. EGO algorithms balance sampling an output (surface) where the predictor indicates the minimum is located, with improving our knowledge of the surface at points having large prediction error. &. The EGO algorithms can be modified to accomodate optimization with noisy

  • utput (simulation error/numerical or modeling bias) or from calibrated computer

and physical experiments 'Þ Pattern search and other algorithms have been used for the same purpose (Booker, Dennis, Torozon, Trosset). Their usefulness depends on the cost of

  • evaluations.

CÐ Ñ

 " 

slide-44
SLIDE 44

Some References

Booker, A. J., A. R. Conn, J. E. Dennis, P. D. Frank, M. Trosset, and V. Torczon (1995). Global modeling for

  • ptimization. Technical Report ISSTECH-95-032. Boeing Information & Support Services, Seattle, WA.

Boeing/IBM/Rice Collaborative Project 1995 Final Report. Booker, A. J., J. E. Dennis, P. D. Frank, D. B. Serafini, V. Torczon, and M. W. Trosset (1999). A rigorous framework for

  • ptimization of expensive functions by surrogates.

17, 1-13. Structural Optimization Dennis, J. E. and V. Torczon (1997). Managing approximation models in optimization. In: Alexandrov, N. and M. Y. Hussaini, Eds., , SIAM, Philadelphia, 330-347. Multidisciplinary Design Optimization: State of the Art Glad, T. and A. Goldstein (1977). Optimization of functions whose values are subject to small errors. 17, 160-169. BIT Lehman, J., T. J. Santner, and W. I. Notz (2004) Design of Computer Experiments to Determine Robust Control Variables, Statistica Sinica, , 571-580. 14 Lewis, R. M. and V. Torczon (1998a). A globally convergent augmented Lagrangian pattern search method for

  • ptimization with general constraints and simple bounds. Technical Report 98-31. Institute for Computer Applications

in Science and Engineering, NASA Langley Research Center, Hampton, VA 23681-0001. Sacks, J. W. J. Welch, T. J. Mitchell, and H. P. Wynn (1989). Design and analysis of computer experiments, Statistical Science 4, 409-435. Includes discussion. Sasena, M. J., P. Papalambros, and P. Goovaerts (2002). Exploration of metamodeling sampling criteria for constrained global optimization. In: , pages 263-278. Engineering Optimization Santner, T. J. B.~J. Williams, and W.~I. Notz (2003) , ß The Design and Analysis of Computer Experiments Springer-Verlag, Inc. Schonlau, M. (1997). Computer Experiments and Global Optimization. PhD thesis, University of Waterloo, Waterloo, Ontario.

slide-45
SLIDE 45

Schonlau, M., and W. J. Welch (1996). Global optimization with nonparametric function fitting. In Proceedings of the Section on Physical and Engineering Sciences, American Statistical Association, 183-186. Schonlau, M., W. J. Welch, and D. R. Jones (1997). A data-analytic approach to Bayesian global optimization. In: Proceedings of the Section on Physical and Engineering Sciences, American Statistical Association, 186-191. Schonlau, M., W. J. Welch, and D. R. Jones (1998). Global versus local search in constrained optimization of computer

  • models. In: Flournoy, N., W. F. Rosenberger, and W. K. Wong, Eds., New Developments and Applications in

Experimental Design. , 11-28, Institute of Mathematical Statistics, Hayward, CA. 34 Trosset, M. W. and V. Torczon (1997). Numerical optimization using computer experiments. Technical Report 97-38. Institute for Computer Applications in Sciences & Engineering, NASA Langley Research Center, Hampton, VA 23681-0001. Williams, B. J., T. J. Santner, and W. I. Notz (2000) Sequential Design of Computer Experiments to Minimize Integrated Response Functions, , , 1133-1152. Statistica Sinica 10 Williams, B. J., T. J. Santner, and W. I. Notz Sequential Design of Computer Experiments to Minimize Integrated Response Functions with Constraints, in preparation.