efficient simulation sampling allocation using multi
play

Efficient Simulation Sampling Allocation Using Multi-Fidelity Models - PowerPoint PPT Presentation

Efficient Simulation Sampling Allocation Using Multi-Fidelity Models Jie Xu Dept. of Systems Engineering & Operations Research George Mason University Fairfax, VA jxu13@gmu.edu Joint work with Y. Peng, C.-H. Chen, L.-H. Lee, J.-Q. Hu


  1. Efficient Simulation Sampling Allocation Using Multi-Fidelity Models Jie Xu Dept. of Systems Engineering & Operations Research George Mason University Fairfax, VA jxu13@gmu.edu Joint work with Y. Peng, C.-H. Chen, L.-H. Lee, J.-Q. Hu Supported by NSF and AFOSR under Grants ECCS-1462409 and CMMI-1462787

  2. S IMULATION - BASED DECISION MAKING • Simulation provides a predictive tool for decision making when problems are intractable to analytical approaches • This talk considers a special case known as ranking & selection ‒ [�] �∈{�,�,…,�} � ‒ Stochastic black-box objective functions, observed by running iid replications of a simulation model • Fruitful research on simulation-based decision making ‒ Efficient sampling/allocation of simulation budget, convergent fast local search, parallelization, surrogate model ‒ Open-source solver ISC (www.iscompass.net) has been used by MITRE and the Argonne National Lab in real-world problems air traffic management and power systems applications ‒ What if the full-scale simulation model runs for hours? G E O R G E M A S O N U N I V E R S I T Y

  3. C AN APPROXIMATION MODELS HELP ? Full-featured model Approximation model High-fidelity/full-scale Low-fidelity/reduced-scale discrete-event simulation, simulation, analytical agent-based model, etc. approximation, full-model with archived data Complex Simple Accurate Approximate Time-consuming Fast G E O R G E M A S O N U N I V E R S I T Y

  4. M ULTI - FIDELITY OPTIMIZATION METHODS • A naïve way of multi-fidelity optimization ‒ Find some most promising designs using the approximation model ‒ Evaluation using high-fidelity simulations • Most approaches use interpolation/regression to “correct” low-fidelity model ‒ Autoregressive framework with kriging/Gaussian process regression (Kennedy and O’Hagan 2000) ‒ Radial basis function, Polynomial chaos • Significant challenges arise when ‒ Solution space is high-dimensional ‒ High-fidelity simulation samples have heterogeneous noise ‒ Quality of low-fidelity model is low ‒ Mixed decision variables (integer, categorical) G E O R G E M A S O N U N I V E R S I T Y

  5. S IMULATION O PTIMIZATION : A N I LLUSTRATIVE E XAMPLE Resource allocation problem in a flexible manufacturing system P1 P1 P2 P2 ‒ 2 product types ‒ 5 workstations Workstation 1 Workstation 1 ‒ Non-exponential service times Workstation 2 Workstation 2 ‒ Re-entrant manufacturing process ‒ Product 1 has higher priority than product 2 Workstation 3 Workstation 3 Optimization problem: Workstation 4 Workstation 4 Decision variable Number of machines at each workstation Workstation 5 Workstation 5 Objective Minimize Expected Total Processing Time G E O R G E M A S O N U N I V E R S I T Y

  6. E XAMPLE : R ESOURCE A LLOCATION P ROBLEM Decision variables: number of machines allocated to each workstation Total 780  Simulation/evaluation can be time consuming  Solution space dimension can be large G E O R G E M A S O N U N I V E R S I T Y

  7. F ULL S IMULATION & A PPROXIMATION M ODELS Bias is non-homogeneous and can be quite large G E O R G E M A S O N U N I V E R S I T Y

  8. O RDINAL RANKINGS OF DESIGNS BY LOW - FIDELITY MODEL Designs with similar performance are grouped together, which may potentially enhance search/optimization efficiency G E O R G E M A S O N U N I V E R S I T Y

  9. A BAYESIAN FRAMEWORK FOR MULTI - FIDELITY MODELS • For design , , we model the prior distribution of high-fidelity ( ) prediction and d -dimensional low-fidelity predictions ( ) by a Gaussian mixture model � ‒ � � � � � � ��� � � � � �� ‒ , � � � � � � � � � � � • can only be observed with a Gaussian noise • is completely observed (negligible computing cost) ‒ � � � • We allocate a total of high-fidelity simulation replications to designs ‒ Let � denote the samples collected after simulation replications ‒ Let � be the number of simulation replications allocated to design after simulation replications G E O R G E M A S O N U N I V E R S I T Y

  10. M ODEL ESTIMATION • We extend classical model-based clustering results to the multi-fidelity setting with stochastic observations of ‒ Binary hidden state random variable �,� assigns design to cluster ‒ �,� follows a multinomial distribution with parameters � �,� � � • The maximal likelihood estimate of model parameters , (�) ‒ � � � � � ∈� � ‒ � � � � � � � ��� � �,� � � � � � ��� � ��� ℝ • The Expectation-maximization (EM) algorithm is applied to compute G E O R G E M A S O N U N I V E R S I T Y

  11. M ODEL ESTIMATION -C ONT . • We estimate the number of components using the completely observed low-fidelity estimates • Bayesian information criterion (BIC) is used to select (���)(���) ��� � ‒ � � � � � � ‒ � , where � � , � � � � ��� � � ∈� � � � ‒ � � ��� � � � � ��� ‒ We select the M from a specified interval that has the largest � G E O R G E M A S O N U N I V E R S I T Y

  12. T HEOREM 1: S TOCHASTIC MODEL - BASED CLUSTERING • After EM iteration t , the posterior probability of conditional on and given is (�,�) � �,� (�,�) � � � (�,�) ‒ , where �,� (�,�) � �,� (�,�) � ∑ � � � ��� (�,�) � �,� � 𝓌 �,� � � (�) �,� �,� �,� �,� ‒ � � � � �,� �,� �,� � � � 𝓌 � � � �,� �,� �,� �,� �,� , � � � � � � � (�,���) � (�,���) (�,���) � � � � ∑ ∑ �̂ �,� �̂ �,� � �,� �,� �,� ��� ��� �,� ‒ , � , � �� (�,���) (�,���) �,� �,� � � ∑ �̂ �,� ∑ �̂ �,� � �� � � ��� ��� �� G E O R G E M A S O N U N I V E R S I T Y

  13. T HEOREM 1: S TOCHASTIC MODEL - BASED CLUSTERING • The posterior distribution of conditional on { , , , and given is normal with density function (�,�) � � � (�,�) (�) (�) (�,�) (�,�) (�,�) ‒ � � � � � � �,� �,� �,� � � Weighted high-fidelity Weighted prediction Weighted cluster mean simulation sample mean using low-fidelity predictions • The estimates of the model parameters are updated in the next EM iteration accordingly • The above results can be extended for noisy G E O R G E M A S O N U N I V E R S I T Y

  14. A SYMPTOTIC RESULTS • Corollary 1: Suppose that design is sampled infinitely often as , then (�,�) ,� � (�,�) �,� � � � |� � � � �,� ‒ almost surely (�,�) (�,�) ,� � �,� � � �→� ∑ ∑ � � � �,� � � � |� ��� ��� ‒ This result is consistent with the classical model-based clustering (�,�) playing the role of (�,�) �,� result with �,� when the � � � effect of stochastic simulation noise is eliminated • Using asymptotic results, we obtain lightweight approximations for posterior estimates that do not require EM iteration G E O R G E M A S O N U N I V E R S I T Y

  15. A SYMPTOTICALLY O PTIMAL SAMPLING ALLOCATION P OLICY • Allocate high-fidelity simulations to to maximize the large deviation rate of incorrect selection event • The large deviation rate of when is given by � � � �,� � � � � � � � � • Define an approximate large deviation rate (ALDR) (�) � , where � � ,� � � � � ��� � ���,…,� • It can be shown that converges with probability 1 to an upper bound on the large deviation rate of incorrect selection event G E O R G E M A S O N U N I V E R S I T Y

  16. M ULTI - FIDELITY BUDGET ALLOCATION POLICY • Based on the clustering statistics, we define the following posterior means and variances (�) , (�) (�) , (�) ‒ , where � is the cluster with � � � � � � �,� �,� the largest clustering statistic for design ‒ Let be the design index after sorting all designs in descending (�) (�) order posterior means, i.e., [�] [�] (�) � (�) (�) ‒ Let � [�] � • The (approximately) optimal sampling allocation policy can be obtained by solving (�) � [�] (�) � 𝓌 [�] � [�] � [�] (�) (�) for [�] ��� (�) � [�] (�) [�] � [�] 𝓌 [�] 𝓌 [�] G E O R G E M A S O N U N I V E R S I T Y

  17. UNDERSTANDING THE SAMPLING ALLOCATION POLICY Signal to Noise Ratio (�) [�] (�) [𝟒] (�) (�) (�) [�] [�] [�] (�) [�] [𝟑] (�) (�) [�] [�] [�] [1] [2] [3] Design inversely proportional to the square of the signal to noise ratio G E O R G E M A S O N U N I V E R S I T Y

  18. M ACHINE ALLOCATION RESULTS • Compare the PCS achieved by the new multi-fidelity budget allocation policy (MFBA) with optimal computing budget allocation (OCBA) for one fidelity level and equal allocation (EQ) G E O R G E M A S O N U N I V E R S I T Y

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend