1 Outline Problem Setting Instance-Based vs. Model-Based - - PowerPoint PPT Presentation

1
SMART_READER_LITE
LIVE PREVIEW

1 Outline Problem Setting Instance-Based vs. Model-Based - - PowerPoint PPT Presentation

1 Outline Problem Setting Instance-Based vs. Model-Based Model-Based Algorithms Estimation of Distribution Algorithms (EDAs) Cross-Entropy (CE) Method Model Reference Adaptive Search (MRAS) Convergence of MRAS


slide-1
SLIDE 1

1

slide-2
SLIDE 2

2

Outline

  • Problem Setting
  • Instance-Based vs. Model-Based
  • Model-Based Algorithms

– Estimation of Distribution Algorithms (EDAs) – Cross-Entropy (CE) Method – Model Reference Adaptive Search (MRAS)

  • Convergence of MRAS
  • Numerical Examples
  • Extension to Stochastic Optimization and MDPs
  • A New Particle Filtering Framework (if time)
slide-3
SLIDE 3

3

  • Solution space
  • continuous or discrete (combinatorial)
  • Objective function H(·):
  • Objective: find optimal such that
  • Assumptions: existence, uniqueness

(but possibly many local minima)

Problem Setting

slide-4
SLIDE 4

4

  • Instance-based approaches: search for new solutions

depends directly on previously generated solutions

  • simulated annealing (SA)
  • genetic algorithms (GAs)
  • tabu search
  • nested partitions

Overview of Global Optimization Approaches

slide-5
SLIDE 5

5

Model-Based Search Methods

probability model gk sampling updating mechanism new candidate solutions Xk selection

slide-6
SLIDE 6

6

Model-Based Approach: Graphical Depiction

slide-7
SLIDE 7

7

Combinatorial Optimization Example: TSP

How do we formulate this problem to use a probability distribution?

  • routing matrix of probability of arc i  j.
  • Example: four cities

[0 0.5 0.4 0.1] [0.2 0 0.6 0.2] [0.4 0.4 0 0.2] [0.3 0.3 0.4 0 ]

  • What is convergence?
  • single 1 in each row
  • single 1 in each column
slide-8
SLIDE 8

8

Model-Based Methods

similarities to genetic algorithms

  • uses a population
  • selection process
  • randomized algorithm,

but uses “model” (distribution) instead of operators

slide-9
SLIDE 9

9

Main Model-Based Methods

  • estimation of distribution algorithms (EDAs)

Muhlenbein and Paas (1996); book by Larranaga and Lozano (2001) [other names, e.g., probabilistic model-building GAs]

  • cross-entropy method (CE)

Rubinstein (1997, 1999) (www.cemethod.org); book by Rubinstein and Kroese (2004)

  • probability collectives (Wolpert 2004)
  • model reference adaptive search (MRAS)
slide-10
SLIDE 10

10

Model-Based Methods (continued)

BIG QUESTION: How to update distribution?

  • traditional EDAs use an explicit construction,

can be difficult & computationally expensive

  • CE method uses single fixed target distribution

(optimal importance sampling measure)

  • MRAS approach:

sequence of implicit model reference distributions

slide-11
SLIDE 11

11

  • ALTERNATIVE: sample from a parameterized

family of distributions, and update parameters by minimizing “distance” to desired distributions (reference distributions in MRAS)

MRAS and CE Methods

parameterized distribution

samples

parameter selection

parameterized family reference distributions

slide-12
SLIDE 12

12

  • Main characteristics
  • Given sequence of reference distributions {gk(·)}
  • works with a family of parameterized probability

distributions {f (·,θ)} over the solution space

  • fundamental steps at iteration k :

* generate candidate solutions according to the current probability distribution f (·, θk) * calculate θk+1 using data collected in previous step to bias future search toward promising regions, by minimizing distance between {f (·,θ)} and gk+1(·)

  • Algorithm converges to optimal if {gk(·)} does

Model Reference Adaptive Search

slide-13
SLIDE 13

13

  • reference distribution construction:

Next distribution obtained by tilting previous where S(.) is non-negative and strictly decreasing (increasing for max problems)

Properties:

  • selection parameter ρ determines the proportion of

solutions used in updating θk+1

MRAS: Specific Instantiation

slide-14
SLIDE 14

14

  • (1- ρ)-quantiles w.r.t. f (·,θk)
  • update θk+1 as

where

MRAS: Parameter Updating

slide-15
SLIDE 15

15

  • covers broad class of distributions
  • closed-form solution for θk+1
  • global convergence can be established under some

mild regularity conditions

*

multivariate Gaussian case

*

independent univariate case Restriction to Natural Exponential Family (NEF)

slide-16
SLIDE 16

16

MRAS: Monte-Carlo version

slide-17
SLIDE 17

17

Comparison of MRAS & CE

slide-18
SLIDE 18

18

Numerical Examples (deterministic problems)

slide-19
SLIDE 19

19

slide-20
SLIDE 20

20

slide-21
SLIDE 21

21

slide-22
SLIDE 22

22

slide-23
SLIDE 23

23

  • Numerical results for ATSPs
  • DISCRETE distribution (matrix: probability ij on tour)
  • Good performance with modest number of tours generated
  • ft70 case: total number of admissible tours = 70! ≈ 10100

Numerical Examples (deterministic problems)

slide-24
SLIDE 24

24

where are i.i.d. random observations at x.

Extension to Stochastic Optimization

slide-25
SLIDE 25

25

Extension to Stochastic Optimization

slide-26
SLIDE 26

26

  • Xt : inventory position in period t.
  • Dt : the i.i.d exponential demand in period t
  • h : per period per unit holding cost; p: demand

lost penalty cost ; c: per unit ordering cost; K: fixed set-up cost

  • The objective is to minimize the long run

average cost per period:

(s,S) Inventory Control Problem

slide-27
SLIDE 27

27

Case 1: c = h = 1, p=10, K=100, E[D]=200 Case 2: c = h = 1, p=10, K=10000, E[D]=200

(s,S) Inventory Control Problem

slide-28
SLIDE 28

28

S2 S3 S4 buffer1 buffer2 buffer3

  • Input:
  • μi: service rate of server i
  • fi : failure rate of server i
  • ri : repair rate of server i
  • n : total number of buffers available
  • Let ni be the number of buffers allocated to Si

satisfying Σni= n, the objective is to choose ni to maximize the steady-state throughput

Buffer Allocation in Unreliable Production Lines

S1

slide-29
SLIDE 29

29

Buffer Allocation in Unreliable Production Lines

slide-30
SLIDE 30

30

Extension to MDPs

slide-31
SLIDE 31

31

Filtering (with Enlu Zhou and M. Fu)

slide-32
SLIDE 32

32

Optimization via Filtering

slide-33
SLIDE 33

33

Optimization via Filtering

Result: Using particle filtering (Monte Carlo simulation), EDAs, CE, MRAS can all be viewed in this framework.

slide-34
SLIDE 34

34

  • Summary
  • new general framework for problems with little structure
  • guaranteed theoretical convergence
  • good experimental performance
  • Future Work
  • incorporate known structure (e.g., local search)
  • convergence rate, computational complexity
  • more new algorithm instantiations in this framework
  • more comparisons with other algorithms