Monte Carlo methods for sampling-based Stochastic Optimization - - PowerPoint PPT Presentation

monte carlo methods for sampling based stochastic
SMART_READER_LITE
LIVE PREVIEW

Monte Carlo methods for sampling-based Stochastic Optimization - - PowerPoint PPT Presentation

Monte Carlo methods for sampling-based Stochastic Optimization Monte Carlo methods for sampling-based Stochastic Optimization Gersende FORT LTCI CNRS & Telecom ParisTech Paris, France Joint works with B. Jourdain, T. Leli` evre, G.


slide-1
SLIDE 1

Monte Carlo methods for sampling-based Stochastic Optimization

Monte Carlo methods for sampling-based Stochastic Optimization

Gersende FORT

LTCI CNRS & Telecom ParisTech Paris, France

Joint works with

  • B. Jourdain, T. Leli`

evre, G. Stoltz from ENPC and E. Kuhn from INRA.

  • A. Schreck and E. Moulines from Telecom ParisTech
  • P. Priouret from Paris VI
slide-2
SLIDE 2

Monte Carlo methods for sampling-based Stochastic Optimization Introduction Simulated Annealing

Simulated Annealing (1/2)

Let U denote the objective function one wants to minimize. min

x∈X U(x) ⇐

⇒ max

x∈X exp(−U(x)) ⇐

⇒ max

x∈X exp

  • −U(x)

T

  • ∀T > 0

In order to sample from πT⋆ where πT (x) = exp

  • −U(x)

T

  • sample successively from a sequence of tempered distributions πT1, πT2,

· · · with T1 > T2 > · · · > T⋆.

slide-3
SLIDE 3

Monte Carlo methods for sampling-based Stochastic Optimization Introduction Simulated Annealing

Simulated Annealing (1/2)

Let U denote the objective function one wants to minimize. min

x∈X U(x) ⇐

⇒ max

x∈X exp(−U(x)) ⇐

⇒ max

x∈X exp

  • −U(x)

T

  • ∀T > 0

In order to sample from πT⋆ where πT (x) = exp

  • −U(x)

T

  • sample successively from a sequence of tempered distributions πT1, πT2,

· · · with T1 > T2 > · · · > T⋆.

  • r sample successively from a sequence of (nt-iterated) kernels (PTt(x, ·))t

such that πT PT = πT .

slide-4
SLIDE 4

Monte Carlo methods for sampling-based Stochastic Optimization Introduction Simulated Annealing

Simulated Annealing (2/2)

Under conditions on X, on the cooling schedule (Tt)t, on the kernels (Pt)t, on the dominating measure and the set of minima, · · ·

Kirkpatrick, Gelatt and Vecchi. Optimization via Simulated Annealing. Science (1983) Geman and Geman. Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images. IEEE Trans. on

  • PAMI. (1984).

Van Laarhoven and Aarts, Simulated Annealing : theory and applications. Mathematics and its Applications, Reidel, Dordrecht (1987). Chiang and Chow. On the convergence rate of annealing processes. SIAM J. Control Optim. (1988) Hajek, B. Cooling schedule for optimal annealing. Math. Operat. Res. (1988). Haario and Saksman. Simulated Annealing process in general state space. Adv. Appl. Probab. (1991)

Xt converges to the minima of U

slide-5
SLIDE 5

Monte Carlo methods for sampling-based Stochastic Optimization Introduction Sampling a density

Sampling a density

Monte Carlo methods are numerical tools to solve some computational problems in bayesian statistics, for the exploration of the a posteriori distribution π computation of integrals (w.r.t. π) stochastic optimization (of U, π ∝ exp(U)) · · · Monte Carlo methods draw points (Xt)t approximating π π ≈ 1 T

T

  • t=1

δXt even in difficult situations when perfect sampling under π is not possible π known up to a normalization constant complex expression of π if explicit large dimension of the state space · · ·

slide-6
SLIDE 6

Monte Carlo methods for sampling-based Stochastic Optimization Introduction Sampling a density

Two main strategies : Importance Sampling & MCMC (1/2)

  • 1. Importance Sampling :

Choose an auxiliary distribution π⋆ Draw points approximating π⋆ Reweight these draws to approximate π

  • Ex. (Xt)t i.i.d. under π⋆,

π ≈ 1 T

T

  • t=1

π(Xt) π⋆(Xt)δXt Main drawback in large dimension : not robust at all when the dimension is large : degeneracy of the weights, large and even infinite variance if π⋆ is not selected in accordance with π. MCMC far more robust to the dimension

slide-7
SLIDE 7

Monte Carlo methods for sampling-based Stochastic Optimization Introduction Sampling a density

Two main strategies : Importance Sampling & MCMC (2/2)

  • 2. Markov Chain Monte Carlo (MCMC) : Sample a Markov chain, with unique

invariant distribution π

  • Ex. Hastings-Metropolis type algorithms :

Choose an auxiliary transition kernel q(x, y) Starting from the current point Xt, propose a candidate Y ∼ q(Xt, ·) Accept or Reject the candidate Xt+1 = Y with probability α(Xt, Y ) Xt with probability 1 − α(Xt, Y ) where α(x, y) = 1 ∧ π(y)q(y, x) π(x)q(x, y). Main drawback of classical MCMC samplers for multimodal densities on large dimensional space have to scale the size of the proposed moves as a function of the dimension remain trapped in some modes, unable to jump and visit the sampling space in a “correct” time.

slide-8
SLIDE 8

Monte Carlo methods for sampling-based Stochastic Optimization Introduction Sampling a density

Example 1

  • Ex. MCMC - Size of the proposed moves w.r.t. the dimension.

−4 −3 −2 −1 1 2 3 4 −4 −3 −2 −1 1 2 3 4 d= 2 (α = 0.24) −4 −3 −2 −1 1 2 3 −4 −3 −2 −1 1 2 3 4 d= 8 (α = 0.01) 0.5 1 1.5 2 2.5 −0.5 0.5 1 1.5 2 d= 32 (α = 0) −4 −3 −2 −1 1 2 3 4 −5 −4 −3 −2 −1 1 2 3 4 d= 2 (α = 0.36) −4 −3 −2 −1 1 2 3 4 −4 −3 −2 −1 1 2 3 4 d= 8 (α = 0.27) −4 −3 −2 −1 1 2 3 4 −3 −2 −1 1 2 3 4 d= 32 (α = 0.24)

Plot of the first two components of the chain in Rd with target π = Nd(0, I) for d ∈ {2, 8, 32} : the candidate is Y = Xt + Nd(0, σ2I). σ does not depend on d (top) and σ is of the form c/ √ d (bottom)

slide-9
SLIDE 9

Monte Carlo methods for sampling-based Stochastic Optimization Introduction Sampling a density

Example 2

The target density π is a mixture of Gaussian in R2 π ∝

20

  • i=1

N2(µi, Σi) We compare N i.i.d. points (left) to N points from a Hastings-Metropolis chain (right)

1 2 3 4 5 6 7 8 9 10 −2 2 4 6 8 10 Target density : mixture of 2−dim Gaussian draws means of the components 1 2 3 4 5 6 7 8 9 1 2 3 4 5 6 7 8 9 10 Hastings−Metropolis draws means of the components

Classical adaptive MCMC are not robust to the multimodality problem

slide-10
SLIDE 10

Monte Carlo methods for sampling-based Stochastic Optimization Introduction How to tackle multimodality in large dimension ?

How to tackle the multimodality question ?

Here are some directions recently proposed in the Statistic literature :

  • 1. Biasing potential approach

Identify (few) “directions of metastability” ξ(x) and a biasing potential A(ξ(x)) such that π⋆(x) ∝ π(x) exp(−A(ξ(x))) has better mixing properties Sample under π⋆ and add a reweighting mecanism to approximate π.

  • Ex. the Wang-Landau sampler
slide-11
SLIDE 11

Monte Carlo methods for sampling-based Stochastic Optimization Introduction How to tackle multimodality in large dimension ?

How to tackle the multimodality question ?

Here are some directions recently proposed in the Statistic literature :

  • 1. Biasing potential approach

Identify (few) “directions of metastability” ξ(x) and a biasing potential A(ξ(x)) such that π⋆(x) ∝ π(x) exp(−A(ξ(x))) has better mixing properties Sample under π⋆ and add a reweighting mecanism to approximate π.

  • Ex. the Wang-Landau sampler
  • 2. Tempering methods and Interactions

Choose a set of inverse temperature 0 < β1 < · · · < βK−1 < 1 Sample points approximating the tempered densities πβi by allowing interactions between these points.

  • Ex. the Equi-Energy sampler
slide-12
SLIDE 12

Monte Carlo methods for sampling-based Stochastic Optimization The Wang-Landau algorithm

Outline

Introduction The Wang-Landau algorithm The proposal distribution A toy example Approximation of π Efficiency of the Wang-Landau algorithm Convergence issues Combining WL and simulated annealing

slide-13
SLIDE 13

Monte Carlo methods for sampling-based Stochastic Optimization The Wang-Landau algorithm

The Wang-Landau algorithm

The algorithm was proposed by Wang and Landau in 2001, in the molecular dynamics field.

F.G. Wang and D.P. Landau, Determining the density of states for classical statistical models : A random walk algorithm to produce a flat histogram, Phys. Rev. E 64 (2001).

  • G. Fort, B. Jourdain, E. Kuhn, T. Leli`

evre and G. Stoltz. Convergence of the Wang-Landau algorithm. Accepted for publication in Mathematics of Computation, March 2014.

  • G. Fort, B. Jourdain, E. Kuhn, T. Leli`

evre and G. Stoltz. Efficiency of the Wang-Landau algorithm. Accepted for publication in Applied Mathematics Research Express, February 2014.

  • L. Bornn, P. Jacob, P. Del Moral and A. Doucet. An Adaptive Wang-Landau Algorithm for Automatic Density Exploration.

Journal of Computational and Graphical Statistics (2013).

  • P. Jacob and R. Ryder. The Wang-Landau algorithm reaches the flat histogram criterion in finite time. Ann. Appl. Probab. (2013).

Approximate the target π Definition of the proposal distribution − Sample points approximating the proposal distribution − Compute the associated weights

slide-14
SLIDE 14

Monte Carlo methods for sampling-based Stochastic Optimization The Wang-Landau algorithm The proposal distribution

The proposal distribution (1/3)

Wang-Landau is an importance sampling algorithm with proposal π⋆

π⋆(x) ∝ d

i=1 π(x) π(Xi) 1

IXi(x)

where X1, · · · , Xd is a partition of the sampling space. The proposal distribution π⋆ consists in reweighting locally π so that ∀i, π⋆(Xi) = 1 d ֒ → This last property will force the sampler π⋆ to visit all the strata, with the same frequency.

slide-15
SLIDE 15

Monte Carlo methods for sampling-based Stochastic Optimization The Wang-Landau algorithm The proposal distribution

The proposal distribution (2/3)

Unfortunately, π(Xi) is unknown and we can not sample from π⋆ (even with MCMC) The algorithm will use a family of biased distributions πθ(x) ∝

d

  • i=1

π(x) θ(i) 1 IXi(x) where θ = (θ(1), · · · , θ(d)) is a weight vector. Key property : π⋆ is among this family π⋆(x) = πθ⋆(x) with θ⋆ = π(X1) Zπ , · · · , π(Xd) Zπ

  • The algorithm will simultaneously (a) learn the target weight θ⋆ and (b)

produce points approximating π⋆.

slide-16
SLIDE 16

Monte Carlo methods for sampling-based Stochastic Optimization The Wang-Landau algorithm The proposal distribution

The proposal distribution (3/3)

The algorithm (step 1) Given the current biasing weight θt and the current sample Xt sample the new point : Xt+1 ∼ Pθt(Xt, ·) where Pθ is a Markov kernel s.t. πθPθ = πθ Update the biasing weight : if Xt+1 ∈ Xi, penalize the stratum i in order to favor the visits to the other stratum. Since πθ(x) ∝ π(x)/θ(ℓ) when x ∈ Xℓ, θt+1(i) > θt(i) θt+1(k) < θ(k)

slide-17
SLIDE 17

Monte Carlo methods for sampling-based Stochastic Optimization The Wang-Landau algorithm The proposal distribution

The proposal distribution (3/3)

The algorithm (step 1) Given the current biasing weight θt and the current sample Xt sample the new point : Xt+1 ∼ Pθt(Xt, ·) where Pθ is a Markov kernel s.t. πθPθ = πθ Update the biasing weight : if Xt+1 ∈ Xi, penalize the stratum i in order to favor the visits to the other stratum. Since πθ(x) ∝ π(x)/θ(ℓ) when x ∈ Xℓ, θt+1(i) > θt(i) θt+1(k) < θ(k)

  • Ex. of updating strategy :

θt+1(i) = θt(i) + γt+1 θt(i)(1 − θt(i)) θt+1(k) = θt(k) − γt+1 θt(i)θt(k) based on a Stochastic Approximation algorithm, with deterministic (non increasing) stepsize sequence (γt)t

slide-18
SLIDE 18

Monte Carlo methods for sampling-based Stochastic Optimization The Wang-Landau algorithm A toy example

A toy example (1/3)

Target density : π(x1, x2) ∝ exp(−β H(x1, x2))1 I[−R,R](x1)

−2.5 −2 −1.5 −1 −0.5 0.5 1 1.5 2 2.5 −2 −1.5 −1 −0.5 0.5 1 1.5 2 2.5 −2 −1 1 2 3 −2.5 −2 −1.5 −1 −0.5 0.5 1 1.5 2 2.5 1 2 3 4 5 6 7 8 −2 −1.5 −1 −0.5 0.5 1 1.5 2 2.5 −4 −2 2 4 1 2 3 4 5 6 7 8

Figure: [left] Level curves of the potential H. [center, right] Density π up to a normalizing constant.

−2 −1 1 2 3 −4 −2 2 4 10 20 30 40 50 60 beta=1 −2 −1 1 2 3 −4 −2 2 4 1 2 3 4 5 x 10

8

beta=5

The larger β is, the lar- ger is the ratio between the weight of the strata located near to the main metastable states and the weight of the tran- sition region (near x1 = 0).

slide-19
SLIDE 19

Monte Carlo methods for sampling-based Stochastic Optimization The Wang-Landau algorithm A toy example

A toy example (2/3)

−2.5 −2 −1.5 −1 −0.5 0.5 1 1.5 2 2.5 −2 −1.5 −1 −0.5 0.5 1 1.5 2 2.5

d = 48 strata, partition along the x-axis. Pθ are Hastings-Metropolis kernels with proposal distribution N(0, (2R/d)2I) and target πθ. X0 = (−1, 0). R = 2.4. The stepsize sequence is γt ∼ c/t0.8.

0.5 e6 1 e6 1.5 e6 2 e6 2.5 e6 3 e6 0.02 0.04 0.06 0.08 0.1 0.12 0.14 −2.5 −2 −1.5 −1 −0.5 0.5 1 1.5 2 2.5 0.02 0.04 0.06 0.08 0.1 0.12

Figure: [left] The sequences (θt(i))t. [right] The limiting value θ⋆(i)

slide-20
SLIDE 20

Monte Carlo methods for sampling-based Stochastic Optimization The Wang-Landau algorithm A toy example

A toy example (3/3)

Path of the x1-component of (Xt)t, when Xt is the WL chain (left) and the Hastings-Metropolis chain (right).

2 4 6 8 10 12 x 10

4

−2 −1.5 −1 −0.5 0.5 1 1.5 2 beta=4 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 x 10

6

−2 −1.5 −1 −0.5 0.5 1 1.5 2 beta=4

Figure: [left] Wang Landau, T = 110 000. [right] Hastings Metropolis, T = 2 106 ; the red line is at x = 110 000

slide-21
SLIDE 21

Monte Carlo methods for sampling-based Stochastic Optimization The Wang-Landau algorithm Approximation of π

Approximation of π (1/2)

Definition of the proposal distribution − Sample points approximating the proposal distribution − Compute the associated weights Approximate the target π

slide-22
SLIDE 22

Monte Carlo methods for sampling-based Stochastic Optimization The Wang-Landau algorithm Approximation of π

Approximation of π (2/2)

It is expected π⋆ ≈ 1 T

T

  • t=1

δXt lim

t θt =

π(X1) Zπ , · · · , π(Xd) Zπ

  • In addition, by definition of π⋆

x ∈ Xi = ⇒ π(x) Zπ π⋆(x) = d lim

t θt(i)

Theis yields the algorithm (step 2)

  • f dπ

Zπ ≈ d T

T

  • t=1

f(Xt) d

  • i=1

θt(i)1 IXi(Xt)

slide-23
SLIDE 23

Monte Carlo methods for sampling-based Stochastic Optimization The Wang-Landau algorithm Efficiency of the Wang-Landau algorithm

Efficiency of Wang-Landau (1/3)

We compare the efficiency of the algorithm based on their ability to jump from one mode to another mode in a short time. Not possible to explicitly compute this time for general problems. We therefore considered toy examples. We report the results for a very simple example, for which explicit computations of the hitting-time is possible.

slide-24
SLIDE 24

Monte Carlo methods for sampling-based Stochastic Optimization The Wang-Landau algorithm Efficiency of the Wang-Landau algorithm

Efficiency of Wang-Landau (2/3)

State space : X = {1, 2, 3} Target distribution : π(1) ∝ 1 π(2) ∝ ǫ π(3) ∝ 1 Let us compare

1

Hastings-Metropolis P with proposal kernel Q and target π Q =   2/3 1/3 1/3 1/3 1/3 1/3 2/3  

2

Wang-Landau where the kernels Pθ are Hastings-Metropolis kernels with proposal Q and target πθ

1 2 3 ε/3 ε/3 ε/3 1/3 1/3 1/3

1 - ε/3 1-ε/3

1 2 3

  • (ε)
  • (ε)

1/3 1/3 1/3

1 - o(ε) 1-o(ε)

(left) Transition matrix P ; (right) Behavior of the transition matrix Pθ when ǫ → 0 (θ fixed)

slide-25
SLIDE 25

Monte Carlo methods for sampling-based Stochastic Optimization The Wang-Landau algorithm Efficiency of the Wang-Landau algorithm

Efficiency of Wang-Landau (3/3)

Comparison based on the hitting time T1→3 : hitting-time of state 3, given the chain started from state 1 and how it behaves when ǫ → 0. When ǫ → 0, we obtain T1→3 scales like For Hastings-Metropolis : 6/ǫ

lim ǫ→0 ǫ 6 E [T1→3] = 1 ǫ 6 T1→3 → E(1) in distribution

For Wang-Landau applied with γt = γ⋆/ta and 1/2 < a < 1 : C(a, γ⋆) | ln ǫ|1/(1−a) For Wang-Landau applied with γt = γ⋆/t ǫ−1/(1+γ⋆)

slide-26
SLIDE 26

Monte Carlo methods for sampling-based Stochastic Optimization Convergence issues

Outline

Introduction The Wang-Landau algorithm Convergence issues Adaptive and Interacting MCMC Sufficient conditions for the convergence Convergence of Wang-Landau Combining WL and simulated annealing

slide-27
SLIDE 27

Monte Carlo methods for sampling-based Stochastic Optimization Convergence issues Adaptive and Interacting MCMC

Adaptive and Interacting MCMC (1/2)

In the two previous examples, the conditional distribution of the points (Xt)t satisfies E [h(Xt+1)|pastt] =

  • h(y) Pθt(Xt, dy)

where Pθ is a Markov transition kernel (θt)t is a random process. Even in the simple situation when there exists π such that πPθ = π for any θ and lim

n P n θ (x, ·) − π = 0

Is π the stationary distribution of the process (Xt)t ?

slide-28
SLIDE 28

Monte Carlo methods for sampling-based Stochastic Optimization Convergence issues Adaptive and Interacting MCMC

Adaptive and Interacting MCMC (2/2)

Consider the following adapted chain : Fix t0, t1 ∈ (0, 1). Define an adapted chain as follows : Xk+1|pastk ∼ Pt0(Xk, ·) if Xk = 0 Pt1(Xk, ·) if Xk = 1 where

Ptℓ =

  • 1 − tℓ

tℓ tℓ 1 − tℓ

  • Pt0 and Pt1 both converge to the stationary distribution

π = 1/2 1/2

  • .

Then, (Xk)k is a Markov chain, with transition matrix 1 − t0 t0 t1 1 − t1

  • and it converges to the distribution

˜ π∝ t1 t0

  • = π.
slide-29
SLIDE 29

Monte Carlo methods for sampling-based Stochastic Optimization Convergence issues Adaptive and Interacting MCMC

Adaptive and Interacting MCMC (2/2)

Consider the following adapted chain : Fix t0, t1 ∈ (0, 1). Define an adapted chain as follows : Xk+1|pastk ∼ Pt0(Xk, ·) if Xk = 0 Pt1(Xk, ·) if Xk = 1 where

Ptℓ =

  • 1 − tℓ

tℓ tℓ 1 − tℓ

  • Pt0 and Pt1 both converge to the stationary distribution

π = 1/2 1/2

  • .

Then, (Xk)k is a Markov chain, with transition matrix 1 − t0 t0 t1 1 − t1

  • and it converges to the distribution

˜ π∝ t1 t0

  • = π.

Adaption cas destroy the convergence !

slide-30
SLIDE 30

Monte Carlo methods for sampling-based Stochastic Optimization Convergence issues Sufficient conditions for the convergence

Sufficient conditions for the convergence (1/3)

The literature provides sufficient conditions so that convergence in distribution of (Xt)t Strong law of large numbers for (Xt)t Central Limit Theorem for (Xt)t

G.O. Roberts, J.S. Rosenthal. Coupling and Ergodicity of Adaptive Markov chain Monte Carlo algorithms. J. Appl. Prob. (2007)

  • G. Fort, E. Moulines, P. Priouret. Convergence of adaptive MCMC algorithms : ergodicity and law of large numbers. Ann. Stat.

2012

  • G. Fort, E. Moulines, P. Priouret and P. Vandekerkhove. A Central Limit Theorem for Adaptive and Interacting Markov Chain.

Bernoulli, 2013.

slide-31
SLIDE 31

Monte Carlo methods for sampling-based Stochastic Optimization Convergence issues Sufficient conditions for the convergence

Sufficient conditions for the convergence (2/3)

Xt-N θt-N Xt-N+1 θt-N+1 Xt θt

Time t-N Time t-N+1 Time t

Kernel Pθt-N (Xt-N, ) Kernel Pθt-1 (Xt-1, )

Adaptive MCMC Frozen chain

Xt-N Xt

Kernel Pθt-N iterated N times

Xt

In the limit Under πθt-N

slide-32
SLIDE 32

Monte Carlo methods for sampling-based Stochastic Optimization Convergence issues Sufficient conditions for the convergence

Sufficient conditions for the convergence (3/3)

E

  • h(Xt)|pastt−N
  • h(y) πθ⋆ (dy) = E
  • h(Xt)|pastt−N
  • h(y) P N

θt−N (Xt−N , dy)

+

  • h(y) P N

θt−N (Xt−N , dy) −

  • h(y) πθn−N (dy)

+

  • h(y) πθn−N (dy) −
  • h(y) πθ⋆ (dy)

Diminishing adaption condition Roughly speaking : dist(Pθ, Pθ′) ≤ dist(θ, θ′) If θt − θt−1 are close, then the transition kernels Pθt and Pθt−1 are close also. Containment condition Roughly speaking : lim

N→∞ dist(P N θ , πθ) = 0

at some rate depending smoothly on θ. Regularity in θ of πθ so that lim

t θt = θ⋆ =

⇒ dist (πθt − πθ⋆) → 0

slide-33
SLIDE 33

Monte Carlo methods for sampling-based Stochastic Optimization Convergence issues Convergence of Wang-Landau

Convergence of Wang-Landau

Fort et al. (2014) provide sufficient conditions on the transition kernels Pθ the target distribution π the step size (γt)t controlling the adaption rate of the weight sequence implying that almost-surely lim

t θt =

π(X1) Zπ , · · · , π(Xd) Zπ

  • lim

T

1 T

T

  • t=1

f(Xt) =

  • f dπ⋆

lim

T

d T

T

  • t=1

f(Xt)

d

  • i=1

θt(i)1 IXi(Xt) =

  • f dπ

Zπ The rate of convergence of (θt)t is also established.

slide-34
SLIDE 34

Monte Carlo methods for sampling-based Stochastic Optimization Combining WL and simulated annealing

Outline

Introduction The Wang-Landau algorithm Convergence issues Combining WL and simulated annealing

slide-35
SLIDE 35

Monte Carlo methods for sampling-based Stochastic Optimization Combining WL and simulated annealing

WL and simulated annealing (1/2)

Liang, Cheng and Lin. Simulated Stochastic Approximation Annealing for Global Optimization with a Square-Root-Cooling

  • Schedule. JASA (2014)

Let a cooling schedule (Tt)t such that limt ↓ Tt = T⋆ > 0. Choose Xi = {x : ui−1 < U(x) ≤ ui}. Set πT,θ(x) ∝

d

  • i=1

1 θ(i) exp

  • −U(x)

T

  • 1

IXi(x) Algorithm : repeat

  • Given the past, draw Xt under a transition kernel PTt,θt with invariant

distribution πTt,θt.

  • Update the weight parameter θt as in the WL algorithm.
slide-36
SLIDE 36

Monte Carlo methods for sampling-based Stochastic Optimization Combining WL and simulated annealing

WL and simulated annealing (2/2)

Results Liang et al. (2014)

1

Law of large numbers : a.s. lim

N

1 N

N

  • t=1

f(Xt) =

  • f(x)πT⋆,θ⋆(x)

Z⋆ dλ

2

Let u⋆ be the minimal value, necessarily reached in stratum X1 lim

t→∞ P (U(Xt) ≤ u⋆ + ǫ|Xt ∈ X1) = P (U(Y ) ≤ u⋆ + ǫ|Y ∈ X1)

where Y ∼ exp(−U(x)/T⋆).

3

“When T⋆ → 0, P (U(Y ) ≤ u⋆ + ǫ|Y ∈ X1) → 1, thus showing the convergence of the algorithm to the minima of U(x).” But we could imagine other methods to combine WL and Simulates Annealing, or WL and stochastic optimization