Efficiency of the Cross-Entropy Method for Markov Chain Problems
Ad Ridder1 Bruno Tuffin2
1Vrije Universiteit, Amsterdam, Netherlands
aridder@feweb.vu.nl http://staff.feweb.vu.nl/aridder/
2INRIA, Rennes, France
bruno.tuffin@irisa.fr
Efficiency of the Cross-Entropy Method for Markov Chain Problems Ad - - PowerPoint PPT Presentation
Efficiency of the Cross-Entropy Method for Markov Chain Problems Ad Ridder 1 Bruno Tuffin 2 1 Vrije Universiteit, Amsterdam, Netherlands aridder@feweb.vu.nl http://staff.feweb.vu.nl/aridder/ 2 INRIA, Rennes, France bruno.tuffin@irisa.fr Rare
1Vrije Universiteit, Amsterdam, Netherlands
aridder@feweb.vu.nl http://staff.feweb.vu.nl/aridder/
2INRIA, Rennes, France
bruno.tuffin@irisa.fr
◮ Suppose that {An : n = 1, 2, . . .} is a family of events in a
◮ such that P(An) → 0 as n → ∞. ◮ Furthermore, suppose that P(An) is difficult to compute
◮ but easy to estimate by simulation.
◮ There might be many simulation algorithms. ◮ Denote by Yn the associated unbiased estimator of P(An).
n→∞
n]
n→∞
n]
◮ Specific: importance sampling with exponentially twisted
◮ More general: Dupuis and co-authors ([2], [3]) developed an
◮ More abstract: assume large deviations probabilities, i.e.,
n log P(An) = −θ, and derive conditions under which
◮ Importance sampling with a biasing scheme in a highly reliable
◮ Zero-variance approximation importance sampling in a highly
◮ Combination of conditioning and importance sampling for tail
◮ State-dependent importance sampling (based on zero-variance
◮ The cross-entropy method is a heuristic for rare event simulation
◮ A summary on one of the next slides. ◮ Then proving analytically the efficiency of the resulting estimator
◮ The usual approach is to estimate the efficiency by empirical
◮ Markov chain is {X(t) : t = 0, 1, . . .}. ◮ Statespace X; transition prob’s p(x, y). ◮ Markov chain starts off in a reference state 0, X(0) = 0. ◮ A ‘good’ set G ⊂ X of absorbing states. ◮ A failure set F ⊂ X of absorbing states. ◮ No other absorbing states. ◮ The time to absorption is T = inf{t > 0 : X(t) ∈ G ∪ F}. ◮ Absorption probabilities γ(x) = P(X(T) ∈ F|X(0) = x). ◮ Rare event A = 1{X(T) ∈ F} with probability
◮ Optimal change of measure Popt = P(·|A) gives Varopt[Y] = 0. ◮ Popt is feasible,
◮ not implementable, since it requires knowledge of the unknown
P∗∈P D(dPopt, dP∗),
z∈X N(x, z)
P∗∈P D(dPopt, dP∗)
P∗∈P
P∗∈P
k
n and Popt n
n , Pce n ) = o(log P(An))
n , Pce n ) = Eopt[log dPopt n /dPce n (X)] ≥ 0.
n
n
n
n
n
n
n
n
n
n]
dPopt
n
dPce
n (X)
dPopt
n
dPce
n (X)
n→∞
dPopt
n
dPce
n (X)
n→∞
dPopt
n
dPce
n (X)
n
dPce
n (X)
n
dPce
n (X)
n , Pce n )/ log P(An).
n]/ log P(An).
n (x, y).
n→∞
(x,y)∈X×X p(x,y)>0
n (x, y)
n|
n
n
◮ Cross-entropy coincides with zero-variance in rare-event
◮ Sufficient condition for logarithmic efficiency. ◮ Further investigations include verification of this condition. ◮ Probabilistic condition for strong efficiency. ◮ Further investigations to elaborate the strong efficiency and