A regenerative modification of the multilevel splitting Alexandra - - PowerPoint PPT Presentation

a regenerative modification of the multilevel splitting
SMART_READER_LITE
LIVE PREVIEW

A regenerative modification of the multilevel splitting Alexandra - - PowerPoint PPT Presentation

A regenerative modification of the multilevel splitting Alexandra Borodina and Evsey Morozov Institute of Applied Mathematical Research, Russian Academy of Sciences and Petrozavodsk University, Russia Supported by Russian Foundation for Basic


slide-1
SLIDE 1

A regenerative modification of the multilevel splitting Alexandra Borodina and Evsey Morozov Institute of Applied Mathematical Research, Russian Academy of Sciences and Petrozavodsk University, Russia Supported by Russian Foundation for Basic Research, Grant 07-07-00088.

1

slide-2
SLIDE 2

1 Introduction

The multilevel splitting (MS): huge simulation time reduction vs crude Monte Carlo (CMC); statistical properties of the estimators are not studied in detail. Known: – Markov queues: the overflow probability estimate is unbiased [Garvels]; – consistency and asymptotic normality in adaptive MS [Cerou, Guyader, 2005].

2

slide-3
SLIDE 3

For a regenerative queue: splitted processes as a family of (dependent) regenerative cycles ⇒ regenerative MS. We estimate the probability of overflow within a cycle for: I.1) workload process in M/G/1 (and GI/G/1); I.2) non-Markov queue using an embedded Markov chain (MC);

  • II. A stationary probability to exceed a level.

As a result: a randomization of the thresholds and a change of algorithm.

3

slide-4
SLIDE 4

2 Preliminaries

  • I. A Markov queueing process X = {Xn, n ≥ 1}: state space E;

MS is (typically) used to estimate the probability γc = P(X hits set A before 0)

  • 1. Given thresholds 0 < L1 < · · · < LM ≡ L;
  • 2. Partition of E : A = CM ⊂ · · · ⊂ C1: Ci = {x ∈ E : x ≥ Li};
  • 3. Ri copies upon reaching Li; termination at L.

4

slide-5
SLIDE 5

Then γc = p1p2 · · · pM with unbiased estimate γ(m) =

M

  • i=1

ˆ pi = N mR1 · · · · · RM ; (1) estimate ˆ pi of pi = P(Ci|Ci−1) is evaluated by CMC; N = #{ reaching A = [L, ∞)}; m = the number of i.i.d. runs starting at 0. (More details: [5, 6, 7, 16].)

  • II. A process Xn, n ≥ 1, with regeneration instants:

T(0) = 0, T(n + 1) = min(k > T(n) : Xk = 0), n ≥ 0; (2)

5

slide-6
SLIDE 6

the same distribution of Xn+T(k), n ≥ 0, for each k ≥ 1 and independent of the Xn, n < T(k); i.i.d. regeneration cycles Gn = (Xk, T(n − 1) ≤ k < T(n)); i.i.d. cycle periods βn = T(n) − T(n − 1). We estimate the cycle overflow probability γc = P( max

1≤n<β Xn ∈ A),

(3)

6

slide-7
SLIDE 7

and, if weak limit Xn ⇒ X exists, the stationary probability γs = P(X ∈ A), (4) for queue size and for workload in M/G/1 queue. (A generic element with no index.)

  • I. Estimating γc (under stability condition). Consistency: i.i.d indicators Ik =

I( max

T(k)≤n<T(k+1) Xn ∈ A);

γc(n) = 1 n

n

  • k=1

Ik → EI = γc; (5) asymptotical normality: by regenerative CLT.

7

slide-8
SLIDE 8
  • II. Estimating γs: positive recurrence, Eβ < ∞; (dependent) indicators Ik =

I(Xk ∈ A) and i.i.d. variables Yi =

T(i)−1

  • k=T(i−1)

Ik, i ≥ 1. (6) Since Y ≤ β, by regenerative theory, γs(n) = 1 n

n

  • k=1

Ik → EY Eβ = γs, n → ∞, (7) the time average EY/Eβ = P(X ∈ A) if weak limit X exists (e.g. if β is aperiodic).

8

slide-9
SLIDE 9

Confidence 100(1 − δ)% interval for γs (by a regenerative CLT):

  • γn − zδ
  • vs(n)

√n , γn + zδ

  • vs(n)

√n

  • ,

(8) P(N(0, 1) ≤ zδ) = 1 − δ/2; where empirical variance vs(n) = 1 n n

i=1(Yi − γs(n)βi)2

β

2 n

⇒ σ2

s ≡ V ar(Y − βγ),

(9) (βn = sample mean cycle length.) A minimal condition for the convergence: E(Y − γβ)2 < ∞.

9

slide-10
SLIDE 10

(Under EY 2 < ∞, Eβ2 < ∞, vn is strongly consistent [8].) Applicability to dependent cycles [1, 12, 15]: below D-groups of dependent cycles with D = R1 × · · · × RM.

  • III. M/G/1 queue: input rate λ, service time S, distribution F, stability: ρ =

λES < 1. The queue size at the departure instants: a positive recurrent (aperiodic) Markov chain (MC): νn+1 = (νn − 1)+ + ∆n, n ≥ 0, (ν0 = 0), (10)

10

slide-11
SLIDE 11

∆n = #{arrivals} during service of customer n + 1. Waiting time (workload) Markov chain: Wn, n ≥ 1; Wn ⇒ W exists. If distribution of W is analytically available [Ross,Seshadri] [13]: γc = P( max

1≤n<β Wn ≥ L) = 1 − P(W + S ≤ L)

P(W ≤ L) . (11) M/M/1: ρ = λ/µ < 1, P(W ≤ x) = 1 − ρe−(λ−µ)x: γc = (1 − ρ)e−(µ−λ)L 1 − ρe−(µ−λ)L , (12)

11

slide-12
SLIDE 12

while (applying technique from the birth-death processes) P( max

1≤n<β νn ≥ L) = ρL−1 − ρL

1 − ρL . (13) NOTE 1. M/M/1 has been used to justify the quality of MS [5, 6, 7, 9, 10, 14]. If service time has subexponential integrated-tail distribution Fe(x) = 1 ES ∞

x

F(y)dy, (F = 1 − F) and lim

x→∞

F(xey/√x) F(x) = 1, locally uniformly in y, ∈ R, (14)

12

slide-13
SLIDE 13

(holds for Pareto distribution) then stationary queue ν satisfies [2] γs = P(ν ≥ L) ∼ ρ 1 − ρFe(L − 1 λ ), L → ∞. (15)

  • 3. Splitting and regenerations in M/G/1

by splitting: dependent regeneration (sub)cycles: ν(i) = {ν(i)

n , τi−1 ≤ n < τi}, τ0 = 0, i = 1, . . . , mD (D = R1 × · · · × RM)

with periods αi = τi−τi−1. D-group: (ν(i) : kD < i ≤ (k+1)D), k = 0, . . . , m−1.

13

slide-14
SLIDE 14
  • Fig. 1: splitted processes with termination instants b, c, d (case a)) and cycles with

regenerations τn pasted together with common pre-history, with no idle-time, case b).

(t) L2 L1 L0 t a b c d SetA (t) L2 L1 L0 t

  • SetA

a) b)

14

slide-15
SLIDE 15
  • 1. Estimating γc = P(max1≤n<β νn ∈ A):

1.a) Markov queue: standard MS, termination upon reaching A with (at most D- dependent) indicators Ii = I( max

τi−1≤n<τi ν(i) n ∈ A), i = 1, . . . , mD.

(16) To compensate overestimation if a process starting at Li reaches 0 before Li+1: take extra Ri+1Ri+2 · · · RM (virtual) cycles, i = 1, . . . , M − 1. 1.b) Non-Markov M/G/1 queue, to keep independence after splitting: use MC

15

slide-16
SLIDE 16

(10): makes transitions between regions Gi = [Li, Li+1) ⇒: A randomization of the thresholds and modification of the algorithm: Condition A: Under transition y ∈ Gi → x ∈ Gk, k > i (crossing the thresholds Li+1, . . . , Lk), generate Ri+1 · · · Rk processes at state x, i = 0, . . . , M − 1; k = i + 1, . . . , M (RM = 1) .

  • 2. Estimating γs = P(ν ≥ L) (weak limit νn ⇒ ν exists): evaluate the time

(=number of arrivals) the process is in A using indicators Ii,n = I(ν(i)

n ∈ A).

Apply condition A evaluating the number of arrivals in all processes.

16

slide-17
SLIDE 17

For Wn, n ≥ 1 with jumps Xn(= ±1): the same randomization. NOTE 2. By PASTA: estimating overflow probability for continuous time. Randomization is widely used, [3, 11].

  • 4. Consistency and asymptotic normality
  • 1. Consistency: estimate γ(m)= regenerative estimate (5) based on n = mD

dependent indicators (16), so: 1 n

n

  • i=1

Ii → E D

i=1 Ii

D = γc. (17)

17

slide-18
SLIDE 18

NOTE 2. No need n = mD for the convergence.

  • 2. Estimating γs = P(ν ≥ L): cycle structure, Fig 2.

L0 t

  • 1

2 D

1

=T mD

m

=T 1 2 3 D 1 m mD D+1 2 ... ... .. . ... ... ...

Variables Zi =

τi−1

  • k=τi−1

I(ν(i)

k ≥ L),

(18)

18

slide-19
SLIDE 19

belonging to the same D-group are dependent; i.i.d variables Yi =

iD

  • j=(i−1)D+1

Zj, i = 1, . . . , m; (19) the length of the ith D-group is βi =

iD

  • j=(i−1)D+1

αj, so βi = T(i) − T(i − 1), i = 1, . . . , m. Since Y ≤ β =st α1 + · · · + αD (see Fig. 2) and Eα < ∞ by ρ < 1, then EY < ∞,

19

slide-20
SLIDE 20

and (by regenerative theory) the estimate γs(n) = n

i=1 Zi

n

i=1 αi

→ EY Eβ = P(ν ≥ L) ≡ γs, n → ∞ (20) is consistent. The limit νn ⇒ ν exists in the system M/G/1. NOTE 4. To estimate γs in discrete-time setting one can choose any discrete- event scale because, for each cycle, the number of arrivals equals the number

  • f departures.

An asymptotic normality: by regenerative interpretation.

20

slide-21
SLIDE 21

Applicability of CLT for estimating γs = P(ν ≥ L): assumption ES2 < ∞ implies Eα2 < ∞, Wolff [20]. Since β =st α1 + · · · + αD then Eβ2 < ∞. Similarly, consistency and asymptotic normality of the estimates for the workload. In confidence interval (8): variance includes Yi depending on target probability.

21

slide-22
SLIDE 22
  • 5. Simulation results
  • Fig. 3 estimate of γs = P(ν ≥ L) in M/Pareto/1 with service time distribution

F(x) = 1/x4, x ≥ 1(= 1, x ≤ 1), (21) input rate λ = 0.45 (ρ = 0.6), NM+1 = 106.

22

slide-23
SLIDE 23

1e-06 2e-06 3e-06 4e-06 5e-06 6e-06 7e-06 8e-06 9e-06 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 L Asymptotic Splitting (regenerative) Splitting (regenerative, EMC) 23

slide-24
SLIDE 24

The simulation stopping rule: a given number N of the target events to be achieved. For instance, for L = 50 we take Li = {7, 14, 21, 28, 35, 42} (that is M = 6), N = 10000 and Ri ≡ 10. More details in Table 1: S=standard MS, EMC= embedded MC. The estimate γs(n) is (20), and ”asymp” = (15). (In all experiments the number of replications m = nD is widely varied in [103, 104].)

24

slide-25
SLIDE 25

Table 1 L asymp γs(n)[S] γs(n)[EMC] t[S] t[EMC] Var[S] Var[EMC] 31 1.53 · 10−6 2.98 · 10−7 1.53 · 10−6 18.5 24.2 1.01 · 10−8 2.40 · 10−10 36 9.77 · 10−7 3.49 · 10−8 1.78 · 10−7 18.7 34.3 7.20 · 10−9 7.33 · 10−11 40 7.12 · 10−7 9.21 · 10−9 2.35 · 10−7 17.2 45.3 1.49 · 10−10 6.15 · 10−11 45 5.0 · 10−7 4.46 · 10−8 1.24 · 10−7 19.7 38.6 2.56 · 10−8 5.77 · 10−10 50 3.65 · 10−7 2.24 · 10−8 3.59 · 10−7 51.4 91.1 4.20 · 10−11 3.54 · 10−13

25

slide-26
SLIDE 26

EMC: an increasing of simulation time t and HUGE variance reduction. NOTE: a consistency between EMC simulation and asymptotic formula (15).

26

slide-27
SLIDE 27
  • Fig. 4: estimation of P( max

1≤n<β Wn ≥ L) in M/M/1 (with ρ = 0.6).

γc(EMC) = 3.12 · 10−11, 4.09 · 10−12, 4.49 · 10−13, (12): γc = 1.51 · 10−11, 2.04 · 10−12, 2.77 · 10−13.

1e-07 2e-07 3e-07 4e-07 5e-07 6e-07 40 45 50 55 60 65 70 75 80 85 L Analytical solution Splitting

27

slide-28
SLIDE 28
  • Fig. 5: estimate of γc = P( max

1≤n<β Wn ≥ L) in M/Pareto/1, service time distribution (21).

CMC vs EMC splitting , ρ = 0.6, N[EMC] = 1000, N[CMC] = 100

5e-07 1e-06 1.5e-06 2e-06 2.5e-06 3e-06 3.5e-06 4e-06 55 60 65 70 75 80 85 90 95 100 105 110 115 120 125 130 135 140 145 150 155 160 165 L Monte Carlo Regenerative Splitting 28

slide-29
SLIDE 29
  • Fig. 6: estimate of γc = P( max

1≤n<β Wn ≥ L) in Pareto/Pareto/1, input distribution P(u ≥ x) =

1/x3, x ≥ 1(= 1 x ≤ 1), service time distribution (21) (ρ = 8/9) EMC vs CMC.

1e-05 2e-05 3e-05 4e-05 5e-05 6e-05 7e-05 8e-05 50 55 60 65 70 75 80 L Monte Carlo Regenerative Splitting 29

slide-30
SLIDE 30

NOTE: i) a consistency between EMC and CMC; ii) more unstable CMC estimate; iii) LRD stationary workload: ES3 < ∞, ES4 = ∞ [4, 12]. In all experiments, the cycles from the same D-group are highly dependable.

30

slide-31
SLIDE 31

3 Concluding remarks

  • 1. Extension of experiments;
  • 2. Source of variance reduction (Markovity)?
  • 3. Selection of the thresholds;
  • 4. Combination randomization with given thresholds;
  • 5. More...

31

slide-32
SLIDE 32

References

[1] S. Asmussen. Applied Probability and Queues, 2nd ed., Springer, NY, 2003. [2] S. Asmussen, C. Kl¨ uppelberg, K. Sigman. Sampling at subexponential times, with queueing applications. Stochas- tic Process. Appl. 79 ,1999, 265-286. [3] F.Cerou, A. Guyader. Adaptive multilevel splitting for rare event analysis. INRIA, Research report No 5710, Oct. 2005. [4] D. J. Daley. The serial correlation coefficients of waiting times in a stationary single server queue. J.Aust.Math.Soc. 8, 1968, 683-699. [5] M. Garvels. PhD Thesis The splitting method in rare event simulation. The University of Twente, The Nether-

32

slide-33
SLIDE 33

lands May, 2000. [6] P. Glasserman, P. Heidelberger, P. Shahabuddin, and T. Zajic. A look at multilevel splitting. In H. Niederreiter (ed.): Monte Carlo and Quasi Monte Carlo Methods, Lecture Notes in Statistics, vol. 127, 1996, 99-108, Springer-Verlag. [7] P. Glasserman, P. Heidelberger, P. Shahabuddin, and T. Zajic. Splitting for rare event simulation: analysis of simple cases. Proceedings of the 1996 Winter Simulation Conference, 1996, 302-308. [8] P. W. Glynn, D. L. Iglehart. Conditions for the applicability of the regenerative method. Management Science 39, 1993, 1108-1111. [9] P. E. Heegaard. A survey of Speedup simulation techniques. Workshop tutorial on Rare Event Simulation, Aachen, Germany, 1997.

33

slide-34
SLIDE 34

[10] P. Heidelberger. Fast simulation of rare events in queuieng and relaibility models. Performance Evaluation of Computers and Communications Systems Springer-Verlag, LN in Computer Sci., 729, 1993, 165-202. [11] S.Ermakov, V.Melas Mathematical experiment with complex stochstic models, S.-Petersburg University, 1993. (In Russian) [12] E. Morozov. Weak regeneration in modeling of queueing processes. Queueing Systems, 46, 2004, 295-315. [13] S. M. Ross, S. Seshadri. Hitting Time in an M/G/1 Queue. J. Appl. Prob., 36, 1999, 934-940. [14] P. Shahabuddin. Rare event simulation in stochastic models. Proceedings of the WSC 1995, IEEE Press., 178-185 1995. [15] K. Sigman, R. Wolff. A review of regenerative processes. SIAM Review, 35, No. 2, 1993, 269-288.

34

slide-35
SLIDE 35

[16] M. Villen-Altamirano, J. Villen-Altamirano. RESTART: A Method for Accelerating Rare Event Simulations. Proceedings of the 13-th International Teletraffic Congress, Queueing, Perform-ance and Control in ATM, 1991, 71-76. [17] M. Villen-Altamirano, J. Villen-Altamirano. RESTART: A Straightforward Method for Fast Simulation of Rare

  • Events. Proceedings of the 1994 Winter Simulation Conference, 1994, 282-289.

[18] M. Villen-Altamirano, J. Villen-Altamirano. About the Efficiency of RESTART. Proceedings of the RESIM’99 Workshop, University of Twente, The Netherlands, 1999, 99-128. [19] M. Villen-Altamirano, J. Villen-Altamirano. Analysis of RESTART Simulation: Theoretical Basis and Sensitivity

  • Study. European Transactions on Telecommunications vol. 13, No. 4, 2002, 373-385.

[20] R.W. Wolff. Stochastic Modeling and the Theory of Queues, Prentice-Hall, 1989.

35

slide-36
SLIDE 36

4 APPENDIX

Example A. γs = P(ν ≥ L) in M/Pareto/1, ρ = 0.6 Standard S vs EMC for L small; EMC: consistency with asymptotic formula; EMC: less variance.

36

slide-37
SLIDE 37

Table A L asymp ˆ γs[S] ˆ γs[EMC] t[S] t[EMC] Var[S] Var[EMC] 10 4.55 · 10−5 4.11 · 10−4 4.41 · 10−4 2.1 1.6 1.85 · 10−9 1.56 · 10−9 15 1.35 · 10−5 9.77 · 10−5 2.23 · 10−5 4.2 3.2 5.32 · 10−9 5.63 · 10−10 20 5.69 · 10−6 3.52 · 10−6 4.14 · 10−6 3.2 3.6 1.41 · 10−7 8.21 · 10−9 25 2.91 · 10−6 6.67 · 10−6 2.32 · 10−6 3.2 5.3 7.54 · 10−9 2.04 · 10−10 30 1.68 · 10−6 5.23 · 10−7 1.03 · 10−6 11.4 10.9 4.21 · 10−9 4.74 · 10−11

37

slide-38
SLIDE 38

Example A. (continued) Table B: M (Ri) - number of thresholds (copies); D= number of dependent paths; N=number of rare events to be achieved=stopping rule; N[S],N[EMC] =actual number of reaching set A m= number of the i.i.d. trials.

38

slide-39
SLIDE 39

Table B L M(Ri) D N N[S] N[EMC] m[S] m[EMC] 10 1(10) 10 10000 10051 10131 491749 520347 15 2(10) 100 10000 17882 11265 367591 1145046 20 3(10) 1000 10000 10028 10176 71631 954783 25 4(10) 10000 10000 13587 129396 117678 451050 30 5(10) 100000 100000 252883 188500 971845 1251389

39

slide-40
SLIDE 40

Example B. CMC vs EMC: γs = P(ν ≥ L) in M/Pareto/1.

2e-07 4e-07 6e-07 8e-07 1e-06 1.2e-06 1.4e-06 1.6e-06 1.8e-06 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 L Crude Monte Carlo Sptitting (at departure moments)

40

slide-41
SLIDE 41

Example B(continued): simulation time (sec) S vs EMC

100 200 300 400 500 600 700 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 L time [regenerative S] time [regenerative, EMC] 41