A variance reduction method for computing VaR 1. Computing Value at - - PowerPoint PPT Presentation

a variance reduction method for computing var
SMART_READER_LITE
LIVE PREVIEW

A variance reduction method for computing VaR 1. Computing Value at - - PowerPoint PPT Presentation

A variance reduction method for computing VaR 1. Computing Value at Risk by Monte Carlo simulations 2. Importance Sampling for variance reduction 3. Interacting Particle Systems for Importance Sampling (IPS-IS) 4. Simulation results Nadia


slide-1
SLIDE 1

A variance reduction method for computing VaR

  • 1. Computing Value at Risk by Monte Carlo simulations
  • 2. Importance Sampling for variance reduction
  • 3. Interacting Particle Systems for Importance Sampling (IPS-IS)
  • 4. Simulation results

Nadia Oudjane - EDF R&D- Journ´ ees MAS 2008

1

slide-2
SLIDE 2
  • 1. Computing Value at Risk by Monte Carlo simulations

2

Value at Risk and quantile

◮ P&L of a portfolio on [0, T] ∆V (X) = VT − V0 + T

0 CF

with

X ∈ Rd

the risk factors impacting the portfolio value on [0, T]

◮ Value at Risk V aRα = | inf{s ∈ R | P(∆V ≤ s) ≥ 1 − α} | V aRα = |F −(α)| ,

where

F(s) = P(∆V ≤ s) ,

for all s ∈ R

  • N. Oudjane - EDF R&D- Journ´

ees MAS 2008 2

slide-3
SLIDE 3
  • 1. Computing Value at Risk by Monte Carlo simulations

3

Monte Carlo method for VaR estimation

◮ The distribution function F can be viewed as an expectation F(s) = E[I∆V (X)≤s] ,

for all

s ∈ R ◮ Traditional Monte Carlo Method for computing VaR

  • 1. Monte Carlo simulations give an approximation of F(s) :

ˆ FN(s) = 1 N

N

  • i=1

I(∆V (Xi) ≤ s) ,

for all s ∈ R

⇒ Too many evaluations of ∆V for a given accuracy

  • 2. Inversion of ˆ

F N and interpolation for approximating VaR

  • N. Oudjane - EDF R&D- Journ´

ees MAS 2008 3

slide-4
SLIDE 4
  • 2. Importance Sampling for variance reduction

4

Importance Sampling for variance reduction

◮ Change of measure p − → q

where q dominates Hp

m = Ep[H(X)] = Eq[H(Y )p q (Y )] ,

where

X ∼ p

and

Y ∼ q ◮ Optimal change of measure p − → q∗

achieves zero variance if H ≥ 0

q∗ = Hp

  • H(x)p(x) dx =

Hp Ep[H(X)] = H · p ◮ Monte Carlo approximation Ep[H(X)] ≈ ˆ mq

M = 1

M

M

  • i=1

H(Yi)p q (Yi) ,

where

(Y1, · · · , YM)

i.i.d. ∼ q

⇒ How to simulate and evaluate approximately q∗ ?

  • N. Oudjane - EDF R&D- Journ´

ees MAS 2008 4

slide-5
SLIDE 5
  • 2. Importance Sampling for variance reduction

5

Variance of the Importance Sampling estimate

◮ Let q be a (possibly random) importance probability density dominating q∗ V ar( ˆ mq

M) = E

  • V ar[ ˆ

mq

M | Fq]

  • + V ar
  • E[ ˆ

mq

M | Fq]

  • =0

Fq denotes the sigma-algebra generated by the random variables involved in q ◮ The variance of the IS estimate depends on the ”distance” between q and q∗ V ar( ˆ mq

M) = m2

M E [(q∗ − q)q∗ q ](x)dx

  • ◮ Idea : use Interacting Particle Systems for Importance Sampling (IPS-IS) to

approximate q∗ by qN based on an N-particle system to achieve

V ar( ˆ mqN

M ) ≤

C MN α

with

0 < α < 1/2

  • N. Oudjane - EDF R&D- Journ´

ees MAS 2008 5

slide-6
SLIDE 6
  • 2. Importance Sampling for variance reduction

6

Some alternative approaches

◮ Large deviation approximation for rare events simulation ◮ Approximation of H to obtain a simple form fo q∗

ex : [Glasserman&al00] for computing VaR, ∆-Γ approximation of the portfolio

◮ Cross-entropy [Homem-de-Mello&Rubinstein02] qθ is chosen in a parametric family such as to minimize the entropy K(qθ, q∗) ◮ Interacting Particle Systems whithout Importance Sampling

[DelMoral&Garnier05], [Cerou&al06] Interacting Particle Systems for Importance Sampling (IPS-IS) can be viewed as a non parametric version of cross entropy approach

  • N. Oudjane - EDF R&D- Journ´

ees MAS 2008 6

slide-7
SLIDE 7
  • 2. Importance Sampling for variance reduction

7

Progressive correction [Musso&al01]

◮ We introduce a sequence of non negative functions (Gk)0≤k≤n such that

for all

x ∈ Rd ,        G0(x) = 1

The product

G0(x) · · · Gn(x) = H(x)

If

Gk(x) = 0

then

Gk+1(x) = 0 ◮ In our case H(x) = I(∆V (x) ≤ s)

then we choose

Gk(x) = I∆V (x)≤sk ,

with

s = sn ≤ · · · ≤ s0 = +∞ ◮ Dynamical system on the space of probability measures (νk)0≤k≤n      ν0 = p dx νk = Gkνk−1

  • Rd Gk(x)νk−1(x) dx = Gk · νk−1 ,

for all

1 ≤ k ≤ n ⇒ νn = q∗ dx

  • N. Oudjane - EDF R&D- Journ´

ees MAS 2008 7

slide-8
SLIDE 8
  • 2. Importance Sampling for variance reduction

8

Space exploration

◮ We introduce a sequence of Markov kernels (Qk)0≤k≤n

such that

νk ≈ νkQk

i.e.

νk(dx) ≈

  • Rd νk(du)Qk(u, dx) ,

for all

x ∈ Rd ◮ In our case where Gk(x) = I∆V (x)≤sk

, if p is Gaussian then Qk is easily obtained from a Gaussian kernel Q reversible for p,

Qk(x, dx′) = Q(x, dx′)I∆V (x)≤sk +

  • 1 − Q(x, ∆V −((−∞, sk]))
  • δx(dx′)

◮ Dynamical system on the space of probability measures (νk)0≤k≤n    ν0 = p dx νk = Gk · (νk−1Qk−1) ,

for all

1 ≤ k ≤ n ⇒ νn = q∗ dx

  • N. Oudjane - EDF R&D- Journ´

ees MAS 2008 8

slide-9
SLIDE 9
  • 3. Interacting Particle Systems for Importance Sampling (IPS-IS)

9

Approximation of the dynamical system

◮ The idea is to replace at each iteration k, νk−1Qk−1

by its N-empirical measure

SN(νk−1Qk−1)

such that

SN(νk−1Qk−1) = 1 N

N

  • i=1

δXi

k

where

(X1

k, · · · , XN k ) are i.i.d. ∼ νk−1Qk−1

◮ Dynamical system on the space of dicrete probability measures (νN

k )0≤k≤n

   νN

0 = SN(ν0)

νN

k = Gk · SN(νN k−1Qk−1) ,

for all

1 ≤ k ≤ n ⇒ One can show that νN

n ≈ q∗ dx

[DelMoral]

  • N. Oudjane - EDF R&D- Journes MAS 2008

9

slide-10
SLIDE 10
  • 3. Interacting Particle Systems for Importance Sampling (IPS-IS)

10

Algorithm

◮ Initialization : Generate independently (X1

0, · · · , XN 0 )

i.i.d.

∼ p

then set

νN

0 = 1

N

N

  • i=1

δXi ◮ Selection : Generate independently ( ˜ X1

k, · · · , ˜

XN

k )

i.i.d.

∼ νN

k = N

  • i=1

ωi

k δXi

k

◮ Mutation : Generate independently for each i ∈ {1, · · · , N}, Xi

k+1

∼ Qk( ˜ Xi

k, ·)

◮ Weighting : For each particle i ∈ {1, · · · , N}, compute ωi

k+1 =

Gk+1(Xi

k+1)

N

j=1 Gk+1(Xj k+1)

then set

νN

k+1 = N

  • i=1

ωi

k+1 δXi

k+1

  • N. Oudjane - EDF R&D- Journes MAS 2008

10

slide-11
SLIDE 11
  • 3. Interacting Particle Systems for Importance Sampling (IPS-IS)

11

Adaptive choice of the sequence (Gk)0≤k≤n

[Musso&al01], [Hommem-de-Mello&Rubinstein02], [C´ erou&al06]

◮ The performance of Interacting particle systems is known to deteriorate

when the quantities

max Gk SN(νN

k−1Qk−1)(Gk)

are big The idea is then to chose Gk such that 1

N

N

  • i=1

Gk(Xi

k) is not to small

◮ In our case where Gk(x) = I∆V (x)≤sk

, the threshold sk is chosen as a r.v. depending on the current particle system and on a parameter ρ ∈ (0, 1) :

sk = inf

  • s

such that

N

  • i=1

I∆V (Xi)≤s ≥ ρN

  • ◮ This choice of sk is not prooved to guarantee that the algorithms ends in a

finite number of iterations but this point does not seem to be a problem in our simulations

  • N. Oudjane - EDF R&D- Journes MAS 2008

11

slide-12
SLIDE 12
  • 3. Interacting Particle Systems for Importance Sampling (IPS-IS)

12

Density estimation

◮ At the end of the algorithm, we get νN

n ≈ q∗ dx

But Importance Sampling requires a smooth approximation with density qN

◮ Kernel of order 2 K K ≥ 0

  • K = 1
  • xi K = 0
  • |xi xj| K < ∞

◮ Rescaled kernel Kh Kh(x) = 1 hd K(x h) ◮ νN =

  • ωi δXi

Density estimation

− − − − − − − − − →

Kh∗ ·

qN,h =

  • ωi Kh(· − Xi)

◮ Optimal choice of h => EqN − q∗1 ≤

C N

4 2(d+4)

W1 W2 W3 W4 W5 <−−−−−−−− WEIGHTS <−−−−−−−− SAMPLE

<−−−−−−−− KERNELS <−−−−−−−− DENSITY ESTIMATE <−−−−−−−− SAMPLE

  • N. Oudjane - EDF R&D- Journes MAS 2008

12

slide-13
SLIDE 13
  • 3. Interacting Particle Systems for Importance Sampling (IPS-IS)

13

Some simulation results : Variance ratio

◮ Several test cases depending on the form of function x → ∆V (x) have

been studied : results are all comparable

◮ X is a d dimensional Gaussian variable and m = Ep[I∆V (X)≤s] ◮ Particles N = 500

Iterations n ≈ 10 to 60 Simulations M = 10 000

d = 1 d = 2 d = 3 d = 4 d = 5 m = 10−2 150 10−1 50 50 30 25 m = 10−3 1000 2 300 300 200 140 m = 10−6 2.105 200 105 400 105 300 5.104 460 2.104 480 d = 6 d = 7 d = 8 d = 9 · · · d = 30 m = 10−2 22 14 11 8 · · · 5.10−3 m = 10−3 100 70 55 40 · · · 10−3 m = 10−6 104 250 2.103 480 2.103 300 4.103 300 · · · 1 360

  • N. Oudjane - EDF R&D- Journes MAS 2008

13

slide-14
SLIDE 14
  • 3. Interacting Particle Systems for Importance Sampling (IPS-IS)

14

References

◮ [Glasserman&al00] Glasserman, P. and Heidelberger, P. and Shahabuddin, P.

Reduction Variance techniques for estimating Value at Risk, Management Science, Vol. 46, No 10, 2000.

◮ [DelMoral&Garnier05] Del Moral, P. and Garnier, J. Genealogical particle

analysis of rare events, Annals of Applied Probability, 2005.

◮ [Cerou&al06] Cerou, F. and Del Moral, P. and Le Gland, F. and Guyader, P. and

Lezaud, H. Topart, Some recent improvements to importance splitting, Proceedings of the 6th International Workshop on Rare Event Simulation, Bamberg, October 9-10, 2006.

◮ [Homem-de-Mello&Rubinstein02] Homem-de-Mello, T. and Rubinstein, R.Y.

Estimation of rare event probabilities using cross-entropy, Proceedings of the Winter Simulation Conference, 2002. ◮ [Musso&al01] Musso, C. and Oudjane,

  • N. and Le Gland, F. , Improving regularized particle filters, in Sequential Monte

Carlo Methods in Practice, Doucet A& al. editors, Statistics for Engineering and Information Science, 2001.

  • N. Oudjane - EDF R&D- Journes MAS 2008

14