Beta kernels and transformed kernels applications to copulas and - - PowerPoint PPT Presentation

beta kernels and transformed kernels applications to
SMART_READER_LITE
LIVE PREVIEW

Beta kernels and transformed kernels applications to copulas and - - PowerPoint PPT Presentation

Arthur CHARPENTIER, transformed kernels and beta kernels Beta kernels and transformed kernels applications to copulas and quantiles Arthur Charpentier Universit Rennes 1 arthur.charpentier@univ-rennes1.fr http


slide-1
SLIDE 1

✶✶♣t ✶✶♣t ◆♦t❡ ❊①❡♠♣❧❡ ❊①❡♠♣❧❡ ✶✶♣t Pr❡✉✈❡

Arthur CHARPENTIER, transformed kernels and beta kernels

Beta kernels and transformed kernels applications to copulas and quantiles

Arthur Charpentier

Université Rennes 1 arthur.charpentier@univ-rennes1.fr http ://freakonometrics.blog.free.fr/ Statistical seminar, Université Laval, Québec, April 2011 1

slide-2
SLIDE 2

Arthur CHARPENTIER, transformed kernels and beta kernels

Agenda

  • General introduction and kernel based estimation

Copula density estimation

  • Kernel based estimation and bias
  • Beta kernel estimation
  • Transforming observations

Quantile estimation

  • Parametric estimation
  • Semiparametric estimation, extreme value theory
  • Nonparametric estimation
  • Transforming observations

2

slide-3
SLIDE 3

Arthur CHARPENTIER, transformed kernels and beta kernels

Moving histogram to estimate a density

A natural way to estimate a density at x from an i.i.d. sample {X1, · · · , Xn} is to count (and then normalized) how many observations are in a neighborhood of x, e.g. |x − Xi| ≤ h, 3

slide-4
SLIDE 4

Arthur CHARPENTIER, transformed kernels and beta kernels

Kernel based estimation

Instead of a step function 1(|x − Xi| ≤ h) consider so smoother functions,

  • fh(x) = 1

n

n

  • i=1

Kh(x − xi) = 1 nh

n

  • i=1

K x − xi h

  • ,

where K(·) is a kernel i.e. a non-negative real-valued integrable function such that +∞

−∞

K(u) du = 1 so that fh(·) is a density , and K(·) is symmetric, i.e. K(−u) = K(u). 4

slide-5
SLIDE 5

Arthur CHARPENTIER, transformed kernels and beta kernels

5

slide-6
SLIDE 6

Arthur CHARPENTIER, transformed kernels and beta kernels

Standard kernels are

  • uniform (rectangular) K(u) = 1

2 1{|u|≤1}

  • triangular K(u) = (1 − |u|) 1{|u|≤1}
  • Epanechnikov K(u) = 3

4(1 − u2) 1{|u|≤1}

  • Gaussian K(u) =

1 √ 2π exp

  • −1

2u2

  • 6
slide-7
SLIDE 7

Arthur CHARPENTIER, transformed kernels and beta kernels

Standard kernels are

  • uniform (rectangular) K(u) = 1

2 1{|u|≤1}

  • triangular K(u) = (1 − |u|) 1{|u|≤1}
  • Epanechnikov K(u) = 3

4(1 − u2) 1{|u|≤1}

  • Gaussian K(u) =

1 √ 2π exp

  • −1

2u2

  • 7
slide-8
SLIDE 8

Arthur CHARPENTIER, transformed kernels and beta kernels

word about copulas

A 2-dimensional copula is a 2-dimensional cumulative distribution function restricted to [0, 1]2 with standard uniform margins.

Copula (cumulative distribution function) Level curves of the copula Copula density Level curves of the copula

If C is twice differentiable, let c denote the density of the copula. 8

slide-9
SLIDE 9

Arthur CHARPENTIER, transformed kernels and beta kernels

Sklar theorem : Let C be a copula, and FX and FY two marginal distributions, then F(x, y) = C(FX(x), FY (y)) is a bivariate distribution function, with F ∈ F(FX, FY ). Conversely, if F ∈ F(FX, FY ), there exists C such that F(x, y) = C(FX(x), FY (y). Further, if FX and FY are continuous, then C is unique, and given by C(u, v) = F(F −1

X (u), F −1 Y (v)) for all (u, v) ∈ [0, 1] × [0, 1]

We will then define the copula of F, or the copula of (X, Y ). 9

slide-10
SLIDE 10

Arthur CHARPENTIER, transformed kernels and beta kernels

Motivation

Example Loss-ALAE : consider the following dataset, were the Xi’s are loss amount (paid to the insured) and the Yi’s are allocated expenses. Denote by Ri and Si the respective ranks of Xi and Yi. Set Ui = Ri/n = ˆ FX(Xi) and Vi = Si/n = ˆ FY (Yi). Figure 1 shows the log-log scatterplot (log Xi, log Yi), and the associate copula based scatterplot (Ui, Vi). Figure 2 is simply an histogram of the (Ui, Vi), which is a nonparametric estimation of the copula density. Note that the histogram suggests strong dependence in upper tails (the interesting part in an insurance/reinsurance context). 10

slide-11
SLIDE 11

Arthur CHARPENTIER, transformed kernels and beta kernels

1 2 3 4 5 6 1 2 3 4 5

Log!log scatterplot, Loss!ALAE

log(LOSS) log(ALAE) 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0

Copula type scatterplot, Loss!ALAE

Probability level LOSS Probability level ALAE

Figure 1 – Loss-ALAE, scatterplots (log-log and copula type). 11

slide-12
SLIDE 12

Arthur CHARPENTIER, transformed kernels and beta kernels

Figure 2 – Loss-ALAE, histogram of copula type transformation. 12

slide-13
SLIDE 13

Arthur CHARPENTIER, transformed kernels and beta kernels

Why nonparametrics, instead of parametrics ?

In parametric estimation, assume the the copula density cθ belongs to some given family C = {cθ, θ ∈ Θ}. The tail behavior will crucially depend on the tail behavior of the copulas in C Example : Table below show the probability that both X and Y exceed high thresholds (X > F −1

X (p) and Y > F −1 Y (p)), for usual copula families, where

parameter θ is obtained using maximum likelihood techniques.

p Clayton Frank Gaussian Gumbel Clayton∗ max/min 90% 1.93500% 2.73715% 4.73767% 4.82614% 5.66878% 2.93 95% 0.51020% 0.78464% 1.99195% 2.30085% 2.78677% 5.46 99% 0.02134% 0.03566% 0.27337% 0.44246% 0.55102% 25.82 99.9% 0.00021% 0.00037% 0.01653% 0.04385% 0.05499% 261.85 Probability of exceedances, for given parametric copulas, τ = 0.5.

13

slide-14
SLIDE 14

Arthur CHARPENTIER, transformed kernels and beta kernels

Figure 3 shows the graphical evolution of p → P X > F −1

X (p), Y > F −1 Y (p)

. If the

  • riginal model is an multivariate student vector (X, Y ), the associated probability is the

upper line. If either marginal distributions are misfitted (e.g. Gaussian assumption), or the dependence structure is mispecified (e.g. Gaussian assumption), probabilities are always underestimated.

14

slide-15
SLIDE 15

Arthur CHARPENTIER, transformed kernels and beta kernels

0.6 0.7 0.8 0.9 1.0 0.00 0.05 0.10 0.15 0.20 0.25 0.30

Joint probability of exceeding high quantiles

Quantile levels Student dependence structure, Student margins Gaussian dependence structure, Student margins Student dependence structure, Gaussian margins Gaussian dependence structure, Gaussian margins

0.5 0.6 0.7 0.8 0.9 1.0 0.0 0.2 0.4 0.6 0.8 1.0

Ratios of exceeding probability

Misfitting dependence structure Misfitting margins Misfit margins and dependence

Figure 3 – p → P

  • X > F −1

X (p), Y > F −1 Y (p)

  • when (X, Y ) is a Student t random

vector, and when either margins or the dependence structure is mispectified. 15

slide-16
SLIDE 16

Arthur CHARPENTIER, transformed kernels and beta kernels

Kernel estimation for bounded support density

Consider a kernel based estimation of density f,

  • f(x) = 1

nh

n

  • i=1

K

x − Xi

h

  • ,

where K is a kernel function, given a n sample X1, ..., Xn of positive random variable (Xi ∈ [0, ∞[). Let K denote a symmetric kernel, then E( f(0)) = 1 2f(0) + O(h)

16

slide-17
SLIDE 17

Arthur CHARPENTIER, transformed kernels and beta kernels

Exponential random variables

Density !1 1 2 3 4 5 6 0.0 0.2 0.4 0.6 0.8 1.0 1.2

Exponential random variables

Density !1 1 2 3 4 5 6 0.0 0.2 0.4 0.6 0.8 1.0 1.2

Figure 4 – Density estimation of an exponential density, 100 points. 17

slide-18
SLIDE 18

Arthur CHARPENTIER, transformed kernels and beta kernels 2 4 6 0.0 0.2 0.4 0.6 0.8 1.0 1.2 Density !0.2 0.0 0.1 0.2 0.3 0.4 0.5 0.0 0.2 0.4 0.6 0.8 1.0 Density

Figure 5 – Density estimation of an exponential density, 10, 000 points. 18

slide-19
SLIDE 19

Arthur CHARPENTIER, transformed kernels and beta kernels

0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 1.2

Kernel based estimation of the uniform density on [0,1]

Density 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 1.2

Kernel based estimation of the uniform density on [0,1]

Density

Figure 6 – Density estimation of an uniform density on [0, 1]. 19

slide-20
SLIDE 20

Arthur CHARPENTIER, transformed kernels and beta kernels

How to get a proper estimation on the border

Several techniques have been introduce to get a better estimation on the border, – boundary kernel (Müller (1991)) – mirror image modification (Deheuvels & Hominal (1989), Schuster (1985)) – transformed kernel (Devroye & Györfi (1981), Wand, Marron & Ruppert (1991)) In the particular case of densities on [0, 1], – Beta kernel (Brown & Chen (1999), Chen (1999, 2000)), – average of histograms inspired by Dersko (1998).

20

slide-21
SLIDE 21

Arthur CHARPENTIER, transformed kernels and beta kernels

Local kernel density estimators

The idea is that the bandwidth h(x) can be different for each point x at which f(x) is

  • estimated. Hence,
  • f(x, h(x)) =

1 nh(x)

n

  • i=1

K

  • x − Xi

h(x)

  • ,

(see e.g. Loftsgaarden & Quesenberry (1965)).

Variable kernel density estimators

The idea is that the bandwidth h can be replaced by n values α(Xi). Hence,

  • f(x, α) = 1

n

n

  • i=1

1 α(Xi)K

  • x − Xi

α(Xi)

  • ,

(see e.g. Abramson (1982)).

21

slide-22
SLIDE 22

Arthur CHARPENTIER, transformed kernels and beta kernels

The transformed Kernel estimate

The idea was developed in Devroye & Györfi (1981) for univariate densities. Consider a transformation T : R → [0, 1] strictly increasing, continuously differentiable,

  • ne-to-one, with a continuously differentiable inverse.

Set Y = T(X). Then Y has density fY (y) = fX(T −1(y)) · (T −1)′(y). If fY is estimated by fY , then fX is estimated by

  • fX(x) =

fY (T(x)) · T ′(x).

22

slide-23
SLIDE 23

Arthur CHARPENTIER, transformed kernels and beta kernels −3 −2 −1 1 2 3 0.0 0.1 0.2 0.3 0.4 0.5

Density estimation transformed kernel

0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

Density estimation transformed kernel

Figure 7 – The transform kernel principle (with a Φ−1-transformation). 23

slide-24
SLIDE 24

Arthur CHARPENTIER, transformed kernels and beta kernels

The Beta Kernel estimate

The Beta-kernel based estimator of a density with support [0, 1] at point x, is obtained using beta kernels, which yields

  • f(x) = 1

n

n

  • i=1

K

  • Xi, u

b + 1, 1 − u b + 1

  • where K(·, α, β) denotes the density of the Beta distribution with parameters α and β,

K(x, α, β) = Γ(α + β) Γ(α)Γ(β)xα−1(1 − x)β−11{x∈[0,1]}.

24

slide-25
SLIDE 25

Arthur CHARPENTIER, transformed kernels and beta kernels

0.0 0.2 0.4 0.6 0.8 1.0 2 4 6 8 10

Gaussian Kernel approach

0.0 0.2 0.4 0.6 0.8 1.0 2 4 6 8 10

Beta Kernel approach

0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

Density estimation using beta kernels

Figure 8 – The beta-kernel estimate. 25

slide-26
SLIDE 26

Arthur CHARPENTIER, transformed kernels and beta kernels

Copula density estimation : the boundary problem

Let (U1, V1), ..., (Un, Vn) denote a sample with support [0, 1]2, and with density c(u, v), which is assumed to be twice continuously differentiable on (0, 1)2. If K denotes a symmetric kernel, with support [−1, +1], then for all (u, v) ∈ [0, 1] × [0, 1], in any corners (e.g. (0, 0)) E( c(0, 0, h)) = 1 4 · c(u, v) − 1 2[c1(0, 0) + c2(0, 0)]

1

ωK(ω)dω · h + o(h).

  • n the interior of the borders (e.g. u = 0 and v ∈ (0, 1)),

E( c(0, v, h)) = 1 2 · c(u, v) − [c1(0, v)]

1

ωK(ω)dω · h + o(h). and in the interior ((u, v) ∈ (0, 1) × (0, 1)), E( c(u, v, h)) = c(u, v) + 1 2[c1,1(u, v) + c2,2(u, v)]

1

−1

ω2K(ω)dω · h2 + o(h2). On borders, there is a multiplicative bias and the order of the bias is O(h) (while it is O(h2) in the interior).

26

slide-27
SLIDE 27

Arthur CHARPENTIER, transformed kernels and beta kernels

If K denotes a symmetric kernel, with support [−1, +1], then for all (u, v) ∈ [0, 1] × [0, 1], in any corners (e.g. (0, 0)) V ar( c(0, 0, h)) = c(0, 0)

1

K(ω)2dω

2

· 1 nh2 + o

1

nh2

  • .
  • n the interior of the borders (e.g. u = 0 and v ∈ (0, 1)),

V ar( c(0, v, h)) = c(0, v)

1

−1

K(ω)2dω

1

K(ω)2dω

  • ·

1 nh2 + o

1

nh2

  • .

and in the interior ((u, v) ∈ (0, 1) × (0, 1)), V ar( c(u, v, h)) = c(u, v)

1

−1

K(ω)2dω

2

d · 1 nh2 + o

1

nh2

  • .

27

slide-28
SLIDE 28

Arthur CHARPENTIER, transformed kernels and beta kernels

0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 1 2 3 4 5

Estimation of Frank copula

0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0

Figure 9 – Theoretical density of Frank copula. 28

slide-29
SLIDE 29

Arthur CHARPENTIER, transformed kernels and beta kernels

0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 1 2 3 4 5

Estimation of Frank copula

0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0

Figure 10 – Estimated density of Frank copula, using standard Gaussian (inde- pendent) kernels, h = h∗. 29

slide-30
SLIDE 30

Arthur CHARPENTIER, transformed kernels and beta kernels

Transformed kernel technique

Consider the kernel estimator of the density of the (Xi, Yi) = (G−1(Ui), G−1(Vi))’s, where G is a strictly increasing distribution function, with a differentiable density. Since density f is continuous, twice differentiable, and bounded above, for all (x, y) ∈ R2,

  • f(x, y) =

1 nh2

n

  • i=1

K

x − Xi

h

  • K

y − Yi

h

  • ,

satisfies E( f(x, y)) = f(x, y) + O(h2), as long as ωK(ω) = 0. And the variance is V ar( f(x, y)) = f(x, y) nh2

  • K(ω)2dω

2

+ o

1

nh2

  • ,

and asymptotic normality can be obtained, √ nh2

  • f(x, y) − E(

f(x, y))

  • L

→ N(0, f(x, y)

  • K(ω)2dω

2

).

30

slide-31
SLIDE 31

Arthur CHARPENTIER, transformed kernels and beta kernels

Since f(x, y) = g(x)g(y)c[G(x), G(y)]. (1) can be inverted in c(u, v) = f(G−1(u), G−1(v)) g(G−1(u))g(G−1(v)), (u, v) ∈ [0, 1] × [0, 1], (2)

  • ne gets, substituting

f in (2)

  • c(u, v) =

1 nh · g(G−1(u)) · g(G−1(v))

n

  • i=1

K

  • G−1(u) − G−1(Ui)

h , G−1(v) − G−1(Vi) h

  • ,

(3) Therefore, E( c(u, v, h)) = c(u, v) +

  • (h)

g(G−1(u))g(G−1(v)).

31

slide-32
SLIDE 32

Arthur CHARPENTIER, transformed kernels and beta kernels

Similarly, V ar( c(u, v, h)) = 1 g(G−1(u))g(G−1(v))

  • c(u, v)

nh2

  • K(ω)2dω

2

+ 1 g(G−1(u))2g(G−1(v))2 o

1

nh2

  • .

32

slide-33
SLIDE 33

Arthur CHARPENTIER, transformed kernels and beta kernels

0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 1 2 3 4 5

Estimation of Frank copula

0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8

Figure 11 – Estimated density of Frank copula, using a Gaussian kernel, after a Gaussian normalization. 33

slide-34
SLIDE 34

Arthur CHARPENTIER, transformed kernels and beta kernels

0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 1 2 3 4 5

Estimation of Frank copula

0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8

Figure 12 – Estimated density of Frank copula, using a Gaussian kernel, after a Student normalization, with 5 degrees of freedom. 34

slide-35
SLIDE 35

Arthur CHARPENTIER, transformed kernels and beta kernels

0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 1 2 3 4 5

Estimation of Frank copula

0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8

Figure 13 – Estimated density of Frank copula, using a Gaussian kernel, after a Student normalization, with 3 degrees of freedom. 35

slide-36
SLIDE 36

Arthur CHARPENTIER, transformed kernels and beta kernels

Bivariate Beta kernels

The Beta-kernel based estimator of the copula density at point (u, v), is obtained using product beta kernels, which yields

  • c(u, v) = 1

n

n

  • i=1

K

  • Xi, u

b + 1, 1 − u b + 1

  • · K
  • Yi, v

b + 1, 1 − v b + 1

  • ,

where K(·, α, β) denotes the density of the Beta distribution with parameters α and β.

36

slide-37
SLIDE 37

Arthur CHARPENTIER, transformed kernels and beta kernels

Beta (independent) bivariate kernel , x=0.0, y=0.0 Beta (independent) bivariate kernel , x=0.2, y=0.0 Beta (independent) bivariate kernel , x=0.5, y=0.0 Beta (independent) bivariate kernel , x=0.0, y=0.2 Beta (independent) bivariate kernel , x=0.2, y=0.2 Beta (independent) bivariate kernel , x=0.5, y=0.2 Beta (independent) bivariate kernel , x=0.0, y=0.5 Beta (independent) bivariate kernel , x=0.2, y=0.5 Beta (independent) bivariate kernel , x=0.5, y=0.5

Figure 14 – Shape of bivariate Beta kernels K(·, x/b+1, (1−x)/b+1)×K(·, y/b+ 1, (1 − y)/b + 1) for b = 0.2. 37

slide-38
SLIDE 38

Arthur CHARPENTIER, transformed kernels and beta kernels

Assume that the copula density c is twice differentiable on [0, 1] × [0, 1]. Let (u, v) ∈ [0, 1] × [0, 1]. The bias of c(u, v) is of order b, i.e. E( c(u, v)) = c(u, v) + Q(u, v) · b + o(b), where the bias Q(u, v) is Q(u, v) = (1 − 2u)c1(u, v) + (1 − 2v)c2(u, v) + 1 2 [u(1 − u)c1,1(u, v) + v(1 − v)c2,2(u, v)] . The bias here is O(b) (everywhere) while it is O(h2) using standard kernels. Assume that the copula density c is twice differentiable on [0, 1] × [0, 1]. Let (u, v) ∈ [0, 1] × [0, 1]. The variance of c(u, v) is in corners, e.g. (0, 0), V ar( c(0, 0)) = 1 nb2 [c(0, 0) + o(n−1)], in the interior of borders, e.g. u = 0 and v ∈ (0, 1) V ar( c(0, v)) = 1 2nb3/2 πv(1 − v) [c(0, v) + o(n−1)],

38

slide-39
SLIDE 39

Arthur CHARPENTIER, transformed kernels and beta kernels

and in the interior,(u, v) ∈ (0, 1) × (0, 1) V ar( c(u, v)) = 1 4nbπ

  • v(1 − v)u(1 − u)

[c(u, v) + o(n−1)]. Remark From those properties, an (asymptotically) optimal bandwidth b can be deduced, maximizing asymptotic mean squared error, b∗ ≡

  • 1

16πnQ(u, v)2 · 1

  • v(1 − v)u(1 − u)

1/3

. Note (see Charpentier, Fermanian & Scaillet (2005)) that all those results can be

  • btained in dimension d ≥ 2.

Example For n = 100 simulated data, from Frank copula, the optimal bandwidth is b ∼ 0.05.

39

slide-40
SLIDE 40

Arthur CHARPENTIER, transformed kernels and beta kernels

0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 0.0 0.5 1.0 1.5 2.0 2.5 3.0

Estimation of the copula density (Beta kernel, b=0.1) Estimation of the copula density (Beta kernel, b=0.1)

0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0

Figure 15 – Estimated density of Frank copula, Beta kernels, b = 0.1 40

slide-41
SLIDE 41

Arthur CHARPENTIER, transformed kernels and beta kernels

0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 0.0 0.5 1.0 1.5 2.0 2.5 3.0

Estimation of the copula density (Beta kernel, b=0.05) Estimation of the copula density (Beta kernel, b=0.05)

0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0

Figure 16 – Estimated density of Frank copula, Beta kernels, b = 0.05 41

slide-42
SLIDE 42

Arthur CHARPENTIER, transformed kernels and beta kernels

0.0 0.2 0.4 0.6 0.8 1.0 1 2 3 4

Standard Gaussian kernel estimator, n=100

Estimation of the density on the diagonal Density of the estimator 0.0 0.2 0.4 0.6 0.8 1.0 1 2 3 4

Standard Gaussian kernel estimator, n=1000

Estimation of the density on the diagonal Density of the estimator 0.0 0.2 0.4 0.6 0.8 1.0 1 2 3 4

Standard Gaussian kernel estimator, n=10000

Estimation of the density on the diagonal Density of the estimator

Figure 17 – Density estimation on the diagonal, standard kernel. 42

slide-43
SLIDE 43

Arthur CHARPENTIER, transformed kernels and beta kernels

0.0 0.2 0.4 0.6 0.8 1.0 1 2 3 4

Transformed kernel estimator (Gaussian), n=100

Estimation of the density on the diagonal Density of the estimator 0.0 0.2 0.4 0.6 0.8 1.0 1 2 3 4

Transformed kernel estimator (Gaussian), n=1000

Estimation of the density on the diagonal Density of the estimator 0.0 0.2 0.4 0.6 0.8 1.0 1 2 3 4

Transformed kernel estimator (Gaussian), n=10000

Estimation of the density on the diagonal Density of the estimator

Figure 18 – Density estimation on the diagonal, transformed kernel. 43

slide-44
SLIDE 44

Arthur CHARPENTIER, transformed kernels and beta kernels

0.0 0.2 0.4 0.6 0.8 1.0 1 2 3 4

Beta kernel estimator, b=0.05, n=100

Estimation of the density on the diagonal Density of the estimator 0.0 0.2 0.4 0.6 0.8 1.0 1 2 3 4

Beta kernel estimator, b=0.02, n=1000

Estimation of the density on the diagonal Density of the estimator 0.0 0.2 0.4 0.6 0.8 1.0 1 2 3 4

Beta kernel estimator, b=0.005, n=10000

Estimation of the density on the diagonal Density of the estimator

Figure 19 – Density estimation on the diagonal, Beta kernel. 44

slide-45
SLIDE 45

Arthur CHARPENTIER, transformed kernels and beta kernels

Copula density estimation

Gijbels & Mielniczuk (1990) : given an i.i.d. sample, a natural estimate for the normed density is obtained using the transformed sample ( FX(X1), FY (Y1)), ..., ( FX(Xn), FY (Yn)), where FX and FY are the empirical distribution function of the marginal distribution. The copula density can be constructed as some density estimate based on this sample (Behnen, Husková & Neuhaus (1985) investigated the kernel method). The natural kernel type estimator c of c is c(u, v) = 1 nh2

n

  • i=1

K

  • u −

FX(Xi) h , v − FY (Yi) h

  • , (u, v) ∈ [0, 1].

“this estimator is not consistent in the points on the boundary of the unit square.”

45

slide-46
SLIDE 46

Arthur CHARPENTIER, transformed kernels and beta kernels

Copula density estimation and pseudo-observations

Example : in linear regression, residuals are pseudo observations. εi = H(Xi, Yi) = Yi − α − βXi

  • εi =

Hn(Xi, Yi) = Yi − αn − βnXi Example : when dealing with copulas, ranks Ui, Vi yield pseudo-observations. (Ui, Vi) = H(Xi, Yi) = (FX(Xi), FY (Yi)) ( Ui, Vi) = Hn(Xi, Yi) = ( FX(Xi), FY (Yi)) (see Genest & Rivest (1993)). More formally, let X1, ..., Xn denote a series of observations of X (∈ X), stationary and ergodic. Let H : X → Rd and set εi = H(Xi) (non-observable). If H is estimated by Hn then εi = Hn(Xi) are called pseudo-observations.

46

slide-47
SLIDE 47

Arthur CHARPENTIER, transformed kernels and beta kernels

Let Kn denote the empirical distribution function of those pseudo-observations,

  • Kn(t) = 1

n

n

  • i=1

I( εi ≤ t) where t ∈ Rd. Further, if K denotes the distribution function of ε = H(X), then define the empirical process based on pseudo-observations, Kn(t) = √n

  • Kn(t) − K(t)
  • As proved in Ghoudi & Rémillard (1998, 2004), this empirical process converges

weakly. Figure ?? shows scatterplots when margins are known (i.e. (FX(Xi), FY (Yi))’s), and when margins are estimated (i.e. ( ˆ FX(Xi), ˆ FY (Yi)’s). Note that the pseudo sample is more “uniform”, in the sense of a lower discrepancy (as in Quasi Monte Carlo techniques, see e.g. Niederreiter (1992)).

47

slide-48
SLIDE 48

Arthur CHARPENTIER, transformed kernels and beta kernels 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0

Scatterplot of observations (Xi,Yi)

Value of the Xis Value of the Yis 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0

Scatterplot of pseudo−observations (Ui,Vi)

Value of the Uis (ranks) Value of the Vis (ranks)

Figure 20 – Observations and pseudo-observation, 500 simulated observations from Frank copula (Xi, Yi) and the associate pseudo-sample ( ˆ FX(Xi), ˆ FY (Yi)). 48

slide-49
SLIDE 49

Arthur CHARPENTIER, transformed kernels and beta kernels

Because samples are more “uniform” using ranks and pseudo-observations, the variance

  • f the estimator of the density, at some given point (u, v) ∈ (0, 1) × (0, 1) is usually
  • smaller. For instance, Figure 21 shows the impact of considering pseudo observations,

i.e. substituting ˆ FX and ˆ FY to unknown marginal distributions FX and FY . The dotted line shows the density of ˆ c(u, v) from a n = 100 sample (Ui, Vi) (from Frank copula), and the straight line shows the density of ˆ c(u, v) from the sample ( ˆ FU(Ui), ˆ FV (Vi)) (i.e. ranks of observations).

49

slide-50
SLIDE 50

Arthur CHARPENTIER, transformed kernels and beta kernels

1 2 3 4 0.0 0.5 1.0 1.5 2.0

Impact of pseudo−observations (n=100)

Distribution of the estimation of the density c(u,v) Density of the estimator

Figure 21 – The impact of estimating from pseudo-observations. 50

slide-51
SLIDE 51

Arthur CHARPENTIER, transformed kernels and beta kernels

Roots of ‘transformed kernel’

51

slide-52
SLIDE 52

Arthur CHARPENTIER, transformed kernels and beta kernels

52

slide-53
SLIDE 53

Arthur CHARPENTIER, transformed kernels and beta kernels

53

slide-54
SLIDE 54

Arthur CHARPENTIER, transformed kernels and beta kernels

54

slide-55
SLIDE 55

Arthur CHARPENTIER, transformed kernels and beta kernels

55

slide-56
SLIDE 56

Arthur CHARPENTIER, transformed kernels and beta kernels

Using a parametric approach

If FX ∈ F = {Fθ, θ ∈ Θ} (assumed to be continuous), qX(α) = F −1

θ

(α), and thus, a natural estimator is

  • qX(α) = F −1
  • θ

(α), (4) where θ is an estimator of θ (maximum likelihood, moments estimator...).

56

slide-57
SLIDE 57

Arthur CHARPENTIER, transformed kernels and beta kernels

Using the Gaussian distribution

A natural idea (that can be found in classical financial models) is to assume Gaussian distributions : if X ∼ N(µ, σ), then the α-quantile is simply q(α) = µ + Φ−1(α)σ, where Φ−1(α) is obtained in statistical tables (or any statistical software), e.g. u = −1.64 if α = 90%, or u = −1.96 if α = 95%.

Definition1

Given a n sample {X1, · · · , Xn}, the (Gaussian) parametric estimation of the α-quantile is

  • qn(α) =

µ + Φ−1(α) σ,

57

slide-58
SLIDE 58

Arthur CHARPENTIER, transformed kernels and beta kernels

Using a parametric models

Actually, is the Gaussian model does not fit very well, it is still possible to use Gaussian approximation If the variance is finite, (X − E(X))/σ might be closer to the Gaussian distribution, and thus, consider the so-called Cornish-Fisher approximation, i.e. Q(X, α) ∼ E(X) + zα

  • V (X),

(5) where

  • zα = Φ−1(α) + ζ1

6 [Φ−1(α)2 − 1] + ζ2 24[Φ−1(α)3 − 3Φ−1(α)] − ζ2

1

36[2Φ−1(α)3 − 5Φ−1(α)], where ζ1 is the skewness of X, and ζ2 is the excess kurtosis, i.e. i.e. ζ1 = E([X − E(X)]3) E([X − E(X)]2)3/2 and ζ2 = E([X − E(X)]4) E([X − E(X)]2)2 − 3. (6)

58

slide-59
SLIDE 59

Arthur CHARPENTIER, transformed kernels and beta kernels

Using a parametric models

Definition2

Given a n sample {X1, · · · , Xn}, the Cornish-Fisher estimation of the α-quantile is

  • qn(α) =

µ + zα σ, where µ = 1 n

n

  • i=1

Xi and σ =

  • 1

n − 1

n

  • i=1

(Xi − µ)2,

and

zα = Φ−1(α)+ ζ1 6 [Φ−1(α)2−1]+ ζ2 24[Φ−1(α)3−3Φ−1(α)]− ζ2

1

36[2Φ−1(α)3−5Φ−1(α)], (7)

where

ζ1 is the natural estimator for the skewness of X, and ζ2 is the natural estimator of

the excess kurtosis, i.e.

ζ1 =

  • n(n − 1)

n − 2 √n n

i=1(Xi −

µ)3

n

i=1(Xi −

µ)23/2 and

  • ζ2 =

n − 1 (n − 2)(n − 3)

  • (n + 1)

ζ′

2 + 6

  • where

ζ′

2 = n n i=1(Xi −

µ)4

n

i=1(Xi −

µ)22 − 3.

59

slide-60
SLIDE 60

Arthur CHARPENTIER, transformed kernels and beta kernels

Parametrics estimator and error model

1 2 3 4 5 0.0 0.2 0.4 0.6 0.8

Density, theoritical versus empirical

Theoritical lognormal Fitted lognormal Fitted gamma

−4 −2 2 4 0.0 0.1 0.2 0.3

Density, theoritical versus empirical

Theoritical Student Fitted lStudent Fitted Gaussian

Figure 22 – Estimation of Value-at-Risk, model error. 60

slide-61
SLIDE 61

Arthur CHARPENTIER, transformed kernels and beta kernels

Using a semiparametric models

Given a n-sample {Y1, . . . , Yn}, let Y1:n ≤ Y2:n ≤ . . .≤ Yn:n denotes the associated order statistics. If u large enough, Y − u given Y > u has a Generalized Pareto distribution with parameters ξ and β ( Pickands-Balkema-de Haan theorem). If u = Yn−k:n for k large enough, and if ξ>0, denote by βk and ξk maximum likelihood estimators of the Genralized Pareto distribution of sample {Yn−k+1:n − Yn−k:n, ..., Yn:n − Yn−k:n},

  • Q(Y, α) = Yn−k:n +

βk

  • ξk

n

k (1 − α)

ξk

− 1

  • ,

(8) An alternative is to use Hill’s estimator if ξ > 0,

  • Q(Y, α) = Yn−k:n

n

k (1 − α)

ξk

, ξk = 1 k

k

  • i=1

log Yn+1−i:n − log Yn−k:n. (9)

61

slide-62
SLIDE 62

Arthur CHARPENTIER, transformed kernels and beta kernels

On nonparametric estimation for quantiles

For continuous distribution q(α) = F −1

X (α), thus, a natural idea would be to consider

  • q(α) =

F −1

X (α), for some nonparametric estimation of FX.

Definition3

The empirical cumulative distribution function Fn, based on sample {X1, . . . , Xn} is

Fn(x) = 1 n

n

  • i=1

1(Xi ≤ x).

Definition4

The kernel based cumulative distribution function, based on sample {X1, . . . , Xn} is

  • Fn(x) = 1

nh

n

  • i=1

x

−∞

k

Xi − t

h

  • dt = 1

n

n

  • i=1

K

Xi − x

h

  • where K(x) =

x

−∞

k(t)dt, k being a kernel and h the bandwidth.

62

slide-63
SLIDE 63

Arthur CHARPENTIER, transformed kernels and beta kernels

Smoothing nonparametric estimators

Two techniques have been considered to smooth estimation of quantiles, either implicit,

  • r explicit.
  • consider a linear combinaison of order statistics,

The classical empirical quantile estimate is simply Qn(p) = F −1

n

i

n

  • = Xi:n = X[np]:n where [·] denotes the integer part.

(10) The estimator is simple to obtain, but depends only on one observation. A natural extention will be to use - at least - two observations, if np is not an integer. The weighted empirical quantile estimate is then defined as Qn(p) = (1 − γ) X[np]:n + γX[np]+1:n where γ = np − [np].

63

slide-64
SLIDE 64

Arthur CHARPENTIER, transformed kernels and beta kernels

0.0 0.2 0.4 0.6 0.8 1.0 2 4 6 8

The quantile function in R

probability level quantile level

  • type=1

type=3 type=5 type=7 0.0 0.2 0.4 0.6 0.8 1.0 2 3 4 5 6 7

The quantile function in R

probability level quantile level

  • type=1

type=3 type=5 type=7

Figure 23 – Several quantile estimators in R. 64

slide-65
SLIDE 65

Arthur CHARPENTIER, transformed kernels and beta kernels

Smoothing nonparametric estimators

In order to increase efficiency, L-statistics can be considered i.e. Qn (p) =

n

  • i=1

Wi,n,pXi:n =

n

  • i=1

Wi,n,pF −1

n

i

n

  • =

1

F −1

n

(t) k (p, h, t) dt (11) where Fn is the empirical distribution function of FX, where k is a kernel and h a

  • bandwidth. This expression can be written equivalently

Qn (p) =

n

  • i=1
  • i

n (i−1) n

k

t − p

h

  • dt
  • X(i) =

n

  • i=1
  • I

K

i

n − p

h

  • − I

K

i−1

n − p

h

  • X(i)

(12) where again I K (x) =

x

−∞

k (t) dt. The idea is to give more weight to order statistics X(i) such that i is closed to pn.

65

slide-66
SLIDE 66

Arthur CHARPENTIER, transformed kernels and beta kernels

0.0 0.2 0.4 0.6 0.8 1.0 1 2 3 quantile (probability) level

Figure 24 – Quantile estimator as wieghted sum of order statistics. 66

slide-67
SLIDE 67

Arthur CHARPENTIER, transformed kernels and beta kernels

0.0 0.2 0.4 0.6 0.8 1.0 1 2 3 quantile (probability) level

Figure 25 – Quantile estimator as wieghted sum of order statistics. 67

slide-68
SLIDE 68

Arthur CHARPENTIER, transformed kernels and beta kernels

0.0 0.2 0.4 0.6 0.8 1.0 1 2 3 quantile (probability) level

Figure 26 – Quantile estimator as wieghted sum of order statistics. 68

slide-69
SLIDE 69

Arthur CHARPENTIER, transformed kernels and beta kernels

0.0 0.2 0.4 0.6 0.8 1.0 1 2 3 quantile (probability) level

Figure 27 – Quantile estimator as wieghted sum of order statistics. 69

slide-70
SLIDE 70

Arthur CHARPENTIER, transformed kernels and beta kernels

0.0 0.2 0.4 0.6 0.8 1.0 1 2 3 quantile (probability) level

Figure 28 – Quantile estimator as wieghted sum of order statistics. 70

slide-71
SLIDE 71

Arthur CHARPENTIER, transformed kernels and beta kernels

Smoothing nonparametric estimators

E.g. the so-called Harrell-Davis estimator is defined as Qn(p) =

n

  • i=1
  • i

n (i−1) n

Γ(n + 1) Γ((n + 1)p)Γ((n + 1)q)y(n+1)p−1(1 − y)(n+1)q−1

  • Xi:n,
  • find a smooth estimator for FX, and then find (numerically) the inverse,

The α-quantile is defined as the solution of FX ◦ qX(α) = α. If Fn denotes a continuous estimate of F, then a natural estimate for qX(α) is qn(α) such that Fn ◦ qn(α) = α, obtained using e.g. Gauss-Newton algorithm.

71

slide-72
SLIDE 72

Arthur CHARPENTIER, transformed kernels and beta kernels

Improving Beta kernel estimators

Problem : the convergence is not uniform, and there is large second order bias on borders, i.e. 0 and 1. Chen (1999) proposed a modified Beta 2 kernel estimator, based on k2 (u; b; t) =

      

k t

b , 1−t b

(u) , if t ∈ [2b, 1 − 2b] kρb(t), 1−t

b

(u) , if t ∈ [0, 2b) k t

b ,ρb(1−t) (u)

, if t ∈ (1 − 2b, 1] where ρb (t) = 2b2 + 2.5 −

  • 4b4 + 6b2 + 2.25 − t2 − t

b.

72

slide-73
SLIDE 73

Arthur CHARPENTIER, transformed kernels and beta kernels

Non-consistency of Beta kernel estimators

Problem : k(0, α, β) = k(1, α, β) = 0. So if there are point mass at 0 or 1, the estimator becomes inconsistent, i.e.

  • fb(x)

= 1 n

  • k
  • Xi, 1 + x

b , 1 + 1 − x b

  • , x ∈ [0, 1]

= 1 n

  • Xi=0,1

k

  • Xi, 1 + x

b , 1 + 1 − x b

  • , x ∈ [0, 1]

= n − n0 − n1 n 1 n − n0 − n1

  • Xi=0,1

k

  • Xi, 1 + x

b , 1 + 1 − x b

  • , x ∈ [0, 1]

≈ (1 − P(X = 0) − P(X = 1)) · f0(x), x ∈ [0, 1] and therefore Fb(x) ≈ (1 − P(X = 0) − P(X = 1)) · F0(x), and we may have problem finding a 95% or 99% quantile since the total mass will be lower.

73

slide-74
SLIDE 74

Arthur CHARPENTIER, transformed kernels and beta kernels

Non-consistency of Beta kernel estimators

Gouriéroux & Monfort (2007) proposed

  • f (1)

b

(x) =

  • fb(x)

1

fb(t)dt , for all x ∈ [0, 1]. It is called macro-β since the correction is performed globally. Gouriéroux & Monfort (2007) proposed

  • f (2)

b

(x) = 1 n

n

  • i=1

kβ(Xi; b; x)

1

0 kβ(Xi; b; t)dt

, for all x ∈ [0, 1]. It is called micro-β since the correction is performed locally.

74

slide-75
SLIDE 75

Arthur CHARPENTIER, transformed kernels and beta kernels

Transforming observations ?

In the context of density estimation, Devroye and Györfi (1985) suggested to use a so-called transformed kernel estimate Given a random variable Y , if H is a strictly increasing function, then the p-quantile of H(Y ) is equal to H(q(Y ; p)). An idea is to transform initial observations {X1, · · · , Xn} into a sample {Y1, · · · , Yn} where Yi = H(Xi), and then to use a beta-kernel based estimator, if H : R → [0, 1]. Then qn(X; p) = H−1( qn(Y ; p)). In the context of density estimation fX(x) = fY (H(x))H′(x). As mentioned in Devroye and Györfi (1985) (p 245), “for a transformed histogram histogram estimate, the optimal H gives a uniform [0, 1] density and should therefore be equal to H(x) = F(x), for all x”.

75

slide-76
SLIDE 76

Arthur CHARPENTIER, transformed kernels and beta kernels

Transforming observations ? a monte carlo study

Assume that sample {X1, · · · , Xn} have been generated from Fθ0 (from a familly F = (Fθ, θ ∈ Θ). 4 transformations will be considered – H = F

θ (based on a maximum likelihood procedure)

– H = Fθ0 (theoritical optimal transformation) – H = Fθ with θ < θ0 (heavier tails) – H = Fθ with θ > θ0 (lower tails)

76

slide-77
SLIDE 77

Arthur CHARPENTIER, transformed kernels and beta kernels

  • 0.0

0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 Transformed observations

  • 0.0

0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 Transformed observations

Figure 29 – F(Xi) versus Fˆ

θ(Xi), i.e. PP plot.

77

slide-78
SLIDE 78

Arthur CHARPENTIER, transformed kernels and beta kernels

Estimated density 0.0 0.2 0.4 0.6 0.8 1.0 0.6 0.8 1.0 1.2 1.4

  • 0.0

0.2 0.4 0.6 0.8 1.0 0.6 0.8 1.0 1.2 1.4 Estimated density

Figure 30 – Nonparametric estimation of the density of the Fˆ

θ(Xi)’s.

78

slide-79
SLIDE 79

Arthur CHARPENTIER, transformed kernels and beta kernels

  • ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
  • 0.80

0.85 0.90 0.95 1.00 1.0 1.5 2.0 2.5 3.0 3.5 4.0

Estimated optimal transformation

Probability level Quantile

  • ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
  • 0.80

0.85 0.90 0.95 1.00 1 2 3 4 5

Estimated optimal transformation

Probability level Quantile

Figure 31 – Nonparametric estimation of the quantile function, F −1

ˆ θ

(q). 79

slide-80
SLIDE 80

Arthur CHARPENTIER, transformed kernels and beta kernels

  • 0.0

0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 Transformed observations

  • 0.0

0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 Transformed observations

Figure 32 – F(Xi) versus Fθ0(Xi), i.e. PP plot. 80

slide-81
SLIDE 81

Arthur CHARPENTIER, transformed kernels and beta kernels

Estimated density 0.0 0.2 0.4 0.6 0.8 1.0 0.6 0.8 1.0 1.2 1.4

  • 0.0

0.2 0.4 0.6 0.8 1.0 0.6 0.8 1.0 1.2 1.4 Estimated density

Figure 33 – Nonparametric estimation of the density of the Fθ0(Xi)’s. 81

slide-82
SLIDE 82

Arthur CHARPENTIER, transformed kernels and beta kernels

  • ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
  • 0.80

0.85 0.90 0.95 1.00 1 2 3 4

Estimated optimal transformation

Probability level Quantile

  • ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
  • 0.80

0.85 0.90 0.95 1.00 1 2 3 4

Estimated optimal transformation

Probability level Quantile

Figure 34 – Nonparametric estimation of the quantile function, F −1

θ0 (q).

82

slide-83
SLIDE 83

Arthur CHARPENTIER, transformed kernels and beta kernels

  • 0.0

0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 Transformed observations

  • 0.0

0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 Transformed observations

Figure 35 – F(Xi) versus Fθ(Xi), i.e. PP plot, θ < θ0 (heavier tails). 83

slide-84
SLIDE 84

Arthur CHARPENTIER, transformed kernels and beta kernels

Estimated density 0.0 0.2 0.4 0.6 0.8 1.0 0.6 0.8 1.0 1.2 1.4

  • 0.0

0.2 0.4 0.6 0.8 1.0 0.6 0.8 1.0 1.2 1.4 Estimated density

Figure 36 – Estimation of the density of the Fθ(Xi)’s, θ < θ0 (heavier tails). 84

slide-85
SLIDE 85

Arthur CHARPENTIER, transformed kernels and beta kernels

  • ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
  • 0.80

0.85 0.90 0.95 1.00 2 4 6 8 10 12

Estimated optimal transformation

Probability level Quantile

  • ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
  • 0.80

0.85 0.90 0.95 1.00 2 4 6 8 10 12

Estimated optimal transformation

Probability level Quantile

Figure 37 – Estimation of quantile function, F −1

θ

(q), θ < θ0 (heavier tails). 85

slide-86
SLIDE 86

Arthur CHARPENTIER, transformed kernels and beta kernels

  • 0.0

0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 Transformed observations

  • 0.0

0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 Transformed observations

Figure 38 – F(Xi) versus Fθ(Xi), i.e. PP plot, θ > θ0 (lighter tails). 86

slide-87
SLIDE 87

Arthur CHARPENTIER, transformed kernels and beta kernels

Estimated density 0.0 0.2 0.4 0.6 0.8 1.0 0.6 0.8 1.0 1.2 1.4

  • 0.0

0.2 0.4 0.6 0.8 1.0 0.6 0.8 1.0 1.2 1.4 Estimated density

Figure 39 – Estimation of density of Fθ(Xi)’s, θ > θ0 (lighter tails). 87

slide-88
SLIDE 88

Arthur CHARPENTIER, transformed kernels and beta kernels

  • ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
  • 0.80

0.85 0.90 0.95 1.00 1.0 1.5 2.0 2.5 3.0 3.5

Estimated optimal transformation

Probability level Quantile

  • ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
  • 0.80

0.85 0.90 0.95 1.00 1.0 1.5 2.0 2.5 3.0 3.5

Estimated optimal transformation

Probability level Quantile

Figure 40 – Estimation of quantile function, F −1

θ

(q), θ > θ0 (lighter tails). 88

slide-89
SLIDE 89

Arthur CHARPENTIER, transformed kernels and beta kernels

A universal distribution for losses

Buch-Larsen,Nielsen, Guillen, & Bolancé (2005) considered the Champernowne generalized distribution to model insurance claims, i.e. positive variables, Fα,M,c (y) = (y + c)α − cα (y + c)α + (M + c)α − 2cα where α > 0, c ≥ 0 and M > 0. The associated density is then fα,M,c (y) = α (y + c)α−1 ((M + c)α − cα) ((y + c)α + (M + c)α − 2cα)2 .

89

slide-90
SLIDE 90

Arthur CHARPENTIER, transformed kernels and beta kernels

A Monte Carlo study to compare those nonparametric estimators

As in Buch-Larsen,Nielsen, Guillen, & Bolancé (2005), 4 distributions were considered – normal distribution, – Weibull distribution, – log-normal distribution, – mixture of Pareto and log-normal distributions,

90

slide-91
SLIDE 91

Arthur CHARPENTIER, transformed kernels and beta kernels

5 10 15 20 0.00 0.05 0.10 0.15 0.20 0.25 0.30

Density of quantile estimators (mixture longnormal/pareto)

Estimated value−at−risk density of estimators Benchmark (R estimator) HD (Harrell−Davis) PRK (Park) B1 (Beta 1) B2 (Beta 2)

  • Box−plot for the 11 quantile estimators

5 10 15 20

  • ● ●
  • ● ● ●
  • ● ● ●
  • ●●
  • ●●
  • ●● ●
  • ● ●
  • MICRO Beta2

MACRO Beta2 Beta2 MICRO Beta1 MACRO Beta1 Beta1 PRK Park PDG Padgett HD Harrell Davis E Epanechnikov R benchmark

Figure 41 – Distribution of the 95% quantile of the mixture distribution, n = 200, and associated box-plots. 91

slide-92
SLIDE 92

Arthur CHARPENTIER, transformed kernels and beta kernels

  • MSE ratio, normal distribution, HD (Harrell−Davis)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, normal distribution, B1 (Beta1)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, normal distribution, MACB1 (MACRO−Beta1)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, normal distribution, PRK (Park)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, normal distribution, B1 (Beta1)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, normal distribution, MACB1 (MACRO−Beta1)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

Figure 42 – Comparing MSE for 6 estimators, the normal distribution case. 92

slide-93
SLIDE 93

Arthur CHARPENTIER, transformed kernels and beta kernels

  • MSE ratio, Weibull distribution, HD (Harrell−Davis)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, Weibull distribution, MACB1 (MACRO−Beta1)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, Weibull distribution, MICB1 (MICRO−Beta1)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, Weibull distribution, PRK (Park)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, Weibull distribution, MACB1 (MACRO−Beta1)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, Weibull distribution, MICB1 (MICRO−Beta1)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

Figure 43 – Comparing MSE for 6 estimators, the Weibull distribution case. 93

slide-94
SLIDE 94

Arthur CHARPENTIER, transformed kernels and beta kernels

  • MSE ratio, lognormal distribution, HD (Harrell−Davis)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, lognormal distribution, MACB1 (MACRO−Beta1)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, lognormal distribution, MICB1 (MICRO−Beta1)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, lognormal distribution, B1 (Beta1)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, lognormal distribution, PRK (Park)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, lognormal distribution, MACB2 (MACRO−Beta2)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, lognormal distribution, MICB2 (MICRO−Beta2)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, lognormal distribution, B2 (Beta2)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

Figure 44 – Comparing MSE for 9 estimators, the lognormal distribution case. 94

slide-95
SLIDE 95

Arthur CHARPENTIER, transformed kernels and beta kernels

  • MSE ratio, mixture distribution, HD (Harrell−Davis)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, mixture distribution, MACB1 (MACRO−Beta1)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, mixture distribution, MICB1 (MICRO−Beta1)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, mixture distribution, B1 (Beta1)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, mixture distribution, PRK (Park)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, mixture distribution, MACB2 (MACRO−Beta2)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, mixture distribution, MICB2 (MICRO−Beta2)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

  • MSE ratio, mixture distribution, B2 (Beta2)

Probability, confidence levels (p) MSE ratio 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0

  • n= 50

n=100 n=200 n=500

Figure 45 – Comparing MSE for 9 estimators, the mixture distribution case. 95

slide-96
SLIDE 96

Arthur CHARPENTIER, transformed kernels and beta kernels

Some references

Charpentier, A. , Fermanian, J-D. & Scaillet, O. (2006). The estimation of copulas : theory and practice. In : Copula Methods in Derivatives and Risk Management : From Credit Risk to Market Risk. Risk Book Charpentier, A. & Oulidi, A. (2008). Beta Kernel estimation for Value-At-Risk of heavy-tailed distributions. in revision Journal of Computational Statistics and Data Analysis. Charpentier, A. & Oulidi, A. (2007). Estimating allocations for Value-at-Risk portfolio optimzation. to appear in Mathematical Methods in Operations Research. Chen, S. X. (1999). A Beta Kernel Estimator for Density Functions. Computational Statistics & Data Analysis, 31, 131-145. Devroye, L. & Györfi, L. (1981). Nonparametric Density Estimation : The L1 View. Wiley. Gouriéroux, C., & Montfort, A. 2006. (Non) Consistency of the Beta Kernel Estimator for Recovery Rate Distribution. CREST-DP 2006-32.

96