Central and non-central limit theorems for statistical functionals - - PowerPoint PPT Presentation

central and non central limit theorems for statistical
SMART_READER_LITE
LIVE PREVIEW

Central and non-central limit theorems for statistical functionals - - PowerPoint PPT Presentation

Central and non-central limit theorems for statistical functionals based on weakly and strongly dependent data Eric Beutner, Maastricht University Tokyo, September 3 (atbegshi) Package atbegshi Warning: Ignoring void shipout box Overview


slide-1
SLIDE 1

(atbegshi) Package atbegshi Warning: Ignoring void shipout box

Central and non-central limit theorems for statistical functionals based on weakly and strongly dependent data

Eric Beutner, Maastricht University

Tokyo, September 3

slide-2
SLIDE 2

Overview

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 2

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics

slide-3
SLIDE 3

Motivation

Motivation Statistical functionals Functional delta method Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 3

slide-4
SLIDE 4

Statistical functionals

Motivation Statistical functionals Functional delta method Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 4

Let θ = T(F) be some characteristic of the distribution

function F (df).

slide-5
SLIDE 5

Statistical functionals

Motivation Statistical functionals Functional delta method Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 4

Let θ = T(F) be some characteristic of the distribution

function F (df).

Typical examples are T(F) = −

−∞ K(F(x)) dx +

0 (1 − K(F(x)) dx

so called L-statistics. Distortion risk measures which are quite popular are also of this form.

slide-6
SLIDE 6

Statistical functionals

Motivation Statistical functionals Functional delta method Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 4

Let θ = T(F) be some characteristic of the distribution

function F (df).

Typical examples are T(F) = −

−∞ K(F(x)) dx +

0 (1 − K(F(x)) dx

so called L-statistics. Distortion risk measures which are quite popular are also of this form.

T(F) =

g(x1, x2) dF(x1)dF(x2) so called U- or V-statistic (of degree 2).

slide-7
SLIDE 7

Statistical functionals

Motivation Statistical functionals Functional delta method Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 4

Let θ = T(F) be some characteristic of the distribution

function F (df).

Typical examples are T(F) = −

−∞ K(F(x)) dx +

0 (1 − K(F(x)) dx

so called L-statistics. Distortion risk measures which are quite popular are also of this form.

T(F) =

g(x1, x2) dF(x1)dF(x2) so called U- or V-statistic (of degree 2).

Z-estimators.

slide-8
SLIDE 8

Statistical functionals

Motivation Statistical functionals Functional delta method Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 4

Let θ = T(F) be some characteristic of the distribution

function F (df).

Typical examples are T(F) = −

−∞ K(F(x)) dx +

0 (1 − K(F(x)) dx

so called L-statistics. Distortion risk measures which are quite popular are also of this form.

T(F) =

g(x1, x2) dF(x1)dF(x2) so called U- or V-statistic (of degree 2).

Z-estimators. Given n observations X1, . . . , Xn with df F a natural

estimator is then T(Fn) with Fn the empirical distribution function.

slide-9
SLIDE 9

Functional delta method

Motivation Statistical functionals Functional delta method Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 5

Well-known: If T is Hadamard differentiable at F, then

the asymptotic distribution of T(Fn) follows immediately by the (functional) delta method from the asymptotic distribution of Fn − F.

slide-10
SLIDE 10

Functional delta method

Motivation Statistical functionals Functional delta method Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 5

Well-known: If T is Hadamard differentiable at F, then

the asymptotic distribution of T(Fn) follows immediately by the (functional) delta method from the asymptotic distribution of Fn − F.

Thus, delta method leads asymptotic distribution of

T(Fn) − T(F) whenever we have weak convergence of the empirical process.

slide-11
SLIDE 11

Functional delta method

Motivation Statistical functionals Functional delta method Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 5

Well-known: If T is Hadamard differentiable at F, then

the asymptotic distribution of T(Fn) follows immediately by the (functional) delta method from the asymptotic distribution of Fn − F.

Thus, delta method leads asymptotic distribution of

T(Fn) − T(F) whenever we have weak convergence of the empirical process.

Many results on weak convergence of the empirical

process (iid, short-memory like α-mixing or β-mixing, long memory, etc.).

slide-12
SLIDE 12

Quasi-Hadamard differentiability

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 6

slide-13
SLIDE 13

Illustrative example: sample mean

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 7

If K corresponds to Lebesgue measure on [0, 1], then

−∞

K(Fn(x)) dx + ∞ (1 − K(Fn(x)) dx, corresponds to the sample mean.

slide-14
SLIDE 14

Illustrative example: sample mean

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 7

If K corresponds to Lebesgue measure on [0, 1], then

−∞

K(Fn(x)) dx + ∞ (1 − K(Fn(x)) dx, corresponds to the sample mean.

Proving Hadamard differentiability of this L-statistic (at

F) in the direction of V boils down to prove that

  • [Vn(x) − V (x)] dx
  • → 0,

whenever ||Vn − V ||∞ → 0. || · ||∞ denotes sup-norm.

slide-15
SLIDE 15

Illustrative example: sample mean

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 7

If K corresponds to Lebesgue measure on [0, 1], then

−∞

K(Fn(x)) dx + ∞ (1 − K(Fn(x)) dx, corresponds to the sample mean.

Proving Hadamard differentiability of this L-statistic (at

F) in the direction of V boils down to prove that

  • [Vn(x) − V (x)] dx
  • → 0,

whenever ||Vn − V ||∞ → 0. || · ||∞ denotes sup-norm.

⇒ The simplest L-statistic the sample mean is not

Hadamard differentiable w.r.t. the sup-norm.

slide-16
SLIDE 16

Illustrative example: V-statistic

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 8

Proving Hadamard differentiability of a V-statistic (at F)

in the direction of V , involves among other things showing that ||Vn − V ||∞ → 0 implies

  • Vn(x2) |dgF|(x2) →
  • V (x2) |dgF|(x2).

where |dgF| is the absolute measure generated by gF(x2) =

  • g(x1, x2)dF(x1) .
slide-17
SLIDE 17

Illustrative example: V-statistic

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 8

Proving Hadamard differentiability of a V-statistic (at F)

in the direction of V , involves among other things showing that ||Vn − V ||∞ → 0 implies

  • Vn(x2) |dgF|(x2) →
  • V (x2) |dgF|(x2).

where |dgF| is the absolute measure generated by gF(x2) =

  • g(x1, x2)dF(x1) .

If gF generates a finite (signed) measure, then ||Vn − V ||∞

this implication indeed holds.

slide-18
SLIDE 18

Illustrative example: V-statistic

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 8

Proving Hadamard differentiability of a V-statistic (at F)

in the direction of V , involves among other things showing that ||Vn − V ||∞ → 0 implies

  • Vn(x2) |dgF|(x2) →
  • V (x2) |dgF|(x2).

where |dgF| is the absolute measure generated by gF(x2) =

  • g(x1, x2)dF(x1) .

If gF generates a finite (signed) measure, then ||Vn − V ||∞

this implication indeed holds.

For g(x1, x2) = (1/2)(x1 − x2)2 (the variance kernel) the

measure dgF has density (x2 − c) dx2.

slide-19
SLIDE 19

Illustrative example: V-statistic

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 8

Proving Hadamard differentiability of a V-statistic (at F)

in the direction of V , involves among other things showing that ||Vn − V ||∞ → 0 implies

  • Vn(x2) |dgF|(x2) →
  • V (x2) |dgF|(x2).

where |dgF| is the absolute measure generated by gF(x2) =

  • g(x1, x2)dF(x1) .

If gF generates a finite (signed) measure, then ||Vn − V ||∞

this implication indeed holds.

For g(x1, x2) = (1/2)(x1 − x2)2 (the variance kernel) the

measure dgF has density (x2 − c) dx2. ⇒ The simplest V-statistic the variance is not Hadamard differentiable w.r.t. the sup-norm.

slide-20
SLIDE 20

Way out & Problems

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 9

If we require not only that ||Vn − V ||∞ → 0 but that

(Vn(x) − V (x))(1 + |x|)λ, λ > 0, converges uniformly to zero,

slide-21
SLIDE 21

Way out & Problems

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 9

If we require not only that ||Vn − V ||∞ → 0 but that

(Vn(x) − V (x))(1 + |x|)λ, λ > 0, converges uniformly to zero, then we only need

  • (1 + |x|)−λ |dgF|(x) < ∞.
slide-22
SLIDE 22

Way out & Problems

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 9

If we require not only that ||Vn − V ||∞ → 0 but that

(Vn(x) − V (x))(1 + |x|)λ, λ > 0, converges uniformly to zero, then we only need

  • (1 + |x|)−λ |dgF|(x) < ∞.

Similar, for the sample mean.

slide-23
SLIDE 23

Way out & Problems

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 9

If we require not only that ||Vn − V ||∞ → 0 but that

(Vn(x) − V (x))(1 + |x|)λ, λ > 0, converges uniformly to zero, then we only need

  • (1 + |x|)−λ |dgF|(x) < ∞.

Similar, for the sample mean. However, Hadamard differentiability is defined as

"Let B1 and B2 be normed spaces. Then φ : B1 → B2 is Hadamard differentiable at b1 ∈ B1 if . . . "

slide-24
SLIDE 24

Way out & Problems

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 9

If we require not only that ||Vn − V ||∞ → 0 but that

(Vn(x) − V (x))(1 + |x|)λ, λ > 0, converges uniformly to zero, then we only need

  • (1 + |x|)−λ |dgF|(x) < ∞.

Similar, for the sample mean. However, Hadamard differentiability is defined as

"Let B1 and B2 be normed spaces. Then φ : B1 → B2 is Hadamard differentiable at b1 ∈ B1 if . . . "

But for an arbitrary df F we have

Fλ := F(x)(1 + |x|)λ = ∞.

slide-25
SLIDE 25

Way out & Problems

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 9

If we require not only that ||Vn − V ||∞ → 0 but that

(Vn(x) − V (x))(1 + |x|)λ, λ > 0, converges uniformly to zero, then we only need

  • (1 + |x|)−λ |dgF|(x) < ∞.

Similar, for the sample mean. However, Hadamard differentiability is defined as

"Let B1 and B2 be normed spaces. Then φ : B1 → B2 is Hadamard differentiable at b1 ∈ B1 if . . . "

But for an arbitrary df F we have

Fλ := F(x)(1 + |x|)λ = ∞.

Hence, with a weighted sup-norm · λ Hadamard

differentiability at F cannot be shown.

slide-26
SLIDE 26

Way out & Problems

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 9

If we require not only that ||Vn − V ||∞ → 0 but that

(Vn(x) − V (x))(1 + |x|)λ, λ > 0, converges uniformly to zero, then we only need

  • (1 + |x|)−λ |dgF|(x) < ∞.

Similar, for the sample mean. However, Hadamard differentiability is defined as

"Let B1 and B2 be normed spaces. Then φ : B1 → B2 is Hadamard differentiable at b1 ∈ B1 if . . . "

But for an arbitrary df F we have

Fλ := F(x)(1 + |x|)λ = ∞.

Hence, with a weighted sup-norm · λ Hadamard

differentiability at F cannot be shown.

⇒: (Functional) Delta method cannot be applied.

slide-27
SLIDE 27

Quasi-Hadamard differentiability

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 10

Definition: (Quasi-Hadamard differentiability)

Let V be a vector space, and V0 ⊂ V be equipped with a norm · V0. Let (V′, · V′) be a normed vector space, and T : VT → V′, VT ⊂ V.

slide-28
SLIDE 28

Quasi-Hadamard differentiability

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 10

Definition: (Quasi-Hadamard differentiability)

Let V be a vector space, and V0 ⊂ V be equipped with a norm · V0. Let (V′, · V′) be a normed vector space, and T : VT → V′, VT ⊂ V. Then T is said to be quasi-Hadamard differentiable at θ ∈ VT tangentially to C0, C0 ⊂ V0, if for a continuous map DHad

θ;T : C0 → V′

lim

n→∞

  • DHad

θ;T (v) − T(θ + hnvn) − T(θ)

hn

  • V′ = 0

holds for each triplet (v, (vn), (hn)) with hn → 0, and v ∈ C0, (vn) ⊂ V0 satisfying vn − vV0 → 0.

slide-29
SLIDE 29

Quasi-Hadamard differentiability

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 10

Definition: (Quasi-Hadamard differentiability)

Let V be a vector space, and V0 ⊂ V be equipped with a norm · V0. Let (V′, · V′) be a normed vector space, and T : VT → V′, VT ⊂ V. Then T is said to be quasi-Hadamard differentiable at θ ∈ VT tangentially to C0, C0 ⊂ V0, if for a continuous map DHad

θ;T : C0 → V′

lim

n→∞

  • DHad

θ;T (v) − T(θ + hnvn) − T(θ)

hn

  • V′ = 0

holds for each triplet (v, (vn), (hn)) with hn → 0, and v ∈ C0, (vn) ⊂ V0 satisfying vn − vV0 → 0.

With this definition we find

slide-30
SLIDE 30

Quasi-Hadamard differentiability (cont.)

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 11

Theorem: The distortion risk measure (L-statistic)

  • x dK(F(x)) is quasi-Hadamard differentiable if

K is continuous and piecewise differentiable, and K′ is bounded above by some constant M > 0.

slide-31
SLIDE 31

Quasi-Hadamard differentiability (cont.)

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 11

Theorem: The distortion risk measure (L-statistic)

  • x dK(F(x)) is quasi-Hadamard differentiable if

K is continuous and piecewise differentiable, and K′ is bounded above by some constant M > 0.

Theorem: The V-statistic

g(x1, x2) dF(x1)dF(x2) is quasi-Hadamard differentiable if for some λ > λ′ ≥ 0 (a) For every x2 ∈ R fixed, the function gx2(·) := g( · , x2) lies in BVloc,rc and gx2(x1)(1 + |x1|)−λ′ is uniformly bounded. (b) The function gF(·) :=

  • g(x1, · )dF(x1) lies in

BVloc,rc, and

  • φ−λ(x) |dgF|(x) < ∞.
slide-32
SLIDE 32

Quasi-Hadamard differentiability (cont.)

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 11

Theorem: The distortion risk measure (L-statistic)

  • x dK(F(x)) is quasi-Hadamard differentiable if

K is continuous and piecewise differentiable, and K′ is bounded above by some constant M > 0.

Theorem: The V-statistic

g(x1, x2) dF(x1)dF(x2) is quasi-Hadamard differentiable if for some λ > λ′ ≥ 0 (a) For every x2 ∈ R fixed, the function gx2(·) := g( · , x2) lies in BVloc,rc and gx2(x1)(1 + |x1|)−λ′ is uniformly bounded. (b) The function gF(·) :=

  • g(x1, · )dF(x1) lies in

BVloc,rc, and

  • φ−λ(x) |dgF|(x) < ∞.

Notice sample mean and variance (with λ′ = 2) are

quasi-Hadamard differentiable. However, results might be completely useless.

slide-33
SLIDE 33

Modified FDM

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 12

Fortunately, they are not. Because

slide-34
SLIDE 34

Modified FDM

Motivation Quasi-Hadamard differentiability Illustrative example: sample mean Illustrative example: V-statistic Way out & Problems Quasi-Hadamard differentiability Quasi-Hadamard differentiability (cont.) Modified FDM Applications Continuous mapping approach to U- and V-statistics

Tokyo, 2013 12

Fortunately, they are not. Because Theorem: (Modified functional delta method)

Let T, θ, V, Vf, V0, C0 be as above. If: (i) T is quasi-Hadamard differentiable at θ tangentially to C0 with quasi-Hadamard derivative DHad

θ;T ,

(ii) Xn − θ takes values only in V0 and satisfies an(Xn − θ)

d

→ V (in (V0, V0, · V0)), V a random element of (V0, V0) taking values only in C0. Then an(T(Xn) − T(θ))

d

→ DHad

θ;T (V )

(in (V′, V′, · V′)).

slide-35
SLIDE 35

Applications

Motivation Quasi-Hadamard differentiability Applications Weakly dependent data Strongly dependent data Strongly dependent data (cont.) Continuous mapping approach to U- and V-statistics

Tokyo, 2013 13

slide-36
SLIDE 36

Weakly dependent data

Motivation Quasi-Hadamard differentiability Applications Weakly dependent data Strongly dependent data Strongly dependent data (cont.) Continuous mapping approach to U- and V-statistics

Tokyo, 2013 14

The quasi-Hadamard derivative of a U-statistic is given by

˙ UF(B◦) := −2

  • B◦(x)dgF(x).
slide-37
SLIDE 37

Weakly dependent data

Motivation Quasi-Hadamard differentiability Applications Weakly dependent data Strongly dependent data Strongly dependent data (cont.) Continuous mapping approach to U- and V-statistics

Tokyo, 2013 14

The quasi-Hadamard derivative of a U-statistic is given by

˙ UF(B◦) := −2

  • B◦(x)dgF(x).

Let (Xi) be α-mixing with α(n) = O(n−θ) for some

θ > 1 + √

  • 2. If F has finite γ-moment for some γ > 2θλ

θ−1,

then (Shao and Yu (1996)) with Dλ càdlàg functions with finite weighted sup-norm √n(Fn − F)

d

→ B◦

F

(in (Dλ, Dλ, · λ)).

slide-38
SLIDE 38

Weakly dependent data

Motivation Quasi-Hadamard differentiability Applications Weakly dependent data Strongly dependent data Strongly dependent data (cont.) Continuous mapping approach to U- and V-statistics

Tokyo, 2013 14

The quasi-Hadamard derivative of a U-statistic is given by

˙ UF(B◦) := −2

  • B◦(x)dgF(x).

Let (Xi) be α-mixing with α(n) = O(n−θ) for some

θ > 1 + √

  • 2. If F has finite γ-moment for some γ > 2θλ

θ−1,

then (Shao and Yu (1996)) with Dλ càdlàg functions with finite weighted sup-norm √n(Fn − F)

d

→ B◦

F

(in (Dλ, Dλ, · λ)).

Asymptotic distribution of √n(U(Fn) − U(F)) follows

then for every df with finite γ-moment for some γ > 2θλ

θ−1.

slide-39
SLIDE 39

Weakly dependent data

Motivation Quasi-Hadamard differentiability Applications Weakly dependent data Strongly dependent data Strongly dependent data (cont.) Continuous mapping approach to U- and V-statistics

Tokyo, 2013 14

The quasi-Hadamard derivative of a U-statistic is given by

˙ UF(B◦) := −2

  • B◦(x)dgF(x).

Let (Xi) be α-mixing with α(n) = O(n−θ) for some

θ > 1 + √

  • 2. If F has finite γ-moment for some γ > 2θλ

θ−1,

then (Shao and Yu (1996)) with Dλ càdlàg functions with finite weighted sup-norm √n(Fn − F)

d

→ B◦

F

(in (Dλ, Dλ, · λ)).

Asymptotic distribution of √n(U(Fn) − U(F)) follows

then for every df with finite γ-moment for some γ > 2θλ

θ−1.

For the variance the assumptions are weaker than in

Dehling and Wendler (2010) whenever γ < 7+8

√ 2 2 √ 2−1.

slide-40
SLIDE 40

Strongly dependent data

Motivation Quasi-Hadamard differentiability Applications Weakly dependent data Strongly dependent data Strongly dependent data (cont.) Continuous mapping approach to U- and V-statistics

Tokyo, 2013 15

Consider the process Xt := ∞

s=0 as εt−s, t ∈ N0,, where

(εi)i∈Z iid random variables with zero mean and finite variance, and ∞

s=0 a2 s < ∞.

slide-41
SLIDE 41

Strongly dependent data

Motivation Quasi-Hadamard differentiability Applications Weakly dependent data Strongly dependent data Strongly dependent data (cont.) Continuous mapping approach to U- and V-statistics

Tokyo, 2013 15

Consider the process Xt := ∞

s=0 as εt−s, t ∈ N0,, where

(εi)i∈Z iid random variables with zero mean and finite variance, and ∞

s=0 a2 s < ∞.

If Cov(X0, Xm) = m1−2β, β ∈ (0.5, 1), then

m=1 Cov(X0, Xm) is not absolute summable, and the

process (Xt) is called a long-memory process.

slide-42
SLIDE 42

Strongly dependent data

Motivation Quasi-Hadamard differentiability Applications Weakly dependent data Strongly dependent data Strongly dependent data (cont.) Continuous mapping approach to U- and V-statistics

Tokyo, 2013 15

Consider the process Xt := ∞

s=0 as εt−s, t ∈ N0,, where

(εi)i∈Z iid random variables with zero mean and finite variance, and ∞

s=0 a2 s < ∞.

If Cov(X0, Xm) = m1−2β, β ∈ (0.5, 1), then

m=1 Cov(X0, Xm) is not absolute summable, and the

process (Xt) is called a long-memory process.

No general result for L- and V-statistics of such processes.

slide-43
SLIDE 43

Strongly dependent data (cont.)

Motivation Quasi-Hadamard differentiability Applications Weakly dependent data Strongly dependent data Strongly dependent data (cont.) Continuous mapping approach to U- and V-statistics

Tokyo, 2013 16

Because quasi-Hadamard differentiability already

established, to apply the Modified Functional Delta Method, we only have

slide-44
SLIDE 44

Strongly dependent data (cont.)

Motivation Quasi-Hadamard differentiability Applications Weakly dependent data Strongly dependent data Strongly dependent data (cont.) Continuous mapping approach to U- and V-statistics

Tokyo, 2013 16

Because quasi-Hadamard differentiability already

established, to apply the Modified Functional Delta Method, we only have to prove weak convergence of weighted empirical processes based on long-memory sequences.

slide-45
SLIDE 45

Strongly dependent data (cont.)

Motivation Quasi-Hadamard differentiability Applications Weakly dependent data Strongly dependent data Strongly dependent data (cont.) Continuous mapping approach to U- and V-statistics

Tokyo, 2013 16

Because quasi-Hadamard differentiability already

established, to apply the Modified Functional Delta Method, we only have to prove weak convergence of weighted empirical processes based on long-memory sequences.

Theorem Let λ ≥ 0, β ∈ (0.5, 1), and assume that

E[|ε0|2+2λ] < ∞, the df G of ε0 is twice differentiable, and 2

j=1

  • |G(j)(x)|2(1 + |x|2λ) dx < ∞.

Then nβ−1/2 Fn(·) − F(·)

  • d

− → −c1,β f(·)Z (in Dλ), where f is the density of X0 and Z is normally distributed with mean 0 and variance 1.

slide-46
SLIDE 46

Continuous mapping approach to U- and V-statistics

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics Motivation Expansion Applications to V-statistics Applications to V-statistics Example I Example II References References

Tokyo, 2013 17

slide-47
SLIDE 47

Motivation

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics Motivation Expansion Applications to V-statistics Applications to V-statistics Example I Example II References References

Tokyo, 2013 18

For the variance & long memory the above approach leads

that the asymptotic distribution of the sample variance multiplied by the rate of the empirical process equals −2

  • B◦(x−) dgF(x) = 2Z1,β
  • f(x−) (x−E[X1]) dx = 0
slide-48
SLIDE 48

Motivation

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics Motivation Expansion Applications to V-statistics Applications to V-statistics Example I Example II References References

Tokyo, 2013 18

For the variance & long memory the above approach leads

that the asymptotic distribution of the sample variance multiplied by the rate of the empirical process equals −2

  • B◦(x−) dgF(x) = 2Z1,β
  • f(x−) (x−E[X1]) dx = 0

In particular for long memory the follwong representation

for U-statistics turns out to be useful an

  • Vg(Fn) − Vg(F)
  • =

2Φ1,g

  • an(Fn − F)
  • +Φ2,g

√an(Fn − F)

  • ,

where Φ1,g(f) := −

  • f(x−) dgF(x) and

Φ2,g(f) :=

  • f(x1−)f(x2−) dg(x1, x2) are continuous

mappings for appropriate weigthed sup-norms.

slide-49
SLIDE 49

Expansion

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics Motivation Expansion Applications to V-statistics Applications to V-statistics Example I Example II References References

Tokyo, 2013 19

For the following expansion (with p ≥ 1)

Fn(·) − F(·) −

p

  • j=1

(−1)j F (j)(·) 1 n

n

  • i=1

Aj;F(Xi)

  • ,

where Aj;F denotes the jth order Appell polynomial associated with F and F (j) is the jth derivative of F, weak convergence at the rate np(β−1/2) to (−1)p F (p)(·)Zp,β in a weighted sup-norm can be shown.

slide-50
SLIDE 50

Expansion

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics Motivation Expansion Applications to V-statistics Applications to V-statistics Example I Example II References References

Tokyo, 2013 19

For the following expansion (with p ≥ 1)

Fn(·) − F(·) −

p

  • j=1

(−1)j F (j)(·) 1 n

n

  • i=1

Aj;F(Xi)

  • ,

where Aj;F denotes the jth order Appell polynomial associated with F and F (j) is the jth derivative of F, weak convergence at the rate np(β−1/2) to (−1)p F (p)(·)Zp,β in a weighted sup-norm can be shown.

Then we can introduce the following statistic

slide-51
SLIDE 51

Applications to V-statistics

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics Motivation Expansion Applications to V-statistics Applications to V-statistics Example I Example II References References

Tokyo, 2013 20 Vn,g;p,q,r(Fn) := Vg(Fn) − Vg(F) +

2

  • ℓ=1

p−1

  • j=1

(−1)j 1 n

n

  • i=1

Aj;F (Xi) F (j)(x−) dgℓ,F (x) −

q−1

  • j=1

(−1)j 1 n

n

  • i=1

Aj;F (Xi)

  • ×
  • F (j)(x1−) (Fn(x2−) − F(x2−)) dg(x1, x2)

r−1

  • k=1

(−1)k 1 n

n

  • i=1

Ak;F (Xi)

  • ×
  • (Fn(x1−) − F(x1−)) F (k)(x2−) dg(x1, x2)

+

q−1

  • j=1

r−1

  • k=1

(−1)j+k 1 n

n

  • i=1

Aj;F (Xi) 1 n

n

  • i=1

Ak;F (Xi)

  • ×
  • F (j)(x1−) F (k)(x2−) dg(x1, x2).
slide-52
SLIDE 52

Applications to V-statistics

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics Motivation Expansion Applications to V-statistics Applications to V-statistics Example I Example II References References

Tokyo, 2013 21

Using the continuous mapping approach and the above result:

(i) Assume q + r > p, then np(β−1/2) Vn,g;p,q,r(Fn)

converges in distribution to (−1)p Zp,β

2

  • ℓ=1
  • F (p)(x−) dgℓ,F(x).
slide-53
SLIDE 53

Applications to V-statistics

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics Motivation Expansion Applications to V-statistics Applications to V-statistics Example I Example II References References

Tokyo, 2013 21

Using the continuous mapping approach and the above result:

(i) Assume q + r > p, then np(β−1/2) Vn,g;p,q,r(Fn)

converges in distribution to (−1)p Zp,β

2

  • ℓ=1
  • F (p)(x−) dgℓ,F(x).

(ii) Assume q + r = p, then np(β−1/2) Vn,g;p,q,r(Fn)

converges in distribution to (−1)p Zp,β

2

  • ℓ=1
  • F (p)(x−) dgℓ,F(x) +

(−1)p Zq,βZr,β

  • F (q)(x1−)F (r)(x2−) dg(x1, x2).
slide-54
SLIDE 54

Example I

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics Motivation Expansion Applications to V-statistics Applications to V-statistics Example I Example II References References

Tokyo, 2013 22

Consider kernel g(x1, x2) = x1(|x2| − 1), and suppose

that F (1) is symmetric about zero and that E[|X1|] = 1.

slide-55
SLIDE 55

Example I

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics Motivation Expansion Applications to V-statistics Applications to V-statistics Example I Example II References References

Tokyo, 2013 22

Consider kernel g(x1, x2) = x1(|x2| − 1), and suppose

that F (1) is symmetric about zero and that E[|X1|] = 1.

Taking n2(β−(1/2)) leads to:

n2β−1 Vn,g;2,1,1(Fn) = n2β−1 Vg(Fn) − Vg(F)

  • d

− → Z2

1,β

  • F (1)(x1−)F (1)(x2−) dg(x1, x2) = 0.
slide-56
SLIDE 56

Example I

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics Motivation Expansion Applications to V-statistics Applications to V-statistics Example I Example II References References

Tokyo, 2013 22

Consider kernel g(x1, x2) = x1(|x2| − 1), and suppose

that F (1) is symmetric about zero and that E[|X1|] = 1.

Taking n2(β−(1/2)) leads to:

n2β−1 Vn,g;2,1,1(Fn) = n2β−1 Vg(Fn) − Vg(F)

  • d

− → Z2

1,β

  • F (1)(x1−)F (1)(x2−) dg(x1, x2) = 0.

However, with n3(β−(1/2)) we have

n3(β−1/2) Vn,g;3,1,2(Fn) = n3(β−1/2) Vg(Fn) − Vg(F)

  • d

− → − Z1,βZ2,β

  • F (1)(x1−)F (2)(x2−) dg(x1, x2)

= −2 Z1,βZ2,β ∞ F (2)(x2) dx2.

slide-57
SLIDE 57

Example II

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics Motivation Expansion Applications to V-statistics Applications to V-statistics Example I Example II References References

Tokyo, 2013 23

Consider the test statistic

Tn := ∞

  • ˆ

Fn(−t) −

  • 1 − ˆ

Fn(t−) 2 dt.

slide-58
SLIDE 58

Example II

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics Motivation Expansion Applications to V-statistics Applications to V-statistics Example I Example II References References

Tokyo, 2013 23

Consider the test statistic

Tn := ∞

  • ˆ

Fn(−t) −

  • 1 − ˆ

Fn(t−) 2 dt.

Taking n3(β−1/2) leads to:

n3(β−1/2) Vn,g;3,1,2(Fn)

d

− → Z1,βZ2,β

  • F (1)(x)F (2)(x) − F (1)(x)F (2)(−x) dx
  • = 0.
slide-59
SLIDE 59

Example II

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics Motivation Expansion Applications to V-statistics Applications to V-statistics Example I Example II References References

Tokyo, 2013 23

Consider the test statistic

Tn := ∞

  • ˆ

Fn(−t) −

  • 1 − ˆ

Fn(t−) 2 dt.

Taking n3(β−1/2) leads to:

n3(β−1/2) Vn,g;3,1,2(Fn)

d

− → Z1,βZ2,β

  • F (1)(x)F (2)(x) − F (1)(x)F (2)(−x) dx
  • = 0.

However, with n4(β−1/2) we find

n4(β−1/2) Vn,g;4,2,2(Fn)

d

− → Z2,βZ2,β F (2)(x)F (2)(x) − F (2)(x)F (2)(−x) dx

slide-60
SLIDE 60

References

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics Motivation Expansion Applications to V-statistics Applications to V-statistics Example I Example II References References

Tokyo, 2013 24

Beutner, E. and Zähle, H. (2010). A modified functional

delta method and its application to the estimation of risk

  • functionals. J. Multivariate Anal. 101, 2452–2463.

Beutner, E. and Zähle, H. (2012). Deriving the asymptotic

distribution of U- and V-statistics of dependent data using weighted empirical processes, Bernoulli 18, 803–822.

Beutner, E., Wu, W.B. and Zähle, H. (2012). Asymptotics

for statistical functionals of long-memory sequences. Stochastic Process. Appl. 122, 910–929.

Beutner, E. and Zähle, H. Continuous mapping approach

to the asymptotics of U- and V-statistics, Bernoulli, to appear.

slide-61
SLIDE 61

References

Motivation Quasi-Hadamard differentiability Applications Continuous mapping approach to U- and V-statistics Motivation Expansion Applications to V-statistics Applications to V-statistics Example I Example II References References

Tokyo, 2013 25

Dehling, H. and Wendler, M. (2010). Central limit

theorem and the bootstrap for U-statistics of strongly mixing data. Journal of Multivariate Analysis, 101(1), 126–137.

Sen, P.K. (1996). Statistical functionals, Hadamard

differentiability and martingales. In A Festschrift for

  • J. Medhi (Eds. Borthakur, A.C, and Chaudhury, H.), New

Age Press, Delhi, 29–47.

Shao, Q.-M. and Yu, H. (1996). Weak convergence for

weighted empirical processes of dependent sequences.

  • Ann. Probab. 24, 2098–2127.

Wu, W.B. (2003). Empirical processes of long-memory

  • sequences. Bernoulli, 9, 809–831.