Distributional Results for Thresholding Estimators in - - PowerPoint PPT Presentation

distributional results for thresholding estimators in
SMART_READER_LITE
LIVE PREVIEW

Distributional Results for Thresholding Estimators in - - PowerPoint PPT Presentation

Distributional Results for Thresholding Estimators in High-Dimensional Gaussian Regression Ulrike Schneider University of G ottingen Workshop on High-Dimensional Problems in Statistics ETH Z urich September 23, 2011 Joint work with


slide-1
SLIDE 1

Distributional Results for Thresholding Estimators in High-Dimensional Gaussian Regression

Ulrike Schneider

University of G¨

  • ttingen

Workshop on High-Dimensional Problems in Statistics ETH Z¨ urich September 23, 2011

Joint work with Benedikt P¨

  • tscher (University of Vienna)
slide-2
SLIDE 2

Penalized LS (ML) Estimators

Linear regression model y = θ1 x.1 + . . . θk x.k + ε = Xθ + ε response y ∈ Rn regressors x.i ∈ Rn, 1 ≤ i ≤ k errors ε ∈ Rn parameter vector θ = (θ1, . . . , θk)′ ∈ Rk A penalized least-squares (PLSE) or maximum-likelihood estimator (PMLE) ˆ θ for θ is given by ˆ θ = arg min

θ∈Rk

y − Xθ2

  • likelihood or LS -part

+ Pn(θ)

penalty

, where X = [x.1, . . . , x.k] is the n × k regressor matrix.

1 / 1

slide-3
SLIDE 3

Penalized LS (ML) Estimators

(cont’d)

General class of Bridge-estimators (Frank & Friedman, 1993) Pn(θ) = λn

k

  • i=1

|θi|γ

γ = 2:

Ridge-estimator (Hoerl & Kennard, 1970)

γ = 1:

Lasso (Tibshirani, 1996). Hard- and soft-thresholding estimators. SCAD estimator (Fan & Li, 2001) Elastic-net estimator (Zou & Hastie, 2005) Adaptive Lasso estimator (Zou, 2006) (thresholded) Lasso with refitting (Van de Geer et al, 2010; Belloni & Chernozhukov, 2011) MCP (Zhang, 2010) . . .

2 / 1

slide-4
SLIDE 4

Relationship to classical PMS-estimators

Brigde-estimators satisfy min

θ∈Rk y − Xθ2 + λn k

  • i=1

|θi|γ (0 < γ < ∞) For γ → 0, get min

θ∈Rk y − Xθ2 + λn card{i : θi = 0}

which yields a minimum Cp-type procedure such as AIC and BIC. (lγ-type penalty with “γ = 0”) → ‘classical’ post-modelselection (PMS) estimators.

3 / 1

slide-5
SLIDE 5

Relationship to classical PMS-estimators

(cont’d)

For “γ = 0” procedures are computationally expensive. For γ > 0 (Bridge) estimators are more computationally tractable, especially for γ ≥ 1 (convex objective function). For γ ≤ 1, estimators perform model selection P(ˆ θi = 0) > 0 if θi = 0. Phenomenon is more pronounced for smaller γ. γ = 1 (Lasso and adaptive Lasso) as compromise between the wish to detect zeros and computational simplicity. The PLSEs (and thresholding estimators) we treat in the following can be viewed to simultaneously perform model selection and parameter estimation.

4 / 1

slide-6
SLIDE 6

Some terminology

Consistent model selection

lim

n→∞ P(ˆ

θi = 0) = 1 whenever θi = 0 (1 ≤ i ≤ k)

Estimator is sparse or sparsely tuned. Conservative model selection

lim

n→∞ P(ˆ

θi = 0) < 1 whenever θi = 0 (1 ≤ i ≤ k)

Estimator is non-sparsely tuned. Consistent vs. conservative model selection can in our context be driven by the (asymptotic) behavior of the tuning parameter.

5 / 1

slide-7
SLIDE 7

Literature on distributional properties of PLSEs

  • fixed-parameter asymptotic framework (non-uniformity issues)
  • sparsely-tuned PLSEs

Oracle property – obtain same asymptotic distribution as ‘oracle estimator’ (infeasible unpenalized estimator using the true zero restrictions). Fan & Li, 2001. (SCAD) Zou, 2006. (Lasso and adaptive Lasso)

Cai, Fan, Li & Zhou (2002), Fan & Li (2002, 2004), Bunea (2004), Fan & Peng (2006), Bunea & McKeague (2005), Hunter & Li (2005), Fan, Li & Zhou (2006), Wang & Leng (2007), Wang, G. Li, & Tsai (2007), Zhang & Lu (2007), Wang, R. Li, & Tsai (2007), Huang, Horowitz & Ma (2008), Li & Liang (2008), Zou & Yuan (2008), Zou & Li (2008), Johnson, Lin, & Zeng (2008), Zou & Li (2008), Zou & Yuan (2008), Lin, Xiang & Zhang (2009), Xie & Huang (2009), Zhu & Zhu (2009), Zou & Zhang (2009) . . .

6 / 1

slide-8
SLIDE 8

Literature on distributional properties of PLSEs

(cont’d)

  • moving-parameter asymptotic framework (taking non-uniformity

into account)

  • Sparsely and non-sparsely tuned PLSEs.

Knight & Fu, 2000. (Non-sparsely tuned Lasso and Bridge estimators for q < 1 in general.) P¨

  • tscher & Leeb (2009), P¨
  • tscher & S., (2009), P¨
  • tscher &
  • S. (2010), P¨
  • tscher & S. (2011).

7 / 1

slide-9
SLIDE 9

Assumptions and Notation

y = Xθ + ε X is non-stochastic (n × k), rk(X) = k (⇒ k ≤ n). No further assumptions on X. k may vary with n. ε ∼ Nn(0, σ2In) Notation: ξ2

i,n :=

  • (X ′X/n)−1

i,i

(X ′X = nIk ⇒ ξi,n = 1) ˆ θLS = (X ′X)−1X ′y ˆ σ2

LS = y − X ˆ

θLS2/(n − k) Consider 3 estimators: hard-, soft- and adaptive soft-thresholding acting componentwise.

8 / 1

slide-10
SLIDE 10

Hard-thresholding ˜ θH,i

˜ θH,i = ˆ θLS,i 1(|ˆ θLS,i| > ˆ σLSξi,nηi,n)

9 / 1

slide-11
SLIDE 11

Hard-thresholding ˜ θH,i

˜ θH,i = ˆ θLS,i 1(|ˆ θLS,i| > ˆ σLSξi,nηi,n)

  • rthogonal case:

equivalent to a pretest estimator based on t-tests or Cp crite- rion such as AIC, BIC (classical post-model selection estima- tor) with penalty term

Pn(θ) = k

i=1 n

σLSξi,nηi,n)2 − (|θi| − ˆ σLSξi,nηi,n)2 1(|θi| < ˆ σLSξi,nηi,n)

  • also equivalent to MCP

9 / 1

slide-12
SLIDE 12

Soft-thresholding ˜ θS,i

˜ θS,i = sign(ˆ θLS,i) (|ˆ θLS,i| − ˆ σLSξi,nηi,n)+

10 / 1

slide-13
SLIDE 13

Soft-thresholding ˜ θS,i

˜ θS,i = sign(ˆ θLS,i) (|ˆ θLS,i| − ˆ σLSξi,nηi,n)+

  • rthogonal case:

equivalent to Lasso with penalty term

Pn(θ) = 2nˆ σLS k

i=1 ξi,nηi,n|θi|

also equivalent to Dantzig selector

10 / 1

slide-14
SLIDE 14

Adaptive soft-thresholding ˜ θAS,i

˜ θAS,i =

  • if |ˆ

θLS,i| ≤ ˆ σLSξi,nηi,n ˆ θLS,i − (ˆ σLSξi,nηi,n)2/ˆ θLS,i if |ˆ θLS,i| > ˆ σLSξi,nηi,n

11 / 1

slide-15
SLIDE 15

Adaptive soft-thresholding ˜ θAS,i

˜ θAS,i =

  • if |ˆ

θLS,i| ≤ ˆ σLSξi,nηi,n ˆ θLS,i − (ˆ σLSξi,nηi,n)2/ˆ θLS,i if |ˆ θLS,i| > ˆ σLSξi,nηi,n

  • rthogonal case:

equivalent to adaptive Lasso with penalty term

Pn(θ) = 2nˆ σ2

LS

k

i=1(ξi,nηi,n)2|θi|/|ˆ

θLS,i|

also equivalent to non-negative Garotte (Breiman, 1995)

11 / 1

slide-16
SLIDE 16

“Infeasible” versions

Known-variance case: ˆ θH,i = ˆ θLS,i 1(|ˆ θLS,i| > σξi,nηi,n) ˆ θS,i = sign(ˆ θLS,i) (|ˆ θLS,i| − σξi,nηi,n)+ ˆ θAS,i =

  • if |ˆ

θLS,i| ≤ σξi,nηi,n ˆ θLS,i − (σξi,nηi,n)2/ˆ θLS,i if |ˆ θLS,i| > σξi,nηi,n

12 / 1

slide-17
SLIDE 17

Variable selection

We shall assume that sup ξi,n/n1/2 < ∞. Let ˇ θi stand for any of the estimators ˆ θH,i, ˆ θS,i, ˆ θAS,i, ˜ θH,i, ˜ θS,i, ˜ θAS,i. Variable selection

Pn,θ,σ(ˇ θi = 0) → 0 for any θ with θi = 0 ⇐ ⇒ ξi,nηi,n → 0 Pn,θ,σ(ˇ θi = 0) → 1 for any θ with θi = 0 ⇐ ⇒ n1/2ηi,n → ∞ Pn,θ,σ(ˇ θi = 0) → ci < 1 for any θ with θi = 0 ⇐ ⇒ n1/2ηi,n → ei

with 0 ≤ ei < ∞

1 (ξi,nηi,n → 0 and) n1/2ηi,n → ei < ∞ leads to (sensible) conserva-

tive selection.

2 (ξi,nηi,n → 0 and) n1/2ηi,n → ∞ leads to (sensible) consistent

selection.

13 / 1

slide-18
SLIDE 18

Parameter estimation, minimax rate

Consistency

ˇ θi is consistent for θi ⇐ ⇒ ξi,nηi,n → 0 and ξi,n/n1/2 → 0

Suppose ξi,nηi,n → 0 and ξi,n/n1/2 → 0. Then ˇ

θi is uniformly

consistent for θi in the sense that for all ε > 0 there exists a real number M > 0 such that

sup

n∈N

sup

θ∈Rk

sup

0<σ<∞

Pn,θ,σ(|ˇ θi − θi| > σM) < ε

Suppose ξi,nηi,n → 0, ξi,n/n1/2 → 0, and bi,n ≥ 0. If for all ε > 0 there exists a real number M > 0 such that

sup

n∈N

sup

θ∈Rk

sup

0<σ<∞

Pn,θ,σ(bi,n|ˇ θi − θi| > σM) < ε.

Then bi,n = O(ai,n), where ai,n = min(n1/2/ξi,n, (ξi,nηi,n)−1)

14 / 1

slide-19
SLIDE 19

Parameter estimation, minimax rate

(cont’d)

Minimax rate is

1

ξi,n/n1/2 in the conservative case, and

2 only ξi,nηi,n = o(ξi,n/n1/2) in the consistent case. 15 / 1

slide-20
SLIDE 20

Finite sample distribution: hard-thresholding ˆ θH,i

F i

H,n,θ,σ(x) = Pn,θ,σ(αi,n/σ(ˆ

θH,i − θi) ≤ x) (known-variance case) dF i

H,n,θ,σ(x) =

  • Φ(n1/2(−θi/(σξi,n) + ηi,n)) − Φ(n1/2(−θi/(σξi,n) − ηi,n))
  • dδ−αi,nθi/σ(x)

+ n1/2/(αi,nξi,n) φ(n1/2x/(αi,nξi,n)) 1(|α−1

i,n x + θi/σ| > ξi,nηi,n) dx,

where φ and Φ are the pdf and cdf of N(0, 1), resp.

16 / 1

slide-21
SLIDE 21

Finite sample distribution: hard-thresholding ˆ θH,i

n = 40, ηi,n = 0.05, θi = 0.16, ξi,n = 1, σ = 1, αi,n = n1/2/ξi,n

16 / 1

slide-22
SLIDE 22

Finite sample distribution: hard-thresholding ˜ θH,i

˜ F i

H,n,θ,σ(x) = Pn,θ,σ(αi,n/σ(˜

θH,i − θi) ≤ x) (unknown-variance case) d ˜ F i

H,n,θ,σ(x) =

∞ {Φ(n1/2(−θi/(σξi,n) + sηi,n)−Φ(n1/2(−θi/(σξi,n) − sηi,n)}ρn−k(s) ds dδ−αi,nθi/σ(x) + n1/2α−1

i,n ξ−1 i,n φ(n1/2x/(αi,nξi,n))

∞ 1(|α−1

i,n x + θi/σ| > ξi,nsηi,n)ρn−k(s) ds dx,

where ρn−k is the density of

  • χ2

n−k/(n − k).

17 / 1

slide-23
SLIDE 23

Finite sample distribution: hard-thresholding ˜ θH,i

.

n = 40, k = 35, ηi,n = 0.05, θi = 0.16, ξi,n = 1, σ = 1, αi,n = n1/2/ξi,n

17 / 1

slide-24
SLIDE 24

Finite sample distribution: soft-thresholding ˆ θS,i

F i

S,n,θ,σ(x) = Pn,θ,σ(αi,n/σ(ˆ

θS,i − θi) ≤ x) (known-variance case) dF i

S,n,θ,σ(x) =

  • Φ(n1/2(−θi/(σξi,n) + ηi,n)) − Φ(n1/2(−θi/(σξi,n) − ηi,n))
  • dδ−αi,nθi/σ(x)

+ n1/2/(αi,nξi,n)

  • φ(n1/2/(αi,nξi,n)x + n1/2ηi,n)1(α−1

i,n x + θi/σ > 0)

+ φ(n1/2/(αi,nξi,n)x − n1/2ηi,n)1(α−1

i,n x + θi/σ < 0)

  • dx

18 / 1

slide-25
SLIDE 25

Finite sample distribution: soft-thresholding ˆ θS,i

n = 40, ηi,n = 0.05, θi = 0.16, ξi,n = 1, σ = 1, αi,n = n1/2/ξi,n

18 / 1

slide-26
SLIDE 26

Finite sample distribution: soft-thresholding ˜ θS,i

˜ F i

S,n,θ,σ(x) = Pn,θ,σ(αi,n(˜

θS,i − θi) ≤ x) (unknown-variance case) d ˜ F i

S,n,θ,σ(x) =

∞ {Φ(n1/2(−θi/(σξi,n) + sηi,n)−Φ(n1/2(−θi/(σξi,n) − sηi,n)}ρn−k(s) ds dδ−αi,nθi/σ(x) + n1/2/(αi,nξi,n) ∞ φ(n1/2/(αi,nξi,n)x + n1/2sηi,n)1(α−1

i,n x + θi/σ > 0)

+ φ(n1/2/(αi,nξi,n)x − n1/2sηi,n)1(α−1

i,n x + θi/σ < 0)

  • ρn−k(s) ds dx

19 / 1

slide-27
SLIDE 27

Finite sample distribution: soft-thresholding ˜ θS,i

n = 40, k = 35, ηi,n = 0.05, θi = 0.16, ξi,n = 1, σ = 1, αi,n = n1/2/ξi,n

19 / 1

slide-28
SLIDE 28

Finite sample distribution: adaptive soft-thresholding ˆ θAS,i

F i

AS,n,θ,σ(x) = Pn,θ,σ(αi,n/σ(ˆ

θAS,i − θi) ≤ x) (known-variance case) dF i

AS,n,θ,σ(x) =

  • Φ(n1/2(−θi/(σξi,n) + ηi,n)) − Φ(n1/2(−θi/(σξi,n) − ηi,n))
  • dδ−αi,nθi/σ(x)

+ (0.5n1/2/(αi,nξi,n))

  • φ(z(2)

n,θ,σ(x, ηi,n))(1 + tn,θ,σ(x, ηi,n))1(α−1

i,n x + θi/σ > 0)

+ φ(z(1)

n,θ,σ(x, ηi,n))(1 − tn,θ,σ(x, ηi,n))1(α−1

i,n x + θi/σ < 0)

  • dx,

where z(1,2)

n,θ,σ(x, y) =

0.5n1/2ξ−1

i,n (α−1 i,n x − θi/σ) ± n1/2

  • (0.5ξ−1

i,n (α−1 i,n x + θi/σ))2 + y 2 and

tn,θ,σ(x, y) = 0.5ξ−1

i,n (α−1 i,n x + θi/σ)/((0.5ξ−1 i,n (α−1 i,n x + θi/σ))2 + y 2)1/2. 20 / 1

slide-29
SLIDE 29

Finite sample distribution: adaptive soft-thresholding ˆ θAS,i

n = 40, ηi,n = 0.05, θi = 0.16, ξi,n = 1, σ = 1, αi,n = n1/2/ξi,n

20 / 1

slide-30
SLIDE 30

Finite sample distribution: adaptive soft-thresholding ˜ θAS,i

˜ F i

AS,n,θ,σ(x) = Pn,θ,σ(αi,n(˜

θAS,i − θi) ≤ x) (unknown-variance case) d ˜ F i

AS,n,θ,σ(x) =

∞ {Φ(n1/2(−θi/(σξi,n) + sηi,n)−Φ(n1/2(−θi/(σξi,n) − sηi,n)}ρn−k(s) ds dδ−αi,nθi/σ(x) + (0.5n1/2/(αi,nξi,n)) ∞ φ(z(2)

n,θ,σ(x, ηi,n))(1+tn,θ,σ(x, ηi,n))1(α−1

i,n x + θi/σ > 0)

+ φ(z(1)

n,θ,σ(x, ηi,n))(1 − tn,θ,σ(x, ηi,n))1(α−1

i,n x + θi/σ < 0)

  • ρn−k(s) ds dx,

21 / 1

slide-31
SLIDE 31

Finite sample distribution: adaptive soft-thresholding ˜ θAS,i

n = 40, k = 35, ηi,n = 0.05, θi = 0.16, ξi,n = 1, σ = 1, αi,n = n1/2/ξi,n

21 / 1

slide-32
SLIDE 32

Large sample distributions

1 Conservative tuning. 22 / 1

slide-33
SLIDE 33

Large sample distribution: hard-thresholding ˆ θH,i

Theorem (known-variance, conservative case) Suppose that for given i ≥ 1 satisfying i ≤ k = k(n) for large enough n we have n1/2ηi,n → ei < ∞. Set the scaling factor αi,n =

n1/2/ξi,n. Suppose that the true parameters θ(n) = (θ1,n, . . . , θkn,n) ∈ Rk(n) and σn ∈ (0, ∞) satisfy n1/2θi,n/(σnξi,n) → νi ∈ R ∪ {−∞, ∞}.

Then F i

H,n,θ(n),σn converges weakly to the distribution with measure

{Φ(−νi + ei) − Φ(−νi − ei)}dδ−νi(x) + φ(x)1(|x + νi| > ei) dx.

[Reduces to N(0, 1) if |νi| = ∞ or ei = 0.] Analogous results for soft-thresholding and adaptive soft- thresholding.

23 / 1

slide-34
SLIDE 34

Uniform closeness of cdfs

Let F i

.,n,θ,σ be the cdf of either (centered and scaled) ˆ

θH,i, ˆ θS,i.

Let ˜

F i

.,n,θ,σ be the cdf of either (centered and scaled) ˜

θH,i, ˜ θS,i.

Uniform closeness Suppose that for given i ≥ 1 satisfying i ≤ k = k(n) for large enough n we have n1/2ηi,n(n − k)−1/2 → 0 as n → ∞. Then, as

n → ∞ sup

θ∈Rk,0<σ<∞

F i

.,n,θ,σ − ˜

F i

.,n,θ,σTV → 0

Result also holds for ad. soft-thresholding with sup-norm instead

  • f TV-norm.

Note: If n1/2ηi,n → ei < ∞ (conservative case) and n − k → ∞, then

n1/2ηi,n(n − k)−1/2 → 0 automatically holds.

24 / 1

slide-35
SLIDE 35

Large sample distribution: hard-thresholding ˜ θH,i

Theorem (unknown-variance, conservative case) Suppose that for given i ≥ 1 satisfying i ≤ k = k(n) for large enough n we have n1/2ηi,n → ei < ∞. Set the scaling factor αi,n =

n1/2/ξi,n. Suppose that the true parameters θ(n) = (θ1,n, . . . , θkn,n) ∈ Rk(n) and σn ∈ (0, ∞) satisfy n1/2θi,n/(σnξi,n) → νi ∈ R ∪ {−∞, ∞}.

Further assume that n − k is eventually constant to m. Then

˜ F i

H,n,θ(n),σn converges weakly to the distribution with measure

∞ {Φ(−νi + sei) − Φ(−νi − sei)ρm(s) ds dδ−νi(x) +φ(x) ∞ 1(|x + νi| > sei) ρm(s) ds dx.

[Reduces to N(0, 1) if |νi| = ∞ or ei = 0.] Analogous results for soft-thresholding and adaptive soft- thresholding.

25 / 1

slide-36
SLIDE 36

Large sample distributions

1 Conservative tuning: Asymptotic distributions capture be-

haviour of finite-sample distribution

in known variance case and in the unknown variance case if n − k does not diverge.

26 / 1

slide-37
SLIDE 37

Large sample distributions

2

Consistent tuning.

27 / 1

slide-38
SLIDE 38

Large sample distribution: hard-thresholding ˆ θH,i

Theorem (known-variance, consistent case) Suppose that for given i ≥ 1 satisfying i ≤ k = k(n) for large enough n we have n1/2ηi,n → ∞. Set the scaling factor

αi,n = (ηi,nξi,n)−1. Suppose that the true parameters θ(n) = (θ1,n, . . . , θkn,n) ∈ Rk(n) and σn ∈ (0, ∞) satisfy θi,n/(σnξi,nηi,n) → ζi ∈ R ∪ {−∞, ∞}. Then F i

H,n,θ(n),σn converges weakly to δ−ζi if |ζi| < 1,

and to δ0 if |ζi| > 1. If |ζi| = 1, and n1/2(ηi,n − ζiθi,n/(σnξi,n)) → ri, for some ri ∈ R then the limit is Φ(ri)δ−ζi + (1 − Φ(ri))δ0. Analogous results for soft-thresholding and adaptive soft- thresholding, except there the distributions collapse to a sin- gle pointmass in all cases.

28 / 1

slide-39
SLIDE 39

Large sample distribution: hard-thresholding ˜ θH,i

Theorem (unknown-variance, consistent case) Suppose that for given i ≥ 1 satisfying i ≤ k = k(n) for large enough n we have n1/2ηi,n → ∞. Set the scaling factor

αi,n = (ηi,nξi,n)−1. Suppose that the true parameters θ(n) = (θ1,n, . . . , θkn,n) ∈ Rk(n) and σn ∈ (0, ∞) satisfy θi,n/(σnξi,nηi,n) → ζi ∈ R ∪ {−∞, ∞}. Then ˜ F i

H,n,θ(n),σn converges weakly to

w(ζi)δ−ζi + (1 − w(ζi))δ0

(a) w(ζi) = Pr(χ2

m > mζ2 i ) if n − k is eventually constant to m ∈ N.

(b) n − k → ∞: w = 1 if |ζi| < 1 and w = 0 if |ζi| > 1. If |ζi| = 1 and n1/2(ηi,n − ζiθi,n/(σnξi,n)) → ri ∈ R ∪ {−∞, ∞}:

  • 1. n1/2ηi,n/(n − k)1/2 → 0: w = Φ(ri).
  • 2. n1/2ηi,n/(n − k)1/2 → 21/2di with 0 < di < ∞: w =

−∞ Φ(dit + ri)φ(t)dt.

  • 3. n1/2ηi,n/(n − k)1/2 → ∞ and n1/2(ηi,n − ζiθi,n/(σnξi,n)) /

(n1/2ηi,n/(n − k)1/2)) → r ′

i ∈ R ∪ {−∞, ∞}: w = Φ(r ′ i ).

29 / 1

slide-40
SLIDE 40

Large-sample distributions

Similar results for soft- and adaptive soft-thresholding, ex- cept that an absolutely continuous part ’survives’ for the case where n − k is eventually constant.

2

Consistent tuning: Asymptotic distributions always collapse at pointmasse(s)

in the known variance case and in the unknown variance case if n − k → ∞. In case of hard-thresholding, some randomness ’survives’ (con- vex combination of two pointmasses, seems to be connected to non-continuity).

(If n1/2/ξi,n-scaling is used, then certain sequences will diverge to ±∞.)

30 / 1

slide-41
SLIDE 41

Large-sample distributions

(cont’d)

Theorems reflect that

ˇ θi − θi = “bias” + “fluctuation”,

where

“bias” is O(ξi,nηi,n) (O(n−1/2) in a pointwise sense) “fluctuation” is O(n−1/2).

31 / 1

slide-42
SLIDE 42

Honest confidence sets

Revert to simpler model:

  • rthogonal design X ′X = nIn (ξi,n = 1)

known variance σ = 1 (for presentation purposes only) Wlog, consider a Gaussian location model: y1, . . . , yn

iid

∼ N(θ, 1).

(k = 1, ˆ

θLS = ¯ y)

Let ˆ θ be one the estimators ˆ θH, ˆ θL or ˆ θAL for θ. We call an interval of the form Cn = [ˆ

θ − a, ˆ θ + b] a valid or honest

confidence interval based on ˆ

θ with significance level δ, if

inf

θ∈R Pn,θ(θ ∈ Cn) ≥ δ

32 / 1

slide-43
SLIDE 43

Minimal coverage probabilities

Hard-thresholding θ vs. Pn,θ(θ ∈ Cn,H), Cn,H = [ˆ θH−an, ˆ θH+bn] (n = 1, an = 0.3, bn = 1, ηn = 0.05)

33 / 1

slide-44
SLIDE 44

Minimal coverage probabilities

Theorem (Hard-thresholding) Let Cn,H = [ˆ

θH − an, ˆ θH + bn] with an, bn ≥ 0. inf

θ∈R Pn,θ(θ ∈ Cn,H)

=    Φ(n1/2(an − ηn)) − Φ(−n1/2bn) for ηn ≤ an + bn und an ≤ bn Φ(n1/2(bn − ηn)) − Φ(−n1/2an) for ηn ≤ an + bn und an > bn for ηn > an + bn

33 / 1

slide-45
SLIDE 45

Minimal coverage probabilities

Lasso θ vs. Pn,θ(θ ∈ Cn,L), Cn,L = [ˆ θL−an, ˆ θL+bn]) (n = 1, an = 0.3, bn = 1, ηn = 0.05)

33 / 1

slide-46
SLIDE 46

Minimal coverage probabilities

Theorem (Lasso) Let Cn,L = [ˆ

θL − an, ˆ θL + bn] with an, bn ≥ 0. inf

θ∈R Pn,θ(θ ∈ Cn,L)

= Φ(n1/2(an − ηn)) − Φ(n1/2(−bn − ηn)) for an ≤ bn Φ(n1/2(bn − ηn)) − Φ(n1/2(−an − ηn)) for an > bn

33 / 1

slide-47
SLIDE 47

Minimal coverage probabilities

Adaptive Lasso θ vs. Pn,θ(θ ∈ Cn,A), Cn,A = [ˆ θAL−an, ˆ θAL+bn]) (n = 1, an = 0.3, bn = 1, ηn = 0.05)

33 / 1

slide-48
SLIDE 48

Minimal coverage probabilities

Adaptive Lasso Let Cn,AL = [ˆ

θAL − an, ˆ θL + bn] with an, bn ≥ 0. inf

θ∈R Pn,θ(θ ∈ Cn,AL) =

           Φ

  • n1/2(an−ηn)
  • − Φ
  • n1/2(an−bn)/2 −
  • ((an + bn)/2)2+η2

n)

  • for an ≤ bn

Φ

  • n1/2((an−bn)/2 +
  • ((an + bn)/2)2+η2

n)

  • − Φ
  • n1/2(−bn+ηn)
  • for an > bn

33 / 1

slide-49
SLIDE 49

The concrete confidence intervals

Let 0 < δ < 1. Hard-thresholding Among the intervals Cn,H with minimal coverage probability not less than δ, there exists a unique shortest interval C ∗

n,H with C ∗ n,H =

[ˆ θH − an,H, ˆ θH + an,H], where an,H is the unique solution of Φ(n1/2(a − ηn)) − Φ(−n1/2a) = δ.

The interval C ∗

n,H has minimal coverage probability equal to δ and

an,H is positive.

Symmetric intervals are the shortest!

34 / 1

slide-50
SLIDE 50

The concrete confidence intervals

Let 0 < δ < 1. Soft-thresholding (Lasso) Among the intervals Cn,L with minimal coverage probability not less than δ, there exists a unique shortest interval C ∗

n,L with C ∗ n,L =

[ˆ θL − an,L, ˆ θL + an,L], where an,L is the unique solution of Φ(n1/2(a − ηn)) − Φ(n1/2(−a − ηn)) = δ.

The interval C ∗

n,L has minimal coverage probability equal to δ and

an,L is positive.

Symmetric intervals are the shortest!

34 / 1

slide-51
SLIDE 51

The concrete confidence intervals

Let 0 < δ < 1. Adaptive Lasso Among the intervals Cn,AL with minimal coverage probability not less than δ, there exists a unique shortest interval C ∗

n,AL with

C ∗

n,AL = [ˆ

θAL − an,AL, ˆ θAL + an,AL], where an,AL is the unique solution of Φ(n1/2(a − ηn)) − Φ(−n1/2 a2 + η2

n) = δ

The interval C ∗

n,AL has minimal coverage probability equal to δ and

aAL is positive.

Symmetric intervals are the shortest!

34 / 1

slide-52
SLIDE 52

Lengths of confidence sets – in finite-samples

For a fixed δ with 0 < δ < 1 and every n ∈ N we have

an,H > an,AL > an,L > an,LS.

35 / 1

slide-53
SLIDE 53

Lengths of confidence sets – asymptotically

1 Conservative case.

an,H ∼ an,AL ∼ an,L ∼ an,LS ∼ n−1/2

All quantities are of the same order n−1/2.

2 Consistent case. an,{L,H,AL} = ηn + n−1/2Φ(δ) + o(n−1/2)

an,H/an,MLE ∼ an,AL/an,MLE ∼ an,L/an,LS ∼ n1/2ηn → ∞

Intervals lengths for PLSEs are larger by an order of magnitude compared to the one based the ’unpenalized’ LS estimator!

36 / 1

slide-54
SLIDE 54

Lengths of confidence sets – illustration

Plot: n1/2an vs n1/2ηn for δ = 0.95.

37 / 1

slide-55
SLIDE 55

Impossibility Results for Estimation of the cdf

Theorem Let ηn → 0 n1/2ηn → m with 0 < e ≤ ∞. Then every estimator ˆ Fn(t) of Fn,θ(t) satisfies

sup

|θ|<c/n1/2 Pn,θ

  • ˆ

Fn(t) − Fn,θ(t)

  • > ε

1 2

for each ε < (Φ(t + n1/2ηn) − Φ(t − n1/2ηn))/2, for each c > |t|, and for each sample size n. Hence

lim inf

n→∞ inf ˆ Fn(t)

sup

|θ|<c/n1/2 Pn,θ

  • ˆ

Fn(t) − Fn,θ(t)

  • > ε

1 2

for each ε < (Φ(t + e) − Φ(t − e))/2, for each c > |t|. In particular, no uniformly consistent estimator for Fn,θ(t) exists.

38 / 1

slide-56
SLIDE 56

Summary

We studied distributional properties of thresholding (PLS) estimators for known and unknown variance in a linear regression setting with a (potentially) growing number of parameters. Fixed-parameter asymptotics paint a misleading picture of the performance of the estimators. Finite- and large-sample distributions are highly non-normal. In case of consistent tuning, the uniform rate of convergence is slower than n−1/2. In the unknown variance case, large-sample behavior depends

  • n whether and how fast n − k diverges in relation to the

tuning parameter. If n − k diverges, distributions collapse at point-mass for con- sistent tuning.

39 / 1

slide-57
SLIDE 57

Summary

(cont’d)

Orthogonal design, fixed dimension: Confidence sets are larger by an order of magnitude compared to the ones based on the LS-estimator in the consistent case. Lengths are of the same order for conservative tuning. Not a criticism on the estimators per se. Distributional properties have to be investigated taking into account non-uniformity issues.

40 / 1

slide-58
SLIDE 58

References

  • A. Belloni and V. Chernozhukov. Post l1-penalized estimators in high-dimensional linear regression models.

manuscript arxiv:1001.0188, 2010.

  • J. Fan and R. Li. Variable selection via nonconcave penalized likelihood and its oracle properties. J. Am.
  • Stat. Ass., 96:1348–1360, 2001.
  • I. E. Frank and J. H. Friedman. A statistical view of some chemometrics regression tools (with discussion).

Technom., 35:109–148, 1993.

  • K. Knight and W. Fu. Asymptotics of lasso-type estimators. Ann. Stat., 28:1356–1378, 2000.
  • B. M. P¨
  • tscher and H. Leeb. On the distribution of penalized maximum likelihood estimators: The LASSO,

SCAD, and thresholding. J. Multivariate Anal., 100:2065–2082,2009.

  • B. M. P¨
  • tscher and U. Schneider. On the distribution of the adaptive lasso estimator. J. Stat. Plan. Inf.,

139:2775–2790, 2009.

  • B. M. P¨
  • tscher and U. Schneider. Confidence sets based on penalized maximum likelihood estimators in

Gaussian regression. J. Electron. Stat., 4:334–360, 2010.

  • B. M. P¨
  • tscher and U. Schneider. Distributional results for thresholding estimators in high-dimensional

Gaussian regression models. manuscript arxiv:1106.6002, 2011.

  • S. van de Geer and P. B¨

uhlmann and S. Zhou. The adaptive and the thresholded Lasso for potentially misspecified models. manuscript arxiv:1001.5176, 2010.

  • H. Zou. The adaptive lasso and its oracle properties. J. Am. Stat. Ass., 101:1418–1429, 2006.
  • H. Zhang. Nearly unbiased variable selection under minimax concave penalty Ann. Stat., 38:894–942, 2010.
  • H. Zou and T. Hastie. Regularization and variable selection via the elastic net. J. Roy. Stat. Soc. B,

67:301–320, 2005. 41 / 1