using R for regression model selection with adaptive penalties - - PowerPoint PPT Presentation

using r for regression model selection with adaptive
SMART_READER_LITE
LIVE PREVIEW

using R for regression model selection with adaptive penalties - - PowerPoint PPT Presentation

using R for regression model selection with adaptive penalties procedures based on the FDR criteria Tal Galili Tel Aviv University Based on the paper by YOAV BENJAMINI and YULIA GAVRILOV A SIMPLE FORWARD SELECTION PROCEDURE BASED ON


slide-1
SLIDE 1

using R for regression model selection with adaptive penalties procedures based on the FDR criteria”

Tal Galili Tel Aviv University Based on the paper by YOAV BENJAMINI and YULIA GAVRILOV

“A SIMPLE FORWARD SELECTION PROCEDURE BASED ONFALSE DISCOVERY RATE CONTROL” (Annals of Applied Statistics 2009)

slide-2
SLIDE 2

Task: Stopping rule (finding the “Best model”

  • n the Forward selection path)

(Some Zeros)

slide-3
SLIDE 3

Why forward selection ? Motivation – Big (m) datasets:

1) Fast results

  • Simple models
  • Simple procedure

2) Good results 3) Easy to use

slide-4
SLIDE 4

Finding variables Over fitting

Penalty Model Size

Minimize

slide-5
SLIDE 5

type examples for “big” models

constant

  • a= 2 (AIC)
  • Over fitting

Non-constant (adaptive)

  • n = log(n) (BIC)
  • m = 2log(m) (universal-threshold)
  • k,m = ?
  • Better results
  • Faster then

bootstraping.

Minimize

How to choose ?

slide-6
SLIDE 6
  • 2

, 1 2 1 1

1

k k m q i i m i q

z k

  • Multiple Step FDR

(MSFDR) Adaptive Penalty

slide-7
SLIDE 7

Model selection multiple testing

  • 0,1

0, 0,

? ? ? 1 ,..., ,...,

0,..., 0,...,

i m

i m H H H

  • Orthogonal X matrix => non changing, coefficients “at once”:

Keeping (Beta) P-values which are bellow forward selection But how should we adjust for multiplicity of the many tests?

  • 1

1

ˆ ˆ ˆ ,...,

m

X X nI n X y

slide-8
SLIDE 8

How to adjust for multiplicity?

q

keeping Approach Principle properties

FWE (familywise error rate) Keeping the probability of making one or more false discoveries.

  • Conservative
  • Low-power

FDR (False discovery rate) Controlling the expected proportion of incorrectly rejected null out of the rejected

  • Not “too

permissive”

  • high-power
slide-9
SLIDE 9

Coefficient P value

(0.16=AIC)

Bonferroni (FWE) BH – (FDR at ) Adaptive - Step down Over fitting Low power More power (Largest) (Smallest) . . . . . . (Smallest) (Largest)

  • 2

1 1

ˆ ˆ SE

  • 2

~

df

t

  • 2

2 2

ˆ ˆ SE

  • 2

ˆ ˆ

m m

SE

  • 1

P

2

P

  • m

P

0.05

m

  • m
  • m
  • m
  • 1

qm 2 qm i qm

i

P

m qm

  • 2

ˆ ˆ

i i

SE

  • m

q m

slide-10
SLIDE 10

Coefficient P value

(0.16=AIC)

Bonferroni (FWE) BH – (FDR at ) Adaptive - Step down Over fitting Low power More power

  • 2

~

df

t

0.05

1 qm 2 qm i qm m qm

  • m

q m

  • 2

, 1 2

1 k

k m q i i m

z k

slide-11
SLIDE 11

Theoretical motivation– results

The minimax properties of the BH procedure were proved (in ABDJ 2006*) asymptotically for:

  • large m,

<and >

  • orthogonal variables,

<and >

  • for sparse signals.

*ABRAMOVICH, F., BENJAMINI, Y., DONOHO, D. and JOHNSTONE, I. (2006). Adapting to unknown sparsity by controlling the false discovery rate. Ann. Statist.

slide-12
SLIDE 12

Coefficient P value

(0.16=AIC)

Bonferroni (FWE) BH – (FDR at ) Adaptive BH – (FDR at level q) Adaptive - Step down Over fitting Low power More power More power for richer models (Largest) (Smallest) . . . . . . (Smallest) (Largest)

  • 2

1 1

ˆ ˆ SE

  • 2

~

df

t

  • 2

2 2

ˆ ˆ SE

  • 2

ˆ ˆ

m m

SE

  • 1

P

2

P

  • m

P

0.05

1 qm 2 qm i qm

i

P

m qm

  • 1

1 1 1 qm q

  • 2

1 2 1 qm q

  • 1

1 i qm i q

  • 1

1 m qm m q

  • 2

ˆ ˆ

i i

SE

  • m

q m

slide-13
SLIDE 13

BH – (FDR at ) Adaptive BH – (FDR at level q) Adaptive - Step down More power More power for richer models

m q m

  • i

qm

  • 1

1 i qm i q

  • 2

, 1 2 1 1

1 k

k m q i i m i q

z k

  • 2

, 1 2

1 k

k m q i i m

z k

  • model size VS

Penalty factor

slide-14
SLIDE 14

Forward-selection - Multiple stage FDR: (a.k.a: MSFDR)

1. Fit Empty model 2. Find the “best” variables (Xi*) to enter (with the smallest P value) 3. Is this true ? 1. Yes - Enter Xi and repeat (step 2) 2. No – Finish.

slide-15
SLIDE 15

R implementation - stepAIC

slide-16
SLIDE 16

Modeling the diabetes data (Efron et al., 2004)

  • n=442 diabetes patients.
  • m= 64 (10 baseline variables

with 45 paired and 9 squared interactions ).

  • Y - disease progression (a year after baseline)
slide-17
SLIDE 17

Factor P-value P-to-enter

bmi 230.740.000000 0.000781 11.29 0.342 ltg 93.86 0.000000 0.001585 10.63 0.457 map 17.36 0.000037 0.002414 10.16 0.477 age.sex 13.56 0.000259 0.003268 9.78 0.491 bmi.map 9.60 0.002076 0.004149 9.47 0.501 hdl 9.00 0.002859 0.005059 9.20 0.510 sex 16.23 0.000066 0.005998 8.96 0.527 glu.2 5.75 0.016920 0.006969 8.75 0.531 age.2 2.58 0.109060 0.007972 8.56 0.533

Modeling the diabetes data (Efron et al., 2004)

2

df

t

, k m

  • (

)

2 adj

R

slide-18
SLIDE 18

Modeling the diabetes data (Efron et al., 2004)

Method Number of variables R^2 MS_FDR (q=.05),

BIC, universal-threshold

7 0.53 AIC 9 0.54 LARS (with Cp) 16 0.55

Over fitting

slide-19
SLIDE 19

Simulation - configurations

  • penalty based model selection procedures
  • m = 20, 40, 80, 160,

Ratio: n = 2*m

  • proportion of non-zero
  • Dependencies in X:

= 0.5, 0,−0.5

  • = 1 constant (with ), 2 rates of

decrease (in one minimal is constant)

  • Computation – avg MSPE over 1000 runs
  • done on 80 computers (distributed computing)

i i i

Y X

  • 3

= m , , , , , 4 3 2 4 m m m m m

  • |

|

0, ;

i j m m m m

N

  • 2

0.75 R

slide-20
SLIDE 20

Simulation – Comparison methodology

1) Compute the ratio: (For each model)

Random Oracle = the “best” model we could find on our search path

2) For each procedure Over all simulation configuration find the worst ratio – and compare them

m odel random oracle

M SP E M SP E

slide-21
SLIDE 21

Simulation – results

  • forward selection procedure
  • Cp
  • the universal threshold in Donoho and Johnstone (1994)
  • Birgé and Massart (2001)
  • Foster and Stine (2004)
  • Tibshirani and Knight (1999)
  • multiple-stage procedure in Benjamini, Krieger and Yekutieli (2006) and Gavrilov, Benjamini and Sarkar (2009)—MSFDR

Comparing the minimax between procedure

slide-22
SLIDE 22

R implementation – biglm + leaps

slide-23
SLIDE 23

Future research

  • Beyond Linear regression? (logistic and

more)

  • Beyond forward selection? (Mixed with

Lasso and more)

  • More variables then observation? (m>n)
slide-24
SLIDE 24

www.R-Statistics.com Tal.Galili@gmail.com

Thank you! Questions?

Tel Aviv University Based on the paper by YOAV BENJAMINI and YULIA GAVRILOV

“A SIMPLE FORWARD SELECTION PROCEDURE BASED ONFALSE DISCOVERY RATE CONTROL” (Annals of Applied Statistics 2009)

slide-25
SLIDE 25

Simulation – Comparison methodology

Challenge (1): Path performance depends on simulation

(while exhaustive search over all subsets – impossible!)

What do we compare to ? Solution (1): a “random oracle” 1)Find the “best” model on the forward path of nested models Example: for the path: X7, X20, X5, X9 … The possible subsets are: {X7},{X7, X20}, {X7, X20, X5} … 2)Compare current models with random oracle

m odel random oracle

M SP E M SP E

slide-26
SLIDE 26

Simulation – Comparison methodology

Challenge (2): MSPE changes per configuration, so how do we compare algorithms? Solution (2): search for “empirical minimax performance” – find the minimum across “maximum relative MSPE

  • ver the configurations”
slide-27
SLIDE 27

Simulation – conclusions

slide-28
SLIDE 28

Simulation – results (extended)

slide-29
SLIDE 29

Earlier studies limitations: 1)Constant coefficients

(mostly)

2)Largest m = 50 3)NOT Compared to other non-constant adaptive penalties

slide-30
SLIDE 30

For orthogonal X matrix:

  • 1

1

ˆ ˆ ˆ ,...,

m

X X nI n X y

  • Forward selection is like sorting the P-values and then keeping only

who ever is smaller then :

1 2 2 2 2 1

1 2 2 2 2 1 2 2 1 1

k i

k k k k i i z z

RSS z RSS z

  • So on which P should we stop ? ?

0.05

  • 2

2 2 1 1 2 1 1 2 2 2

ˆ ˆ ˆ ˆ

n k k k ki i k k k k k

RSS RSS x RSS RSS n SE

  • The difference of each step is of the standardized coefficient, since:
slide-31
SLIDE 31

BH – (FDR at ) Adaptive BH – (FDR at level q) Adaptive - Step down More power More power for richer models

m q m

  • i

qm

  • 1

1 i qm i q

  • 1

2 2 2 2 1

1 2 2 2 2 1 2 2 1 1

k i

k k k k i i z z

RSS z RSS z

  • An adaptive

penalty procedure

2 2 2 2 2 2 1 1

1

k k k k i i

RSS z RSS k z k

  • With this
slide-32
SLIDE 32

What this is NOT:

1. Fit full model – then check the P values 2. Fit m “small” models – then check the P values

slide-33
SLIDE 33
  • :

1 : FWE P V V FDR E q R

  • !
  • !
  • Correction for multiple testing

Very conservative, Low power Different objective, More power