Subsampling versus bootstrap in resampling-based model selection for - - PowerPoint PPT Presentation

subsampling versus bootstrap in resampling based model
SMART_READER_LITE
LIVE PREVIEW

Subsampling versus bootstrap in resampling-based model selection for - - PowerPoint PPT Presentation

Subsampling versus bootstrap in resampling-based model selection for multivariable regression Subsampling versus bootstrap in resampling-based model selection for multivariable regression Riccardo De Bin 1 , Silke Janitza 1 , Willi Sauerbrei 2


slide-1
SLIDE 1

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Riccardo De Bin1, Silke Janitza1, Willi Sauerbrei2 & Anne-Laure Boulesteix1 G¨ unzburg, July 23rd 2014

1Department of Medical Informatics, Biometry and Epidemiology, University of

Munich, Germany

2Department of Medical Biometry and Medical Informatics, University Medical

Center Freiburg, Germany

Statistical Computing 2014, July 23rd 2014 1/ 23

slide-2
SLIDE 2

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Outline

Introduction Methods Data Results Prediction accuracy Conclusions

Statistical Computing 2014, July 23rd 2014 2/ 23

slide-3
SLIDE 3

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Introduction

  • model selection to identify the “best” model to describe an outcome;
  • if the study is replicated, the procedure should ideally produce the

same result → stability;

  • model selection for multivariable regression based on inclusion

frequencies (Gong, 1982; Sauerbrei & Schumacher, 1992);

  • this approach is based on a resampling technique:

◮ the classical (mostly used) choice is the bootstrap; ◮ the bootstrap has some pitfalls (Janitza et al., 2014); ◮ alternatives such as subsampling should be considered.

  • Aim: compare bootstrap and subsampling in a model selection

process for multivariable regression based on inclusion frequencies.

Statistical Computing 2014, July 23rd 2014 3/ 23

slide-4
SLIDE 4

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Methods: Inclusion frequencies

  • we generate, through a resampling technique, several

pseudo-samples containing small perturbation of the original data.

  • in each pseudo-sample, we apply a model selection procedure;
  • we define the proportion of times in which a variable is selected in

the models as “inclusion frequency” (IF);

  • ideally, we can distinguish between:

◮ relevant variables, related to the outcome → high IF; ◮ noise variables, significant only in specific samples → low IF;

  • possible issues:

◮ variables with weak effect (their IF may depend on chance); ◮ co-selection (e.g., two highly correlated variables may be

alternatively selected, leading to an IF around 0.5 for both of them).

Statistical Computing 2014, July 23rd 2014 4/ 23

slide-5
SLIDE 5

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Methods: Model selection

  • we would like to select a model which:

◮ contains all the relevant variables, to correctly explain the outcome

and to avoid underfitting;

◮ contains as few variables as possible, to favor interpretability and to

avoid overfitting;

  • several approaches are available in literature:

◮ backward elimination, forward selection, all subset approach, . . . ◮ here we use backward elimination with no re-inclusion (for arguments

in favor of this choice, see Mantel, 1970);

  • the inclusion criterion is a key aspect:

◮ significance level, information criterion, total number of variables, . . . ◮ we base our analysis on the significance level (here 0.05, 0.10, 0.157); Statistical Computing 2014, July 23rd 2014 5/ 23

slide-6
SLIDE 6

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Methods: Resampling strategies (1/2)

In our study, we consider the following resampling strategies:

  • bootstrap(n)

◮ it is the classical bootstrap technique (Efron, 1979) ◮ n observations drawn from the original data with replacement;

(hereafter n denotes the sample size);

◮ its asymptotic properties have been extensively studied in the last

decades, starting from Bickel & Freedman (1981);

◮ there are counterexamples where the consistency is not achieved

(see, e.g., Mammen, 1992; Bickel et al., 1997);

◮ bootstrap(n) shows pitfalls in several cases (for a recent review, see

Janitza et al., 2014);

Statistical Computing 2014, July 23rd 2014 6/ 23

slide-7
SLIDE 7

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Methods: Resampling strategies (2/2)

  • subsample(m);

◮ intensively investigated in literature (Shao & Wu, 1989; Politis &

Romano, 1994; Politis et al., 1999);

◮ m < n observations drawn from the original data without

replacement;

◮ also known as delete-d jackknife (see Wu, 1986); ◮ shows asymptotic consistency also in cases where the classical

bootstrap fails (Davison et al., 2003);

  • bootstrap(m);

◮ m < n observations drawn from the original data with replacement; ◮ already considered in Bickel & Freedman (1981);

  • here m = 0.632n, the average number of unique observations in a

bootstrap(n) sample;

  • comparability due to same pseudo-sample size.

.

Statistical Computing 2014, July 23rd 2014 7/ 23

slide-8
SLIDE 8

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Methods: Resampling censored data

  • one of our examples (Glioma data) deals with survival data;
  • the presence of censored observations may raise some complications;
  • we directly apply the resampling technique, also if it produces

pseudo-samples with different effective sizes (number of events);

  • alternatives are available (e.g., resample separately events and

censored observations);

  • arguments in favor of the direct approach can be found in Burr

(1994) and Zelterman et al. (1996).

Statistical Computing 2014, July 23rd 2014 8/ 23

slide-9
SLIDE 9

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Data: Glioma dataset

  • original study: Ulm et al. (1989);
  • publicly available at http:

//portal.uni-freiburg.de/imbi/Royston-Sauerbrei-book;

  • time-to-event data: survival time of patients with malignant glioma;
  • 411 patients, 274 events (median follow up: 712 days);
  • 15 variables available: 1 continuous, 8 binary, 6 dummy variables

representing 3 originally categorical variables;

  • the proportional hazards assumption is acceptable → Cox model.

Statistical Computing 2014, July 23rd 2014 9/ 23

slide-10
SLIDE 10

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Data: Ozona dataset

  • original study: Ihorst et al. (2004);
  • we use the subset defined by Buchholz et al. (2008);
  • information about ozone effect on 496 school children’s lung growth;
  • 24 variables available: 7 continuous and 17 binary;
  • classical multivariable linear regression model;

Statistical Computing 2014, July 23rd 2014 10/ 23

slide-11
SLIDE 11

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Results

  • the results are based on 10000 iterations of the following procedure:

◮ we draw a pseudo-sample from the original data; ◮ we select a model applying a backward elimination procedure with

inclusion criterion α = 0.05;

  • we consider:

◮ inclusion frequencies of the variables; ◮ average number of variables included in the models; ◮ number of unique models selected; ◮ structure of the models;

  • for each of the three resampling techniques.

Statistical Computing 2014, July 23rd 2014 11/ 23

slide-12
SLIDE 12

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Glioma data: Variables’ inclusion frequencies

bootstrap(n) bootstrap(m) subsample(m) frequency of inclusion 0.0 0.2 0.4 0.6 0.8 1.0 sex time gradd1 gradd2 age kard1 kard2 surgd1 surgd2 convul cort epi amnesia

  • ps

aph

Statistical Computing 2014, July 23rd 2014 12/ 23

slide-13
SLIDE 13

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Glioma data: Number of unique models and average number of variables

resampling method average number

  • f variables

number of unique models bootstrap(n) 6.864 1787 bootstrap(m) 5.856 1829 subsample(m) 5.057 580

Statistical Computing 2014, July 23rd 2014 13/ 23

slide-14
SLIDE 14

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Glioma data: Models’ selection frequencies

bootstrap(n) bootstrap(m) subsample(m) model rank freq. rank freq. rank freq. basic+kard1 2 124 1 326 1 1615 basic+kard1+epi 8 93 7 128 2 417 basic 8 123 3 398 basic+kard1+surgd2 6 103 3 163 4 352 basic+kard1+cort 5 106 4 148 5 298 basic+kard1+sex 3 108 2 187 6 290 basic+cort+ops 4 148 7 264 basic+epi 8 242 basic+kard1+sex+epi 1 156 6 140 9 225 basic* 10 117 10 205 basic+ops 9 121 basic*+kard1+cort+ops 3 108 basic+cort+ops 7 97 basic*+kard1+cort 8 93 basic+kard1+surgd2+sex+epi 10 89 basic=intercept+gradd1+age+surgd1; basic*=intercept+gradd2+age+surgd1

Statistical Computing 2014, July 23rd 2014 14/ 23

slide-15
SLIDE 15

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Glioma data: Models’ structures

Variable bootstrap(n) bootstrap(m) subsample(m) basic 15 123 398 basic + 1 additional 247 878 2432 basic + 2 additional 1030 1923 2786 basic + 3 additional 2071 2123 1956 basic + 4 additional 2505 1451 653 basic + 5 additional 1742 676 155 basic + > 5 additional 1103 275 27

  • thers

1287 2551 1593 Variable bootstrap(n) bootstrap(m) subsample(m) basic* 17 178 473 basic* + 1 additional 304 1213 2841 basic* + 2 additional 1272 2590 3441 basic* + 3 additional 2472 2772 2309 basic* + 4 additional 2832 1825 730 basic* + 5 additional 1904 803 163 basic* + > 5 additional 1180 321 27 Without at least 1 core* 19 298 16

Statistical Computing 2014, July 23rd 2014 15/ 23

slide-16
SLIDE 16

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Ozone data: Variables’ inclusion frequencies

bootstrap(n) bootstrap(m) subsample(m) frequency of inclusion 0.0 0.2 0.4 0.6 0.8 1.0 alter adheu sex hochozon amatop avatop adekz arauch agebgew fsnight flgross fmilb fnoh24 ftier fpoll fltotmed fo3h24 fspt fteh24 fsatem fsauge flgew fspfei fshlauf Statistical Computing 2014, July 23rd 2014 16/ 23

slide-17
SLIDE 17

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Ozone data: Number of unique models and average number of variables

resampling method average number

  • f variables

number of unique models bootstrap(n) 9.148 5254 bootstrap(m) 7.663 4768 subsample(m) 6.170 1030

Statistical Computing 2014, July 23rd 2014 17/ 23

slide-18
SLIDE 18

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Ozone data: Models’ selection frequencies

bootstrap(n) bootstrap(m) subsample(m) model rank freq. rank freq. rank freq.

basic+fspfei+fpoll

1 80 1 416

basic+fsatem

2 73 2 371

basic+fspfei

4 68 3 340

basic+fsatem+fmilb

6 56 4 318

basic+fsatem+fpoll

3 69 5 312

basic+fspfei+fmilb+hochozon+fnoh24

1 72 5 63 6 295

basic+fspfei+fpoll+hochozon+fnoh24

3 59 10 48 7 269

basic+fspfei+fmilb

8 233

basic+fsatem+fmilb+hochozon+fnoh24

4 56 9 221

basic

7 51 10 206

basic+fsatem+hochozon+fnoh24

7 51

basic+fspfei+fmilb+hochozon+fnoh24+fo3h24+fteh24

2 60 9 49

basic+fspfei+fpoll+hochozon+fnoh24+fo3h24+fteh24+fltotmed

5 54

basic+fspfei+fpoll+fsatem+hochozon+fnoh24+fltotmed

6 46

basic+fspfei+hochozon+fnoh24+fo3h24+fteh24

7 42

basic+fspfei+fmilb+fsatem+hochozon+fnoh24+fltotmed

8 38

basic+fspfei+fpoll+fsatem+hochozon+fnoh24+fo3h24+fteh24+fltotmed

8 38

basic+fspfei+fpoll+hochozon+fnoh24+fltotmed

10 37

basic=intercept+sex+flgross+flgew Statistical Computing 2014, July 23rd 2014 18/ 23

slide-19
SLIDE 19

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Ozone data: Models’ structures

bootstrap(n) bootstrap(m) subsample(m) basic 5 51 206 basic + 1 additional 31 309 953 basic + 2 additional 217 921 2639 basic + 3 additional 643 1666 2333 basic + 4 additional 1333 1989 2167 basic + 5 additional 1648 1830 796 basic + > 5 additional 6123 3189 906 Without at least 1 core 45

Statistical Computing 2014, July 23rd 2014 19/ 23

slide-20
SLIDE 20

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Prediction accuracy: definition

  • dealing with real data, we do not know the true model;
  • we investigate the prediction ability of the selected models through a

10-fold cross-validation procedure:

◮ we apply backward selection on pseudo-samples generated from the

  • bservations belonging to 9 folds;

◮ we evaluate the results on the remaining fold; ◮ we iterate for all the 10 combinations and average the results; ◮ we repeat the cross-validation 10000 times;

  • the prediction ability is measured with a quadratic score:

◮ integrated Brier score (IBS) (Graf et al., 1999); ◮ residuals sum of squares (RSS). Statistical Computing 2014, July 23rd 2014 20/ 23

slide-21
SLIDE 21

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Prediction accuracy: results

resampling method Glioma data Ozone dataset (IBS) (RSS) bootstrap(n) 0.157 2.408 bootstrap(m) 0.160 2.465 subsample(m) 0.156 2.362

Statistical Computing 2014, July 23rd 2014 21/ 23

slide-22
SLIDE 22

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Conclusions

  • we compared subsampling and bootstrap in a model building

procedure for multivariable regression based on inclusion frequencies;

  • the results confirm the overcomplexity issues related to the use of

the bootstrap approach (Janitza et al., 2014);

  • there seem to be some arguments in favor of subsampling:

◮ it leads to a strong consensus for one or very few models; ◮ the relevant and noise variables seem to be well separated;

  • we need a simulation study (work in progress) to confirm the

impressions derived from these two case studies.

Statistical Computing 2014, July 23rd 2014 22/ 23

slide-23
SLIDE 23

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

References

Bickel, P. J. & Freedman, D. A. (1981). Some asymptotic theory for the bootstrap. The Annals of Statistics 9, 1196–1217. Bickel, P. J., G¨

  • tze, F. & van Zwet, W. R. (1997). Resampling fewer than n observations: gains, losses, and remedies for losses.

Statistica Sinica 7, 1–31. Buchholz, A., Holl¨ ander, N. & Sauerbrei, W. (2008). On properties of predictors derived with a two-step bootstrap model averaging approach: a simulation study in the linear regression model. Computational Statistics & Data Analysis 52, 2778–2793. Burr, D. (1994). A comparison of certain bootstrap confidence intervals in the cox model. Journal of the American Statistical Association 89, 1290–1302. Davison, A. C., Hinkley, D. V. & Young, G. A. (2003). Recent developments in bootstrap methodology. Statistical Science 18, 141–157. Efron, B. (1979). Bootstrap methods: another look at the jackknife. The Annals of Statistics 7, 1–26. Gong, G. (1982). Some ideas on using the bootstrap in assessing model variability. In Computer Science and Statistics: Proceedings of the 14th Symposium on the Interface. Springer New York. Graf, E., Schmoor, C., Sauerbrei, W. & Schumacher, M. (1999). Assessment and comparison of prognostic classification schemes for survival data. Statistics in Medicine 18, 2529–2545. Ihorst, G., Frischer, T., Horak, F., Schumacher, M., Kopp, M., Forster, J., Mattes, J. & Kuehr, J. (2004). Long-and medium-term ozone effects on lung growth including a broad spectrum of exposure. European Respiratory Journal 23, 292–299. Janitza, S., Binder, H. & Boulesteix, A.-L. (2014). Pitfalls of hypothesis tests and model selection on bootstrap samples: causes and consequences in biometrical applications. Tech. Rep. 163, Department of Statistics, University of Munich. Mammen, E. (1992). When Does Bootstrap Work? Springer. Mantel, N. (1970). Why stepdown procedures in variable selection. Technometrics 12, 621–625. Politis, D., Romano, J. & Wolf, M. (1999). Subsampling. Springer, New York. Politis, D. N. & Romano, J. P. (1994). Large sample confidence regions based on subsamples under minimal assumptions. The Annals

  • f Statistics 22, 2031–2050.

Sauerbrei, W. & Schumacher, M. (1992). A bootstrap resampling procedure for model building: application to the Cox regression

  • model. Statistics in Medicine 11, 2093–2109.

Shao, J. & Wu, C. J. (1989). A general theory for jackknife variance estimation. The Annals of Statistics 17, 1176–1197. Ulm, K., Schmoor, C., Sauerbrei, W., Kemmler, G., Aydemir, ¨ U., M¨ uller, B. & Schumacher, M. (1989). Strategien zur Auswertung einer Therapiestudie mit der ¨ Uberlebenszeit als Zielkriterium. Biometrie und Informatik in Medizin und Biologie 20, 171–205. Wu, C.-F. J. (1986). Jackknife, bootstrap and other resampling methods in regression analysis. The Annals of Statistics 14, 1261–1295. Zelterman, D., Le, C. T. & Louis, T. A. (1996). Bootstrap techniques for proportional hazards models with censored observations. Statistics and Computing 6, 191–199. Statistical Computing 2014, July 23rd 2014 23/ 23

slide-24
SLIDE 24

Subsampling versus bootstrap in resampling-based model selection for multivariable regression Statistical Computing 2014, July 23rd 2014 24/ 23

slide-25
SLIDE 25

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Glioma data: Cox model

variable estimate std error p-value sex

  • 0.175

0.129 0.17460 time

  • 0.128

0.140 0.36274 gradd1 0.798 0.251 0.00151 gradd2 0.257 0.190 0.17585 age 0.038 0.007 9 × 10−7 kard1

  • 0.317

0.139 0.02305 kard2

  • 0.039

0.172 0.82208 surgd1

  • 1.046

0.213 9 × 10−9 surgd2

  • 0.216

0.139 0.11962 convul 0.095 0.138 0.49361 cort 0.264 0.139 0.05755 epi

  • 0.270

0.150 0.07148 amnesia 0.097 0.198 0.62390

  • ps

0.253 0.164 0.12328 aph

  • 0.119

0.137 0.27478

Statistical Computing 2014, July 23rd 2014 25/ 23

slide-26
SLIDE 26

Subsampling versus bootstrap in resampling-based model selection for multivariable regression

Ozone data: linear model

variable estimate std error p-value intercept

  • 1.721

0.264 2 × 10−10 alter 0.025 0.017 0.15708 adheu

  • 0.038

0.043 0.37135 sex

  • 0.197

0.020 1 × 10−16 hochozon

  • 0.069

0.027 0.01202 amatop

  • 0.003

0.023 0.87883 avatop

  • 0.017

0.024 0.48672 adekz 0.009 0.025 0.70635 arauch 0.007 0.022 0.75821 agebgew 2 × 10−5 2 × 10−5 0.33302 fsnight 0.026 0035. 0.44492 flgross 0.026 0.002 1 × 10−16 fmilb

  • 0.057

0.037 0.12073 fnoh24

  • 0.002

0.001 0.00468 ftier

  • 0.013

0.037 0.71378 fpoll

  • 0.060

0.045 0.18902 fltotmed

  • 0.054

0.028 0.05463 fo3h24 0.001 0.001 0.11463 fspt 0.032 0.049 0.51448 fteh24

  • 0.005

0.003 0.12744 fsatem 0.102 0.054 0.06102 fsauge 0.010 0.032 0.76082 flgew 0.012 0.002 3 × 10−9 fspfei 0.122 0.055 0.02825 fshlauf

  • 0.032

0.043 0.45219

Statistical Computing 2014, July 23rd 2014 26/ 23