Examining the Experimental Designs and Statistical Power of Group - - PowerPoint PPT Presentation

examining the experimental designs and statistical power
SMART_READER_LITE
LIVE PREVIEW

Examining the Experimental Designs and Statistical Power of Group - - PowerPoint PPT Presentation

Examining the Experimental Designs and Statistical Power of Group Randomized Trials Funded by the Institute of Education Sciences Jessaca K. Spybrook A Presentation for the Evaluation Caf at Western Michigan University February 19, 2008


slide-1
SLIDE 1

Examining the Experimental Designs and Statistical Power of Group Randomized Trials Funded by the Institute of Education Sciences

Jessaca K. Spybrook A Presentation for the Evaluation Café at Western Michigan University February 19, 2008

slide-2
SLIDE 2

Slide 2

Background

Evidence-based education Randomized trials Group randomized trials / Cluster randomized

trials

slide-3
SLIDE 3

Slide 3

Background

Institute of Education Sciences (IES)

National Center for Education Research (NCER) National Center for Education Evaluation and

Regional Assistance (NCEE)

Produce research that provides reliable

evidence on which to base education policy and practice

slide-4
SLIDE 4

Slide 4

Background

NCER

Goal 3 Projects – Efficacy and Replication

Test effectiveness of intervention under specific

conditions

~ $250,000 - $700,000 per year

Goal 4 Projects – Effectiveness Evaluations

Test effectiveness of intervention under more typical

conditions

Up to $1.2 million per year

slide-5
SLIDE 5

Slide 5

Background

NCEE

Conduct rigorous evaluations of federal programs Contracts not grants At least $1 million per year

slide-6
SLIDE 6

Slide 6

Background

Group randomized trial Reliable, scientific

evidence

Strong design Large enough sample size to conclusively

determine whether or not an intervention can improve student outcomes by a specified margin (adequate power)

Power of 0.80 is usually considered acceptable in

social sciences

slide-7
SLIDE 7

Slide 7

Background - Terms

Minimum detectable effect size (MDES) –

Smallest effect size that can be detected with power = 0.80

Sample size at all levels Intra-class correlation Covariate-outcome correlation Presence and strength of blocking variable

slide-8
SLIDE 8

Slide 8

Central Goal of this Study

Examine the designs and power analyses for

the group randomized trials funded by the National Center for Education Research (NCER) and the National Center for Education Evaluation and Regional Assistance (NCEE)

slide-9
SLIDE 9

Slide 9

Key Questions

1.

What designs do these studies use?

slide-10
SLIDE 10

Slide 10

Key Questions

2.

Under plausible assumptions about intra- class correlations, covariate-outcome correlations, and explanatory effects of blocking, what are the minimum detectable effect sizes’s (MDES) of the studies in the sample?

slide-11
SLIDE 11

Slide 11

Key Questions

3.

What is the relationship between the MDES stated in the proposal and the MDES under plausible assumptions regarding the design parameters? To the extent that there are discrepancies between the two values, what are the possible sources of the inconsistencies?

  • Is there a power analysis? Is it documented? Does it correspond to the study

description?

  • Are the intra-class correlations documented? If so, what are the estimated

values?

  • Are covariates included in the power analysis? If so, are the covariate-
  • utcome correlations documented? If so, what are the values?
  • Is blocking included in the description of the study? If so, is blocking

included in the power analysis and are the explanatory effects of blocking documented? Is the treatment of the blocks (ie. fixed or random) stated, and if so, is it justified?

slide-12
SLIDE 12

Slide 12

Sample

55 Potential NCER Studies 13 Potential NCEE Studies 40 Received from direct contact with Principal Investigators 15 Sent request via FOIA and still waiting 9 Received from NCEE directly 3 Received from direct contact with Principal Investigators 1 Sent request to Principal Investigator and still waiting 33 Meet criteria 6 Meet criteria 3 Meet criteria

Pool of Studies

slide-13
SLIDE 13

Slide 13

Sample

9 National Center for Education Evaluation and Regional Assistance 8 Goal 4 Study 25 Goal 3 Study 33 National Center for Education Research Number of Studies

slide-14
SLIDE 14

Slide 14

Methods

Classify the study design Determine plausible values for design

parameters – intra-class correlations, covariate-outcome correlations, explanatory power of blocking

Calculate the recomputed MDES Compare recomputed MDES to stated MDES

slide-15
SLIDE 15

Slide 15

Results – Experimental Designs

Two-Level Cluster Randomized Trial Three-level Cluster Randomized Trial Three-level Multi- site cluster randomized triala Four-Level Multi-site cluster randomized trial Number of Levels 2 3 3 4 Level of Randomization 2 3 2 3 Blocking? No No Yes Yes Number of Studies 5 5 20 11 Example of Nesting Students, Schools Students, Classrooms, Schools Students, Classrooms, Schools Students, Classroom, Schools, Districts

slide-16
SLIDE 16

Slide 16

Results – Experimental Design

Experimental Design Number of NCER Proposals Number of NCEE Proposals Two-Level Cluster Randomized Trial 5 Three-Level Cluster Randomized Trial 5 Three-Level Multi-site cluster randomized trial 13 7 Four-Level Multi-site cluster randomized trial 9 2

slide-17
SLIDE 17

Slide 17

Results - The Recomputed MDES

Plausible values for ICCs

Bloom et al., 1999 Schochet, 2005 Hedges & Hedberg, 2007 Bloom, Richburg-Hayes, & Black, 2007 Murray & Blitstein, 2003

slide-18
SLIDE 18

Slide 18

Results – The Recomputed MDES

Plausible values for covariate-correlations

Bloom, Richburg-Hayes, & Black, 2007

Plausible values for variance explained by

blocking

Hedges & Hedberg, 2007

slide-19
SLIDE 19

Slide 19

Results – Recomputed and Stated MDES

Solid Lines=Recomputed Effect Size Dotted Lines=Stated Effect Size

slide-20
SLIDE 20

Slide 20

Results

Studies 1-24, MDES ranges from 0.40-0.90

NCER studies funded in 2002, 2003, 2004 Less likely to use a covariate

Studies 26-J, MDES ranges from 0.18-0.40

NCER studies funded in 2005, 2006 NCEE studies More likely to use a covariate

slide-21
SLIDE 21

Slide 21

Results - NCEE

0.1 0.2 0.3 0.4 0.5 0.6 0.7 B F G A C-F C-R H-F H-R I-F I-R E-F E-R J-F J-R

NCEE Study I D

Solid Lines=Recomputed Effect Size Dotted Lines=Stated Effect Size

slide-22
SLIDE 22

Slide 22

Results - NCEE

Recomputed MDES ranges from 0.10 – 0.40 Majority of recomputed and stated MDES are

in the same range

slide-23
SLIDE 23

Slide 23

Results - NCER

0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 1 2 3 4 5 10 11 12 13 14 15 16 17 18 19 22 23 24 25 26 28 29 30 31 32-R 32-F 34 35

NCER Goal 3 Study

Solid Lines=Recomputed Effect Size Dotted Lines=Stated Effect Size

slide-24
SLIDE 24

Slide 24

Results - NCER

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 6 7 8 9 20 21 27-A 27-B 33

NCER Goal 4 Study

Solid Lines=Recomputed Effect Size Dotted Lines=Stated Effect Size

slide-25
SLIDE 25

Slide 25

Results - NCER

Similar for goal 3 and 4 studies Recomputed MDES ranges from 0.18 – 1.70 Approximately half of the studies have

recomputed and stated MDES in the same range

slide-26
SLIDE 26

Slide 26

Results – Relationship between stated and expected MDES

Number of NCER Proposals Number of NCEE Proposals MDES within the same range 14 7 Stated MDES < Expected MDES 12 Expected MDES < Stated MDES 1 2 The 6 NCER studies without a power analysis are not included.

slide-27
SLIDE 27

Slide 27

Results – Details of Power Analyses

Number of NCER Proposals Number of NCEE Proposals

Same (n=14) Stated<Recomputed (n=12) Recomputed<Stated (n=1) Same (n=7) Recomputed<Stated (n=2)

Simple statement of power with/without brief citation 6 11 Detailed power analysis with software or documented calculations 8 1 1 7 2 Optimal Design 7 1 1 2 Other 1 7

slide-28
SLIDE 28

Slide 28

Results – Details of Power Analyses

N u m b e r o f N C E R P ro p o s a ls N u m b e r o f N C E E P ro p o s a ls

S a m e (n = 1 5 ) S ta te d < R e c o m p u te d (n = 1 1 ) R e c o m p u te d < S ta te d (n = 1 ) S a m e (n = 7 ) R e c o m p u te d < S ta te d (n = 2 )

IC C e s tim a te n o t in c lu d e d in p ro p o s a l 4 7 2 IC C e s tim a te in c lu d e d in p ro p o s a l 1 1 4 1 5 2 A c a d e m ic IC C s W ith in 0 .1 0 to 0 .2 0 7 1 1 4 N o t w ith in 0 .1 0 to 0 .2 0 3 1 1 2 S o c ia l o r h e a lth IC C s W ith in 0 .0 1 to 0 .0 5 1 N o t w ith in 0 .0 1 to 0 .0 5 1

slide-29
SLIDE 29

Slide 29

Results – Details of Power Analyses

Number of NCER Proposals Number of NCEE Proposals

Same (n=15) Stated<Recomputed (n=11) Recomputed<Stated (n=1) Same (n=7) Recomputed<Stated (n=2)

No covariate 6 6 1 Covariate mentioned not documented 5 3 1 2 1 Covariate documented 4 2 4 1 0.01-0.30 1 0.31-0.50 1 0.51-0.70 4 1 1 1 0.71-0.99 2

slide-30
SLIDE 30

Slide 30

Results – Details of Power Analyses

Number of NCER Proposals Number of NCEE Proposals

Same (n=14) Stated<Recomputed (n=7) Recomputed<Stated (n=1) Same (n=7) Recomputed<Stated (n=2)

Blocking included in the description 14 7 1 7 2 Blocking included in the power analysis 1 3 2 Include explanatory power of blocking 3 Explicitly treat blocks as fixed effects 1 Explicitly treat blocks as random effects 1 Specify the effect size variability 1 1

slide-31
SLIDE 31

Slide 31

Conclusions

Blocked designs are most common

Good for precision

NCEE studies tend to have smaller MDES

Differences in funding Differences in methodological guidelines

slide-32
SLIDE 32

Slide 32

Conclusions

NCEE studies tend to be more accurate

Training

Growth is evident in accuracy and precision of

NCER studies

More precise over time (use of covariates, blocked

designs)

More accurate over time

slide-33
SLIDE 33

Slide 33

Limitations

Study proposals as data Use of original funded proposal