Pas ast t Ev Evalua aluations: tions: What Do We Know? - - PowerPoint PPT Presentation

pas ast t ev evalua aluations tions what do we know
SMART_READER_LITE
LIVE PREVIEW

Pas ast t Ev Evalua aluations: tions: What Do We Know? - - PowerPoint PPT Presentation

Pas ast t Ev Evalua aluations: tions: What Do We Know? Presentation at the Secretarys Innovation Group Washing ington, , DC DC April 10, 2014 Peter Schochet, Ph.D., Senior Fellow Rigorous Evaluations Are Feasible! Many


slide-1
SLIDE 1

Pas ast t Ev Evalua aluations: tions:

Presentation at the Secretary’s Innovation Group Washing ington, , DC DC

What Do We Know?

Peter Schochet, Ph.D., Senior Fellow

April 10, 2014

slide-2
SLIDE 2

2

Rigorous Evaluations Are Feasible!

  • Many informative random assignment studies

have been conducted

– Range of interventions, including SNAP – Multiple settings – Diverse populations similar to SNAP recipients

slide-3
SLIDE 3

3

What Employment Strategies Work?

  • Models that combine

– Work experience – Skills training (especially in community colleges) – Intensive case management and support services – Activities that target specific industries

  • Providing only transitional jobs does not have

long-term effects

slide-4
SLIDE 4

4

How Can the Research Be Improved?

  • Unify the class of tested intervention across sites

– Help interpret findings

  • Introduce planned variation

– Go beyond the single treatment and control group – Vary promising intervention components

  • Evaluators should be selected early
slide-5
SLIDE 5

5

For More Information

  • Peter Schochet

pschochet@mathematica-mpr.com

slide-6
SLIDE 6

Quasi-Experimental Designs for Social Policy Interventions

Peter Z. Schochet

Ph.D., Senior Fellow

Presentation at the Secretary’s Innovation Group Washington, DC

slide-7
SLIDE 7

Introduction and Summary

  • There have been significant advances in

the use of quasi-experimental methods to create credible comparison groups

  • Experimental methods are still the best

starting point for impact evaluations

– Ensure unbiased estimates – Most precise estimates

2

slide-8
SLIDE 8
  • Cannot always do RCTs

– Entitlement programs – Undersubscribed programs – Site refusals

  • Takes time to get results

Problems With Random Assignment

3

slide-9
SLIDE 9
  • Pre-post or interrupted time series (ITS)
  • Matched comparison group or propensity

scoring

  • Instrumental variable (IV)
  • Regression discontinuity (RD)

What Are Alternative Designs?

4

slide-10
SLIDE 10
  • Ok if pre-period outcomes are very stable and

there are large post-period effects

Pre-Post or ITS Designs

5

10 20 30 40 50 60 70 80

2000 2001 2002 2003 2004 2005 2006 2007 2008 2009

%% Girls in School, by Year

Girl-Friendly Schools Built

Year

slide-11
SLIDE 11
  • Some studies found that these methods cannot

replicate impacts from experiments

– LaLonde (1986); Fraker and Maynard (1987); Agodini and Dynarski (2004); Peikes et al. (2008)

  • Some studies are more optimistic

– Heckman and Hotz (1989); Deheija and Wahba (1999); Mueser et

  • al. (2007); Shadish et al. (2008)
  • Some have expressed extreme caution

– Smith and Todd (2005); Fortson et al. (2012)

  • Literature on conditions with better replications

– Glazerman et al. (2003); Heckman et al. (1997); Bloom et al. (2005); Cook and Wong (2008)

Matched Comparison Group Designs

6

slide-12
SLIDE 12
  • Scoring rule is used to define who gets the

treatment

– Income threshold – Risk index

  • Becoming increasingly popular
  • Replication studies are promising (Cook &

Wong 2008, Gleason et al. 2012)

RD Designs

7

slide-13
SLIDE 13

Example: Early Reading First Evaluation

8

10 20 30 40 50 60 70 40 45 50 55 60 65 70 75 80 85 90 95 Application Score Print Awareness Score

Unfunded Funded

Impact

Cutoff Score

Grants Were Awarded to Sites with the Highest Application Scores

slide-14
SLIDE 14
  • Credible quasi-experimental designs are

available if RCTs are not an option

– But need the right conditions – Need larger samples than experimental designs

Conclusions

9

slide-15
SLIDE 15

Samp Sample le Siz Size: e:

Presentation at the Secretary’s Innovation Group Washing ington, , DC DC

How many study participants?

Peter Schochet, Ph.D., Senior Fellow

April 10, 2014

slide-16
SLIDE 16

2

Having Sufficient Samples Is Critical

  • Estimates of program effects are measured

with error

  • Need large samples to be able to say that

likely program effects are different than zero

  • Requires sufficient enrollment to generate

large treatment and control groups

slide-17
SLIDE 17

3

What Determines Sample Size Needs?

  • Unit of random assignment

–Smaller samples if individuals are randomized than “groups”

  • Expected effects

–Smaller samples if impacts are likely to be large

  • Whether sites can be pooled
  • How much the outcomes vary across people
slide-18
SLIDE 18

4

Example of Sample Size Requirements

Number of Sites

(100 treatments, 100 controls per site)

Individuals Randomized SNAP Offices Randomized

(10 per site)

1 17 20 5 8 10 10 5 7

Mi Minimum nimum Pr Prog

  • gram Ef

am Effec ects ts on

  • n E

Emplo mploymen yment

(Per ercen centa tage ge Poin

  • ints

ts)