adaptive experiments for policy choice
play

Adaptive Experiments for Policy Choice Maximilian Kasy Anja - PowerPoint PPT Presentation

Adaptive Experiments for Policy Choice Maximilian Kasy Anja Sautmann December 7, 2018 Introduction Consider an NGO who wants to encourage kangaroo care for prematurely born babies know to be effective if used. There are


  1. Adaptive Experiments for Policy Choice Maximilian Kasy Anja Sautmann December 7, 2018

  2. Introduction • Consider an NGO who wants to encourage “kangaroo care” for prematurely born babies – know to be effective if used. • There are numerous implementation choices: • Incentives for health-care providers; • Methods for educating mothers and nurses; • Involvement of fathers and other relatives; • Nurse home visits vs. hospitalization... • We argue: • NGO should run an experiment in multiple waves. • Initially, try many different variants. • Later, focus the experiment on the best performing options. • Once the experiment is concluded, recommend the best performing option. • Principled approach for pilot studies, or “tinkering.” • In the spirit of “the economist as plumber” (Duflo, 2017). 1 / 30

  3. Introduction • Our setting: • Multiple waves. • Objective: 1. After the experiment pick a policy 2. to maximize social welfare. • How to design experiments for this objective? • Contrast with canonical field experiments: • One wave. • Objectives: 1. Estimate average treatment effect. 2. Test whether it equals 0. • Design recommendations: 1. Same number of observations for each treatment. 2. If possible stratify. 3. Choose sample size based on power calculations. 2 / 30

  4. Introduction Preview of findings • The distinction matters: • Optimal designs look qualitatively different for different objective functions. • Adaptive designs for policy choice improve welfare. • Implementation: • Optimal designs are feasible but computationally challenging. • Good and easily computed approximations are available. • Features of optimal designs: • Adapt to the outcomes of previous waves. • Discard treatments that are clearly not optimal. • Marginal value of observations for a given treatment is non-monotonic. 3 / 30

  5. Introduction Literature • Multi-armed bandits – related but different: • Goal is to maximize outcomes of experimental units (rather than choose policy after experiment). • Exploration-exploitation trade-off (we focus on “exploration”). • Units come in sequentially (rather than in waves). • Good reviews: • Gittins index (optimal solution to some bandit problems): Weber et al. (1992) • Adaptive designs in clinical trials: Berry (2006). • Regret bounds for bandit problems: Bubeck and Cesa-Bianchi (2012). • Reinforcement learning: Ghavamzadeh et al. (2015). • Thompson sampling: Russo et al. (2018). • Empirical examples for our simulations: Bryan et al. (2014), Ashraf et al. (2010), Cohen et al. (2015) 4 / 30

  6. Introduction Setup Optimal treatment assignment Modified Thompson sampling Inference Conclusion

  7. Setup • Waves t = 1 , . . . , T , sample sizes N t . • Treatment D ∈ { 1 , . . . , k } , outcomes Y ∈ { 0 , 1 } . • Potential outcomes Y d . • Repeated cross-sections: ( Y 0 it , . . . , Y k it ) are i.i.d. across both i and t . • Average potential outcome: θ d = E [ Y d it ] . • Key choice variable: Number of units n d t assigned to D = d in wave t . • Outcomes: Number of units s d t having a “success” (outcome Y = 1). 5 / 30

  8. Setup Treatment assignment, outcomes, state space • Treatment assignment in wave t : n t = ( n 1 t , . . . , n k t ) . • Outcomes of wave t : s t = ( s 1 t , . . . , s k t ) . • Cumulative versions: � � � M t = N t ′ , m t = n t , r t = s t . t ′ ≤ t t ′ ≤ t t ′ ≤ t • Relevant information for the experimenter in period t + 1 is summarized by m t and r t . • Total trials for each treatment, total successes. 6 / 30

  9. Setup Design objective • Policy objective SW ( d ): Average outcome Y , net of the cost of treatment. • Choose treatment d after the experiment is completed. • Posterior expected social welfare: SW ( d ) = E [ θ d | m T , r T ] − c d , where c d is the unit cost of implementing policy d . 7 / 30

  10. Setup Bayesian prior and posterior • By definition, Y d | θ ∼ Ber ( θ d ). • Prior: θ d ∼ Beta ( α d 0 , β d 0 ), independent across d . • Posterior after period t : θ d | m t , r t ∼ Beta ( α d t , β d t ) α d t = α d 0 + r d t β d t = β d 0 + m d t − r d t . • In particular, α d 0 + r d T − c d . SW ( d ) = α d 0 + β d 0 + m d T 8 / 30

  11. Introduction Setup Optimal treatment assignment Modified Thompson sampling Inference Conclusion

  12. Optimal treatment assignment Optimal assignment: Dynamic optimization problem • Dynamic stochastic optimization problem: • States ( m t , r t ), • actions n t . • Solve for the optimal experimental design using backward induction. • Denote by V t the value function after completion of wave t . • Starting at the end, we have � � α d 0 + r d T − c d V T ( m T , r T ) = max . α d 0 + β d 0 + m d d T • Finite state and action space. ⇒ Can, in principle, solve directly for optimal rule. • But: Computation time quickly explodes. 9 / 30

  13. Optimal treatment assignment Simple examples • Consider a small experiment with 2 waves, 3 treatment values (minimal interesting case). • The following slides plot expected welfare as a function of: 1. Division of sample size between waves, N 1 + N 2 = 10. N 1 = 6 is optimal. 2. Treatment assignment in wave 2, given wave 1 outcomes. N 1 = 6 units in wave 1, N 2 = 4 units in wave 2. • Keep in mind: α 1 = (1 , 1 , 1) + s 1 β 1 = (1 , 1 , 1) + n 1 − s 1 10 / 30

  14. Optimal treatment assignment Dividing sample size between waves • N 1 + N 2 = 10. • Expected welfare as a function of N 1 . • Boundary points ≈ 1-wave experiment. • N 1 = 6 (or 5) is optimal. 0.700 V 0 0.698 0.696 0 1 2 3 4 5 6 7 8 9 10 N 1 11 / 30

  15. Optimal treatment assignment α = ( 2, 2, 2 ), β = ( 2, 2, 2 ) n2=N 0.564 0.594 0.594 0.585 0.595 0.585 0.594 0.595 0.595 0.594 0.564 0.594 0.585 0.594 0.564 n3=N n1=N 12 / 30

  16. Optimal treatment assignment α = ( 2, 2, 3 ), β = ( 2, 2, 1 ) n2=N 0.750 0.756 0.750 0.755 0.755 0.750 0.758 0.758 0.755 0.750 0.754 0.755 0.756 0.750 0.758 n3=N n1=N 13 / 30

  17. Optimal treatment assignment α = ( 3, 3, 1 ), β = ( 1, 1, 3 ) n2=N 0.804 0.804 0.812 0.800 0.805 0.805 0.788 0.788 0.805 0.812 0.750 0.788 0.800 0.804 0.804 n3=N n1=N 14 / 30

  18. Introduction Setup Optimal treatment assignment Modified Thompson sampling Inference Conclusion

  19. Modified Thompson sampling A simpler alternative • Old proposal by Thompson (1933) for clinical trials; popular in online experimentation. • Assign each treatment with probability equal to the posterior probability that it is optimal. • Easily implemented: Sample draws � θ it from the posterior, assign ˆ θ d D it = argmax it . d • We propose two modifications : 1. Don’t assign the same treatment twice in a row. 2. Re-run the algorithm several times, and use average n d t for each treatment d . 15 / 30

  20. Modified Thompson sampling Justifications 1. Mimics the qualitative behavior of optimal assignment in examples. 2. Thompson sampling has strong theoretical justifications (regret bounds) in multi armed bandit setting. 3. Modifications motivated by differences in setting: a) No exploitation motive. b) Waves rather than sequential arrival. 4. Performs well in calibrated simulations (coming up). 5. Is easy to compute. 6. Is easy to adapt to more general models. 16 / 30

  21. Modified Thompson sampling Extension: Covariates and treatment targeting • Suppose now that 1. We additionally observe a (discrete) covariate X . 2. The policy to be chosen can target treatment by X . • Implications for experimental design? 1. Simple solution: Treat each covariate cell as its separate experiment; all the above applies. 2. Better solution: Set up a hierarchical Bayes model, to optimally combine information across treatment cells. • Example of a hierarchical Bayes model: Y d | X = x , θ dx , ( α d 0 , β d 0 ) ∼ Ber ( θ dx ) θ dx | ( α d 0 , β d 0 ) ∼ Beta ( α d 0 , β d 0 ) ( α d 0 , β d 0 ) ∼ π, 17 / 30

  22. Modified Thompson sampling Calibrated simulations • Simulate data calibrated to estimates of 3 published experiments. • Set θ equal to observed average outcomes for each stratum and treatment. • Total sample size same as original. Ashraf, N., Berry, J., and Shapiro, J. M. (2010). Can higher prices stimulate product use? Evidence from a field experiment in Zambia. American Economic Review , 100(5):2383–2413 Bryan, G., Chowdhury, S., and Mobarak, A. M. (2014). Underinvestment in a profitable technology: The case of seasonal migration in Bangladesh. Econometrica , 82(5):1671–1748 Cohen, J., Dupas, P., and Schaner, S. (2015). Price subsidies, diagnostic tests, and targeting of malaria treatment: evidence from a randomized controlled trial. American Economic Review , 105(2):609–45 18 / 30

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend