a computationally practical simulation estimation
play

A Computationally Practical Simulation Estimation Algorithm for - PDF document

A Computationally Practical Simulation Estimation Algorithm for Dynamic Panel Data Models with Unobserved Endogenous State Variables and Implications of Classification Error for the Dynamics of Female Labor Supply: A Comment on Hyslop (1999)


  1. A Computationally Practical Simulation Estimation Algorithm for Dynamic Panel Data Models with Unobserved Endogenous State Variables and Implications of Classification Error for the Dynamics of Female Labor Supply: A Comment on Hyslop (1999) Michael P. Keane University of Technology Sydney Arizona State University Robert M. Sauer University of Southampton

  2. Introduction  Missing endogenous state variables is a widespread problem in panel discrete choice models  Present when unobserved initial conditions and missing choices during sample period  Assess and implement a new SML algorithm that specifically addresses these issues  Main advantage: computationally easy  relies on unconditional simulation of data from a model  conditional simulation of choice probabilities when history not observed is difficult (e.g., GHK, MCMC, EM)

  3. Plan of Presentation  Specify Panel Data Probit Model  Discuss Models of Classification Error (CE)  Describe the SML Algorithm  Show Monte Carlo Results  Application to Female Labor Supply with PSID data  Implications of CE itself for endogeneity of fertility and nonlabor income

  4. Panel Probit Model t − 1 u it   0   1 x it  ∑ d i      it   0 1 if u it ≥ 0 d it  0 otherwise.  it   i   it  it   1  i , t − 1   it x it   2 x i , t − 1   it 2   i  N  0,    it  N  0,   2   it  N  0,  v 2  simulate data from model so algorithm can easily handle wider range of distributions

  5. Classification Error (CE)  Required to form likelihood in our approach  CE probabilistically "matches" simulated choice and reported choice  General model of CE we consider ∗  0 | d it  1  10 t  Pr d it ∗  1 | d it  0  01 t  Pr d it  00 t  1 −  01 t  11 t  1 −  10 t as in Poterba and Summers  1986,1995  and HAS  1998   HAS  1998  develops identification conditions (more later)  Only need tractable expression for  jkt ’s to form likelihood

  6. CE Model 1: Unbiased Classification Error  imposes unconditional prob report an option equals true prob ∗  1   Pr  d it  1  Pr  d it generates tractable linear expressions for  jkt ’s ∗  1 | d it  1  11 t  Pr d it  E   1 − E  Pr  d it  1  ∗  1 | d it  0  01 t  Pr d it   1 − E  Pr  d it  1  because ∗  1    11 t Pr  d it  1    01 t Pr  d it  0  Pr  d it  11 t  E   1 − E  Pr  d it  1  ∗  1   Pr  d it  1   Pr  d it

  7. CE Model 1: Unbiased Classification Error  In  11 t  E   1 − E  Pr  d it  1   E is estimable parameter where  low prob events have prob equal to E of being classified correctly  prob of correct classification increases linearly in E  Pr  d it  1  is easily simulated

  8. CE Model 2: Biased Classification Error ∗  1   Pr  d it  1   Do not impose Pr  d it rather assume: ∗ l it   0   1 d it   2 d it − 1   it 1 if l it ≥ 0 ∗  d it 0 otherwise.  it  logistic  Get tractable expressions for  jkt ’s: ∗ e  0   1   2 d it − 1 ∗  1 | d it  1  11 t  Pr d it  ∗ 1  e  0   1   2 d it − 1 ∗ e  0   2 d it − 1 ∗  1 | d it  0  01 t  Pr d it  ∗ 1  e  0   2 d it − 1  Note that easily incorporates dynamic (persistent) misreporting ∗  d it − 1 can be simulated from l it model if missing

  9. Identification ∗  1   Again, rewrite Pr  d it   11 t Pr  d it  1    01 t Pr  d it  0    1 −  10 t  Pr  d it  1    01 t  1 − Pr  d it  1    01 t   1 −  10 t −  01 t  Pr  d it  1   need non-linear Pr  d it  1  and monotonicity  10 t   01 t  1  otherwise standard identification issues  SD effects from causal effect of lagged X’s  SD effects sensitive to modeling of serial correlation (RE or AR(1), etc.)  in biased CE model, lagged reported choice identified because not perfectly correlated with lagged X’s

  10. The SML Estimation Algorithm ∗   d it N , D i T , x i   x it  t  1 ∗ , x i  i  1 ∗  t  1 Data:  D i T 1 . Draw M times from the  it distribution to T  i  1 M N   it m  t  1 form m  1 T  i  1 N and 2 . Given  x it  t  1 M , construct T  i  1 N   it m  t  1 m  1 T  i  1 M N  d it m  t  1 according to model m  1 3 . Construct conditional probs  m  t  1 M T  jkt m  1 4 . Form a simulator likelihood contribution: ∗ |  , x i  P D i ∗ observed 1  I d it 1 M T m  j , d it ∗  k  m I  d it M ∑  ∑ ∑  jkt 1 m  1 t  1 j  0 k  0

  11. Important Things to Note  any observed choice history has non-zero prob conditional on any simulated history  building likelihood off of unconditional simulations  state space updated according to simulated outcomes  completely circumvents problem of partially observed choice history  consistency and asymptotic normality N →  as N →  (as in M require that Pakes and Pollard  1989  and Lee  1992  )  but still need to check small sample properties of estimator

  12. Some More Important Things to Note  if missing x it ’s, can simulate them and modify likelihood to include density of x it ’s  if initial conditions problem can simulate model from t  0,..., T with d i 0  x i 0  0  or can imbed Heckman’s solution - specify marginal distribution for d i   , then simulate from t    ,..., T  or can imbed Wooldridge’s solution - random effect function of d i   and covariates, simulate from t     1,..., T  for standard errors, or to use gradient based optimization, have smooth version of the algorithm

  13. Importance Sampling (Smooth) Version of Algorithm  construct  d it m   0  t  1 and  U it m   0  t  1 T T and hold fixed as vary   calculate importance sampling weights as vary  m   0  ,..., U iT m   0  |  , x i g U i 1 W m     m   0  ,..., U iT m   0  |  0 , x i g U i 1 T m   0  |  , x i    g  U i     a    1 t  1 t − 1 m   0  −  0 −  1 x it − ∑ a  U it m   0    d i    0  likelihood contribution becomes M T ∗ |  , x i M ∑ W m      1 f m  x it  I x it observed P D i t   m  1  ∗ observed 1  I d it 1 m  j , d it ∗  k  m I  d it ∑ ∑  jkt j  0 k  0

  14. Repeated Sampling Experiments  RE  AR(1) Polya Model with Exponential Decay: t − 1 u it   0   1 x it  ∑ d i      it   0     e −   t −  − 1   it   i   it  it   1  i , t − 1   it 2  1 −  1  it  N  0,  1 −   2  x it   2 x i , t − 1   it ,  it  N  0,  v 2   Sample Size: N  500, T  10  Replications: R  50  Simulated histories per person: M  1000  Also do experiments on Markov model

  15. Table 3 Repeated Sampling Experiments Random E ff ects Polya Model Unbiased Classi fi cation Error (Missing X’s, No Initial Conditions Problem) Mean b Median b Std ( c Parameter True Value β β β ) RMSE t-Stat 20% Missing Choices and X’s ( t = 1 , ..., 10) β 0 -.1000 -.1051 -.1023 .0436 .0439 -.83 β 1 1.0000 1.0167 1.0191 .0611 .0634 1.92 ρ 1.0000 1.0479 1.0446 .0444 .0653 7.63 α .5000 .4977 .5031 .0656 .0657 -.24 φ 2 .2500 .2520 .2505 .0176 .0177 .80 σ ν .5000 .5015 .5016 .0057 .0059 1.86 σ μ .8000 .8056 .8017 .0287 .0292 1.38 E .7500 .7428 .7430 .0172 .0187 -2.95 40% Missing Choices and X’s ( t = 1 , ..., 10) β 0 -.1000 -.1087 -.1099 .0539 .0546 -1.15 β 1 1.0000 1.0141 1.0233 .0678 .0692 1.48 ρ 1.0000 1.0458 1.0374 .0636 .0784 5.10 α .5000 .4953 .4949 .0600 .0602 .56 φ 2 .2500 .2521 .2546 .0253 .0254 .59 σ ν .5000 .5012 .5012 .0069 .0070 1.21 σ μ .8000 .8046 .8063 .0347 .0350 .94 E .7500 .7474 .7416 .0245 .0246 -.74 60% Missing Choices and X’s ( t = 1 , ..., 10) β 0 -.1000 -.0997 -.1116 .0542 .0543 .05 β 1 1.0000 1.034 1.0258 .0894 .0924 1.85 ρ 1.0000 1.0401 1.0512 .0682 .0791 4.15 α .5000 .4957 .4973 .0721 .0722 -.42 φ 2 .2500 .2507 .2498 .0372 .0373 .13 σ ν .5000 .5011 .5017 .0089 .0090 .88 σ μ .8000 .8096 .8044 .0421 .0432 1.61 E .7500 .7493 .7440 .0288 .0288 -.16 Note: The number of replications in each experiment is 50 and the number of individuals in the sample is 500 . Std ( c β ) and RMSE refer to the sample standard deviation and the root mean square µ ¶ √ Mean � β − β error, respectively, of the estimated parameters. The t-statistics are calculated as 50 . Std ( � β ) The model is the same as in Table 1 . 1

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend