conditional choice probability estimators of single agent
play

Conditional Choice Probability Estimators of Single-Agent Dynamic - PowerPoint PPT Presentation

Conditional Choice Probability Estimators of Single-Agent Dynamic Discrete Choice Models Hotz and Miller (REStud, 1993), Aguirregabiria and Mira (Econometrica, 2002) presented by Anton Laptiev ECON 565, UBC February 7, 2013 Outline


  1. Conditional Choice Probability Estimators of Single-Agent Dynamic Discrete Choice Models Hotz and Miller (REStud, 1993), Aguirregabiria and Mira (Econometrica, 2002) presented by Anton Laptiev ECON 565, UBC February 7, 2013

  2. Outline Single-Agent Models Hotz & Miller Estimator Aguirregabiria & Mira Estimator

  3. Single-Agent Models � time is discrete t = 0,1,.., T � agents have preferences defined over a sequence of states of the world { s it , a it } � s it a vector of state variables that is known at period t � a it denotes a decision chosen at t with a it ∈ A = {0,1,..., J } � agents’ preferences are represented by a utility function T β j U � � � a i , t + j , s i , t + j , β ∈ (0,1) j = 0 � the optimization problem of an agent � � T β j U � � � max a ∈ A E a i , t + j , s i , t + j | a it , s it j = 0

  4. Single-Agent Models � Bellman equation � � �� � � � V ( s it ) = max U ( a , s it ) + β V s i , t + 1 d F s i , t + 1 | a , s it a ∈ A � choice-specific value function � � � � � v ( a , s it ) ≡ U ( a , s it ) + β V s i , t + 1 d F s i , t + 1 | a , s it � the optimal decision rule α ( s it ) = arg max a ∈ A { v ( a , s it )} � assume that s it is divided into two subvectors s it = ( x it , ǫ it )

  5. Assumptions � Assumption AS � the one-period utility function is additively separable in observable and unobservable components U ( a , x it , ǫ it ) = u ( a , x it ) + ǫ it ( a ) � Assumption IID � the unobserved state variables ǫ it are identically distributed over agents and over time with cdf G ǫ ( ǫ it ) � Assumption CI � conditional on the current values of the decision and ob- servable state variables, next period state variables does nor depend on the current ǫ � � � � F x x i , t + 1 | a it , x it , ǫ it = F x x i , t + 1 | a it , x it

  6. Assumptions � Assumption Logit � the unobserved state variables are independent across al- ternatives and have an extreme value type I distribution � Assumption DS � the support of x it is discrete and finite � by assumptions IID and CI, the solution to DP problem is fully characterized by ex ante value function � V ( x it ) = V ( x it , ǫ it ) d G ǫ ( ǫ it ) � by assumption AS the choice-specific value function can be written as follows � � � � � v ( a , x it ) = u ( a , x it ) + β V x i , t + 1 f x x i , t + 1 | a , x it + ǫ it ( a ) x i , t + 1

  7. Forming the Likelihood � � � � Pr a it , x it | ˜ a i , t − 1 , ˜ x i , t − 1 = Pr ( a it | x it ) f x x it | a i , t − 1 , x i , t − 1 T i T i − 1 � � � � l i ( θ ) = log P ( a it | x it , θ ) + log f x x i , t + 1 | a it , x it , θ f + log Pr ( x i 1 | θ ) t = 1 t = 1 � P ( a | x , θ ) ≡ I { α ( x , ǫ ; θ ) = a }d G ǫ ( ǫ ) �� ∀ a ′ �= a a ′ , x it a ′ � � � � � = v ( a , x it ) + ǫ it ( a ) > v + ǫ it d G ǫ ( ǫ it )

  8. The Hotz and Miller Inversion � suppose that payoff function is linear in parameters u ( a , x it ; θ u ) = z ( a , x it ) ′ θ u � the choice-specific value function v ( a , x t ; θ ) = ˜ z ( a , x t ) θ u + ˜ e ( a , x t ; θ ) T − t β j E ( x t + j , ǫ t + j ) | a t = a , x t � � � � � �� z ( a , x t ; θ ) = z ( a , x t ) + ˜ z α x t + j , ǫ t + j ; θ , x t + j j = 1 � � T − t J β j E x t + j | a t = a , x t � � � � � � = z ( a , x t ) + P a t + j | x t + j ; θ z a t + j , x t + j j = 1 a t + j = 0

  9. The Hotz and Miller Inversion T − t β j E ( x t + j , ǫ t + j ) | a t = a , x t � � � � ��� e ( a , x t ; θ ) = ˜ ǫ α x t + j , ǫ t + j ; θ j = 1 � � T − t J β j E x t + j | a t = a , x t � � � � � � P a t + j | x t + j ; θ e a t + j , x t + j = a t + j = 0 j = 1 ǫ t ( a ) | x t , v ( a , x t ) + ǫ t ( a ) ≥ v ( a ′ , x t ) + ǫ t ( a ′ ) ∀ a ′ �= a � � e ( a , x t ) = E 1 � ǫ t ( a ′ ) − ǫ t ( a ) ≤ v ( a , x t ) − v ( a ′ , x t ) ∀ a ′ �= a � � = ǫ t ( a ) I d G ǫ ( ǫ t ) P ( a | x t ) � ǫ t ( a ′ ) − ǫ t ( a ) ≤ v ( a , x t ) − v ( a ′ , x t ) ∀ a ′ �= a � � P ( a | x t ) = ǫ t ( a ) I d G ǫ ( ǫ t ) = ⇒ e ( a , x t ) = f ( P ( a | x t ), G ǫ , ˜ v ( x t )) P ( a | x t ) = f ( G ǫ , ˜ v ( x t )) � where ˜ v ( x t ) is a vector of value differences

  10. The Hotz and Miller Inversion � H&M prove that the letter mapping is invertible = ⇒ e ( a , x t ) = f ( P ( a | x t ), G ǫ ) � by assumption LOGIT, conditional probabilities look as follows � P ( a , x t ) θ u + ˜ P ( a , x t ) � z ˆ e ˆ exp ˜ P ( a | x t ; θ ) = � � P ( a ′ , x t ) θ u + ˜ P ( a ′ , x t ) � J z ˆ e ˆ ˜ a ′ = 0 exp z ˆ P and ˜ e ˆ P are defined on all a ∈ A � where ˜

  11. Example: Renewal Actions � logit errors �� � V ( x t ) = log exp [ v ( x t , a )] + γ a ∈ A � ��� a ∈ A exp [ v ( x t , a )] �� v ( x t , a ′ ) � V ( x t ) = log exp + γ exp [ v ( x t , a ′ )] �� �� v ( x t , a ) − v ( x t , a ′ ) + v ( x t , a ′ ) + γ � = log exp a ∈ A p ( a ′ | x t ) + v ( x t , a ′ ) + γ � � = − log � conditional value function �� v ( x t + 1 , a ′ ) − log p ( a ′ | x t + 1 � �� v ( x t , a ) = u ( x t , a ) + β f x ( x t + 1 | x t , a )d x t + 1 + βγ

  12. Example: Renewal Actions � let a = R indicate the renewal action so that � � x t + 1 | x t , a ′ � � f x ( x t + 1 | x t , a ) f x ( x t + 2 | x t + 1 , R )d x t + 1 = f x f x ( x t + 2 | x t + 1 , R )d x t + 1 ∀ { a , a ′ }, x t + 2 � substitute into the conditional value function �� � �� v ( x t , a ) = u ( x t , a ) + β v ( x t + 1 , R ) − log p ( R | x t + 1 f x ( x t + 1 | x t , a )d x t + 1 + βγ �� � �� = u ( x t , a ) + β u ( x t + 1 , R ) − log p ( R | x t + 1 f x ( x t + 1 | x t , a )d x t + 1 �� + βγ + β 2 V ( x t + 2 ) f x ( x t + 2 | x t + 1 , R ) f x ( x t + 1 | x t , a )d x t + 2 d x t + 1 � the last term is constant across all choices made at time t

  13. Estimation � first stage � P ( a | x t ;) and f x � � x i , t + 1 | a it , x it , θ f are estimated nonpara- metrically directly from the data P and ˜ P by backward induction (finite T ) or us- z ˆ e ˆ � estimate ˜ ing an iterative procedure (requires stationarity) � second stage � estimate θ by GMM using moment conditions of the form N T i � � I { a it = a } − ˆ � � H ( x it ) P ( a | x it ; θ ) = 0 t = 1 i = 1

  14. Advantages of H&M estimator � computational simplicity relative to full solution methods z ˆ P and ˜ e ˆ P are computed only once and remain fixed � ˜ during the search for θ � conditional logit assumption along with the assump- tion of linearity result in the unique solution for the system of GMM equations � pseudo maximum likelihood version of H&M es- timator is asymptotically equivalent to two-step NFP estimator

  15. Limitations of H&M approach � computational gain relies on (strong) assump- tions regarding the form of utility function and the structure of unobservable state variables � in finite samples produces larger bias than the two step NFP estimator (Aguirregabiria & Mira, 2002) � subject to the "curse of dimensionality" � cannot accommodate unobserved heterogene- ity � consistent estimates of conditional choice probabili- ties cannot be recovered in the first stage

  16. Aguirregabiria & Mira estimator � start from the pseudo likelihood version of H&M estima- tor � � z ˆ P ( a it , x it ) θ u + ˜ e ˆ P ( a it , x it ) exp ˜ N T i P , ˆ θ u , ˆ � � � � Q θ f log = � � � J P ( a ′ , x it ) θ u + ˜ P ( a ′ , x it ) z ˆ e ˆ ˜ a ′ = 0 exp i = 1 t = 1 � obtain estimates ˆ θ u � compute new estimates of the choice probabilities � � z ˆ P ( a , x ) θ u + ˜ e ˆ P ( a , x ) exp ˜ P 1 = ˆ ˆ P 1 ( a | x ) = � � � J P ( a ′ , x ) θ u + ˜ P ( a ′ , x ) z ˆ e ˆ ˜ a ′ = 0 exp

  17. Aguirregabiria & Mira estimator P and ˜ z ˆ e ˆ � given ˆ P P 1 one can compute new values ˜ as well as a new pseudo likelihood function Q ( · ) and maximize it to get a new value of ˆ θ u � iterating in this way we can generate a sequence of estimators of structural parameters and CCPs � ˆ θ u , K , ˆ � P K : K = 1,2,... : ∀ K ≥ 1 ˆ θ u , ˆ P K − 1 , ˆ � � θ u , K = arg max θ u ∈ Θ Q θ f � � z ˆ e ˆ P K − 1 ( a , x ) ˆ P K − 1 ( a , x ) exp ˜ θ u , K + ˜ ˆ P K ( a | x ) = � � � J z ˆ P K − 1 ( a ′ , x ) ˆ e ˆ P K − 1 ( a ′ , x ) a ′ = 0 exp ˜ θ u , K + ˜

  18. Advantages of the Aguirregabiria & Mira estimator � Aguirregabiria & Mira (2002) present Monte Carlo evidence that iterating in this procedure produces significant reductions in finite sample bias � as K → ∞ the recursive procedure converges to the two-step NFP estimator even if the initial CCP estimator was inconsistent � A&M show that CCP mapping used in the iter- ative estimation is a contraction mapping and therefore can be used to compute the solution of the DP problem in the space of CCPs

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend