Conditional Choice Probability Estimators of Single-Agent Dynamic - - PowerPoint PPT Presentation

conditional choice probability estimators of single agent
SMART_READER_LITE
LIVE PREVIEW

Conditional Choice Probability Estimators of Single-Agent Dynamic - - PowerPoint PPT Presentation

Conditional Choice Probability Estimators of Single-Agent Dynamic Discrete Choice Models Hotz and Miller (REStud, 1993), Aguirregabiria and Mira (Econometrica, 2002) presented by Anton Laptiev ECON 565, UBC February 7, 2013 Outline


slide-1
SLIDE 1

Conditional Choice Probability Estimators of Single-Agent Dynamic Discrete Choice Models

Hotz and Miller (REStud, 1993), Aguirregabiria and Mira (Econometrica, 2002)

presented by Anton Laptiev

ECON 565, UBC

February 7, 2013

slide-2
SLIDE 2

Outline

Single-Agent Models Hotz & Miller Estimator Aguirregabiria & Mira Estimator

slide-3
SLIDE 3

Single-Agent Models

time is discrete t = 0,1,..,T agents have preferences defined over a sequence of states

  • f the world {sit,ait}

sit a vector of state variables that is known at period t ait denotes a decision chosen at t with ait ∈ A = {0,1,...,J}

agents’ preferences are represented by a utility function

T

  • j=0

βjU

  • ai,t+j, si,t+j
  • , β ∈ (0,1)

the optimization problem of an agent

max

a∈A E

  • T
  • j=0

βjU

  • ai,t+j, si,t+j
  • | ait, sit
slide-4
SLIDE 4

Single-Agent Models

Bellman equation

V(sit) = max

a∈A

  • U (a, sit)+β
  • V
  • si,t+1
  • dF
  • si,t+1 | a, sit
  • choice-specific value function

v(a, sit) ≡ U (a, sit)+β

  • V
  • si,t+1
  • dF
  • si,t+1 | a, sit
  • the optimal decision rule

α(sit) = arg max

a∈A {v(a, sit)}

assume that sit is divided into two subvectors sit = (xit, ǫit)

slide-5
SLIDE 5

Assumptions

Assumption AS

the one-period utility function is additively separable in

  • bservable and unobservable components

U (a, xit, ǫit) = u(a, xit)+ǫit (a)

Assumption IID

the unobserved state variables ǫit are identically distributed

  • ver agents and over time with cdf Gǫ (ǫit)

Assumption CI

conditional on the current values of the decision and ob-

servable state variables, next period state variables does nor depend on the current ǫ Fx

  • xi,t+1 | ait, xit, ǫit
  • = Fx
  • xi,t+1 | ait, xit
slide-6
SLIDE 6

Assumptions

Assumption Logit

the unobserved state variables are independent across al-

ternatives and have an extreme value type I distribution

Assumption DS

the support of xit is discrete and finite

by assumptions IID and CI, the solution to DP problem is

fully characterized by ex ante value function V (xit) =

  • V (xit, ǫit) dGǫ (ǫit)

by assumption AS the choice-specific value function can

be written as follows v(a, xit) = u(a, xit)+β

  • xi,t+1

V

  • xi,t+1
  • fx
  • xi,t+1 |a, xit
  • +ǫit (a)
slide-7
SLIDE 7

Forming the Likelihood

Pr

  • ait, xit | ˜

ai,t−1, ˜ xi,t−1

  • = Pr (ait | xit) fx
  • xit | ai,t−1, xi,t−1
  • li (θ) =

Ti

  • t=1

log P(ait | xit, θ)+

Ti−1

  • t=1

log fx

  • xi,t+1 | ait, xit, θf
  • +log Pr (xi1 | θ)

P(a | x, θ) ≡

  • I {α(x, ǫ; θ) = a}dGǫ (ǫ)

=

  • v(a, xit)+ǫit (a) > v
  • a′, xit
  • +ǫit
  • a′

∀a′ = a

  • dGǫ (ǫit)
slide-8
SLIDE 8

The Hotz and Miller Inversion

suppose that payoff function is linear in parameters

u(a, xit; θu) = z(a, xit)′ θu

the choice-specific value function

v(a, xt; θ) = ˜ z(a, xt)θu + ˜ e(a, xt; θ)

˜ z(a, xt; θ) = z(a, xt)+

T−t

  • j=1

βj E(xt+j,ǫt+j)|at=a,xt

  • z
  • α
  • xt+j,ǫt+j; θ
  • ,xt+j
  • = z(a, xt)+

T−t

  • j=1

βj Ext+j | at=a,xt

  • J
  • at+j=0

P

  • at+j | xt+j; θ
  • z
  • at+j, xt+j
slide-9
SLIDE 9

The Hotz and Miller Inversion

˜ e(a, xt; θ) =

T−t

  • j=1

βj E(xt+j,ǫt+j)|at=a,xt

  • ǫ
  • α
  • xt+j,ǫt+j; θ
  • =

T−t

  • j=1

βj Ext+j | at=a,xt

  • J
  • at+j=0

P

  • at+j | xt+j; θ
  • e
  • at+j, xt+j
  • e(a, xt) = E
  • ǫt(a) | xt, v(a, xt)+ǫt(a) ≥ v(a′, xt)+ǫt(a′) ∀a′ = a
  • =

1 P(a | xt)

  • ǫt(a)I
  • ǫt(a′)−ǫt(a) ≤ v(a, xt)−v(a′, xt) ∀a′ = a
  • dGǫ(ǫt)

P(a | xt) =

  • ǫt(a)I
  • ǫt(a′)−ǫt(a) ≤ v(a, xt)−v(a′, xt) ∀a′ = a
  • dGǫ(ǫt)

= ⇒ e(a, xt) = f (P(a | xt), Gǫ, ˜ v(xt)) P(a | xt) = f (Gǫ, ˜ v(xt))

where ˜

v(xt) is a vector of value differences

slide-10
SLIDE 10

The Hotz and Miller Inversion

H&M prove that the letter mapping is invertible

= ⇒ e(a, xt) = f (P(a | xt), Gǫ)

by assumption LOGIT, conditional probabilities look as

follows P(a | xt; θ) = exp

  • ˜

P (a, xt)θu + ˜

P (a, xt)

  • J

a′=0 exp

  • ˜

P (a′, xt)θu + ˜

P (a′, xt)

  • where ˜

P and ˜

P are defined on all a ∈ A

slide-11
SLIDE 11

Example: Renewal Actions

logit errors

V(xt) = log

  • a∈A

exp[v(xt, a)]

V(xt) = log

  • exp
  • v(xt, a′)
  • a∈A exp[v(xt, a)]

exp[v(xt, a′)]

= log

  • a∈A

exp

  • v(xt, a)−v(xt, a′)
  • +v(xt, a′)+γ

= −log

  • p(a′ | xt)
  • +v(xt, a′)+γ

conditional value function

v(xt, a) = u(xt, a)+β

  • v(xt+1,a′)−log
  • p(a′|xt+1
  • fx(xt+1|xt,a)dxt+1+βγ
slide-12
SLIDE 12

Example: Renewal Actions

let a = R indicate the renewal action so that

  • fx (xt+1|xt,a)fx (xt+2|xt+1,R)dxt+1 =
  • fx
  • xt+1|xt,a′

fx (xt+2|xt+1,R)dxt+1 ∀ {a, a′}, xt+2

substitute into the conditional value function

v(xt, a) = u(xt, a)+β

  • v(xt+1,R)−log
  • p(R|xt+1
  • fx(xt+1|xt,a)dxt+1+βγ

= u(xt, a)+β

  • u(xt+1,R)−log
  • p(R|xt+1
  • fx(xt+1|xt,a)dxt+1

+βγ+β2

  • V(xt+2) fx (xt+2|xt+1,R)fx (xt+1|xt,a)dxt+2 dxt+1

the last term is constant across all choices made at time t

slide-13
SLIDE 13

Estimation

first stage

P(a | xt;) and fx

  • xi,t+1 | ait, xit, θf
  • are estimated nonpara-

metrically directly from the data

estimate ˜

P and ˜

P by backward induction (finite T) or us-

ing an iterative procedure (requires stationarity)

second stage

estimate θ by GMM using moment conditions of the form

N

  • i=1

Ti

  • t=1

H (xit)

  • I {ait = a}− ˆ

P(a | xit; θ)

  • = 0
slide-14
SLIDE 14

Advantages of H&M estimator

computational simplicity relative to full solution

methods

˜

P and ˜

P are computed only once and remain fixed

during the search for θ

conditional logit assumption along with the assump-

tion of linearity result in the unique solution for the system of GMM equations

pseudo maximum likelihood version of H&M es-

timator is asymptotically equivalent to two-step NFP estimator

slide-15
SLIDE 15

Limitations of H&M approach

computational gain relies on (strong) assump-

tions regarding the form of utility function and the structure of unobservable state variables

in finite samples produces larger bias than the

two step NFP estimator (Aguirregabiria & Mira, 2002)

subject to the "curse of dimensionality" cannot accommodate unobserved heterogene-

ity

consistent estimates of conditional choice probabili-

ties cannot be recovered in the first stage

slide-16
SLIDE 16

Aguirregabiria & Mira estimator

start from the pseudo likelihood version of H&M estima-

tor Q

  • θu, ˆ

P, ˆ θf

  • =

N

  • i=1

Ti

  • t=1

log exp

  • ˜

P (ait, xit)θu + ˜

P (ait, xit)

  • J

a′=0 exp

  • ˜

P (a′, xit)θu + ˜

P (a′, xit)

  • btain estimates ˆ

θu

compute new estimates of the choice probabilities

ˆ P1 = ˆ P1 (a | x) = exp

  • ˜

P (a, x)θu + ˜

P (a, x)

  • J

a′=0 exp

  • ˜

P (a′, x)θu + ˜

P (a′, x)

slide-17
SLIDE 17

Aguirregabiria & Mira estimator

given ˆ

P1 one can compute new values ˜ zˆ

P and ˜

P

as well as a new pseudo likelihood function Q(·) and maximize it to get a new value of ˆ θu

iterating in this way we can generate a sequence

  • f estimators of structural parameters and CCPs

ˆ θu, K, ˆ PK : K = 1,2,...

  • : ∀K ≥ 1

ˆ θu,K = arg max

θu∈Θ Q

  • θu, ˆ

PK−1, ˆ θf

  • ˆ

PK (a | x) = exp

  • ˜

PK−1 (a, x) ˆ

θu, K + ˜ eˆ

PK−1 (a, x)

  • J

a′=0 exp

  • ˜

PK−1 (a′, x) ˆ

θu, K + ˜ eˆ

PK−1 (a′, x)

slide-18
SLIDE 18

Advantages of the Aguirregabiria & Mira estimator

Aguirregabiria & Mira (2002) present Monte Carlo

evidence that iterating in this procedure produces significant reductions in finite sample bias

as K → ∞ the recursive procedure converges to

the two-step NFP estimator even if the initial CCP estimator was inconsistent

A&M show that CCP mapping used in the iter-

ative estimation is a contraction mapping and therefore can be used to compute the solution

  • f the DP problem in the space of CCPs