SLIDE 1
Statistical Modeling and Analysis of Neural Data (NEU 560) Princeton University, Spring 2018 Jonathan Pillow
Lecture 17 notes: Latent variable models and EM
Thurs, 4.12
1 Quick recap of EM
L =
log-likelihood
- log p(x|θ) = log
- total-data
likelihood
- p(x, z|θ)
dx ≥
- q(z|φ)
variational distribution
log p(x, z|θ) q(z|φ)
- dx
negative free energy
F(φ, θ) (1) The negative free energy has two convenient forms, which we exploit in the two alternating phases
- f EM:
F(φ, θ) = log p(x|θ) − KL(q(z|φ)||p(z|x, θ)) (used in E-step) (2) F(φ, θ) =
- q(z|φ) log p(x, z|θ)dz + H[q(z|φ)]
- indep of θ
(used in M-step) (3) Specifically, EM involves alternating between:
- E-step: Update φ by setting q(z|φ) = p(z|x, θ), with θ held fixed.
- M-step: Update θ by maximizing
- q(z|φ) log p(x, z|θ)dz, with φ held fixed.
Note that for discrete latent variable models, where the latent z takes on finite or countably infinitely many discrete values, the integral over z is replaced by a sum: F(φ, θ) =
m
- j=0
q(z = αj|φ) log p(x, z = αj|θ)dz + H[q(z|φ)]
- indep of θ