SLIDE 1
Statistical Modeling and Analysis of Neural Data (NEU 560) Princeton University, Spring 2018 Jonathan Pillow
Lecture 19 notes: FA and Probabilistic PCA
Thurs, 4.19
1 Factor Analysis (FA): quick recap
To recap, the FA model is defined by a low-dimensional Gaussian latent variable, a linear mapping to higher dimensional observation space, and indepedent Gaussian noise with different variances along each dimension: The model: z ∼ N(0, Im) (1) x = Az + ǫ, ǫ ∼ N(0, Ψ) . . . , σ2
n)),
(2)
- r
x|z ∼ N(Az, Ψ) (3) where Im denotes an m × m identity matrix, A is a d × m matrix that maps the latent space to the observation space, and the noise covariance is the diagonal matrix Ψ = diag(σ2
1, . . . , σ2 d). The
model parameters are θ = {A, Ψ}. The marginal likelihood is also Gaussian: x ∼ N(0, AA⊤ + Ψ). (4) Thus FA is basically a model of the covariance of the data, and seeks to represent it as the sum of a low-rank component and a diagonal component.
2 Recognition distribution
The recognition distribution follows from Bayes’ rule: p(z|x) ∝ p(x|z)p(z) = N(x|Az, Ψ) · N(z|0, I) (5) we can ”complete the square” to find the Gaussian distribution in z that results from normalizing the right-hand-side: p(z|x) = N
- (AΨ−1A⊤ + I)−1A⊤Ψ−1x, (A⊤Ψ−1A + I)−1