Recitation 10/8 Mixture Models, PCA Slides borrowed from Prof. - - PowerPoint PPT Presentation

recitation 10 8
SMART_READER_LITE
LIVE PREVIEW

Recitation 10/8 Mixture Models, PCA Slides borrowed from Prof. - - PowerPoint PPT Presentation

Recitation 10/8 Mixture Models, PCA Slides borrowed from Prof. Seyoung Kim, Ryan Tibshirani. Thanks! Law of Total Probability Completely Observed Data Bishop Page 431 Since z uses a 1-of-K representation, we have What if we do not know


slide-1
SLIDE 1

Recitation 10/8

Mixture Models, PCA

Slides borrowed from Prof. Seyoung Kim, Ryan Tibshirani. Thanks!

slide-2
SLIDE 2

Law of Total Probability

slide-3
SLIDE 3

Since z uses a 1-of-K representation, we have

Completely Observed Data

Bishop Page 431

slide-4
SLIDE 4
slide-5
SLIDE 5

What if we do not know ?

slide-6
SLIDE 6
slide-7
SLIDE 7
slide-8
SLIDE 8

Example 2-d data points coming from K = 2 Gaussian distributions

K=2 1-d Gaussian distributions: <x, y> pairs

slide-9
SLIDE 9

Example 2-d data points coming from K = 2 Gaussian distributions

K=2 1-d Gaussian distributions: <x, y> pairs

Initialize

slide-10
SLIDE 10

Example 2-d data points coming from K = 2 Gaussian distributions

Initialize iteration t =1

slide-11
SLIDE 11

Example 2-d data points coming from K = 2 Gaussian distributions

Initialize iteration t =1 2 4 7 0.953 0.047

slide-12
SLIDE 12

Example 2-d data points coming from K = 2 Gaussian distributions

Initialize iteration t =1 2 4 7 0.953 0.047

slide-13
SLIDE 13

PCA

Principal components are a sequence of projections of the data, mutually uncorrelated and ordered in variance.

slide-14
SLIDE 14
slide-15
SLIDE 15

Assume X is a normalized Nxp data matrix for N samples and p features Assume data is normalized. ⇐> each column of X is normalized. Variance of projected data where S = <- Want to maximize this over v

slide-16
SLIDE 16
slide-17
SLIDE 17

The proportion of variance explained is a nice way to quantify how much structure is being captured