Principal Component Analysis
Ken Kreutz-Delgado (Nuno Vasconcelos)
UCSD — ECE 175A — Winter 2012
Principal Component Analysis Ken Kreutz-Delgado (Nuno Vasconcelos) - - PowerPoint PPT Presentation
Principal Component Analysis Ken Kreutz-Delgado (Nuno Vasconcelos) UCSD ECE 175A Winter 2012 Curse of dimensionality Typical observation in Bayes decision theory: Error increases when number of features is large Even for simple
UCSD — ECE 175A — Winter 2012
2
axis in order to get a reasonable quantization
dimension 1 2 3 points/bin 10 1 0.1
3
maintain classification accuracy increases exponentially with the dimension of the feature space
4
1.features are not discriminant 2.features are not independent (are highly correlated)
discriminant non-discriminant
5
car loan
projection onto 1D subspace: y = a x
car loan new feature y
6
look very flat when viewed from the full space, e.g.
are going to be highly skewed ellipsoids
data give the Principle Components of the data.
1D subspace in 2D 2D subspace in 3D
7
measures the “natural” units for the problem because it is “adapted” to the covariance of the data
is that it uses S-1
1 2( , )
T
8
9
s1 s2 z2 z1
10
s1 s2 z2 z1 y = F z s1 s2 y2 y1
f1 f2
11
the eigenvectors of S
square roots of the eigenvalues li of S
s1 s2 y2 y1
f1 f2
s1 s2 y2 y1 s1 s2 y2 y1
12
13
14
followed by a flat area
15
along the principle directions (eigenvectors)
contained in the top k eigenvalues
along the associated eigenvectors s1 s2 z2 z1 l1 l2 y2 y1
f1 f2
y = F z
2 1 2 1 k i i k n i i
16
this dataset
17
singular vectors (columns of M)
singular values (which are nonzero and strictly positive)
(columns of N = rows of NT)
T
T T T m m m m
18
1
n
1 1
n i n i
19
1
c n T T T
20
c is the i th column of Xc
1 1
c c c T n c c c n
T T c c i i i i i i
21
1 c T c c n
T T c
T T
2
T T T T c c
22
2 2 i i i
2
T
23
T T T c
T T c
i i i
24
Principal components are
informative about the structure of the data Example:
the principal components for the space of images of faces
show the first 16 eigenvectors (eigenfaces)
structure, etc
25
26
average sound images Eigensounds corresponding to the three highest eigenvalues
27
Flames Eigenflames
28
Eigenrings reconstruction
29
directions of semantic concepts
terms documents = documents concepts concepts terms x x
30
same idea.
resources
anecdotal instances of terms are to be eliminated.
term-document matrix. E.g. lists only words actually in each document, whereas we might be interested in all words related to each document-- much larger set due to synonymy
31
meanings.
direction are added to the components of words that share this sense.
either simply cancel out, or, at worst, to be smaller than components in the directions corresponding to the intended sense.
32
through dot-products only, i.e. the matrix of elements can be kernelized
the inner product form mentioned above
T i j
1
n
33
TXc, and for their
c
T T c
2 1 2
T T c c c
34
TXc
2, )
1
/ n) P
2
T
2
T T T c c
1
c
35
1 1
| | | |
c T c c c c c n c n T c c n n
x K X X x x x x x
36
TXc
2, )
1
/ n ) P
2
components is given by
1 1
T T c c c c
37