Advanced Introduction to Machine Learning, CMU-10715 Manifold - - PowerPoint PPT Presentation

advanced introduction to machine learning cmu 10715
SMART_READER_LITE
LIVE PREVIEW

Advanced Introduction to Machine Learning, CMU-10715 Manifold - - PowerPoint PPT Presentation

Advanced Introduction to Machine Learning, CMU-10715 Manifold Learning Barnabs Pczos Motivation Find meaningful low-dimensional structures hidden in high-dimensional observations. The human brain confronts the same problem in


slide-1
SLIDE 1

Advanced Introduction to Machine Learning, CMU-10715

Manifold Learning

Barnabás Póczos

slide-2
SLIDE 2

Motivation

2

The human brain confronts the same problem in perception: 30,000 auditory nerve fibers 106 optic nerve fibers ) extract small number of perceptually relevant features.

  • Find meaningful low-dimensional structures

hidden in high-dimensional observations.

  • Difficult to visualize data in dimensions greater

than three.

slide-3
SLIDE 3

Manifolds

Informal definition: Manifold = any object which is nearly "flat" on small scales. 1dim manifolds: 2dim manifolds:

3

slide-4
SLIDE 4

X1

R2

X2

x

z

Manifold Learning

4

slide-5
SLIDE 5
  • PCA (1901), kernel PCA
  • Multi-dimensional Scaling (1952)
  • Maximum Variance Unfolding, Colored MVU
  • Mixture of PCA, Mixture of Factor Analyzers
  • Local Linear Embedding (2000)
  • Isomap (2000), C-Isomap
  • Hessian Eigenmaps
  • Laplacian Eigenmaps (2003)
  • Local Tangent Space Alignment
  • … and many more

Algorithms

5

slide-6
SLIDE 6

PCA

PCA is a linear method: it fails to find the nonlinear structure in the data

6

slide-7
SLIDE 7

Nonlinear Manifolds..

A

Unroll the manifold

PCA uses the Euclidean distance What is important is the geodesic distance

Issues with PCA

7

slide-8
SLIDE 8

Multi-dimensional Scaling

8

slide-9
SLIDE 9

MultiDimensional Scaling..

 In PCA we are given a set of points  In MDS we are given pairwise distances instead

  • f the actual data points.

 Question: If we only preserve the pairwise distances do we preserve the structure?

Multi-dimensional Scaling

9

slide-10
SLIDE 10

How to get dot product matrix from pairwise distance matrix?

k ij

d

kj

d

ki

d 

j i

From Distances to Inner Products

10

slide-11
SLIDE 11

MDS..

Similarly: Center the data and then calculate

11

From Distances to Inner Products

MDS cost function:

slide-12
SLIDE 12

MDS..

12

From Distances to Inner Products

MDS algorithm:

Step 1: Build a Gram matrix of inner products Step 2: Find the top k eigenvectors of G with the top k eigenvalues: Step 3:

slide-13
SLIDE 13

Though based on a somewhat different geometric intuition, metric MDS is closely related to PCA. There are many different versions of MDS…

13

Metric MDS = PCA

Observation: If the data is centered, then the Gram matrix can be found this way:

slide-14
SLIDE 14

Example of MDS…

14

slide-15
SLIDE 15
slide-16
SLIDE 16

Isomap

16

A Global Geometric Framework for Nonlinear Dimensionality Reduction

  • J. B. Tenenbaum, V. de Silva and J. C. Langford

Science 290 (5500): 2319-2323, 22 December 2000

slide-17
SLIDE 17

ISOMAP

Comes from Isometric feature mapping Step1: Take a data matrix as input. Step2: Estimate geodesic distance between any two points by

“a chain of short paths”. Approximate the geodesic distance by Euclidean distances.

Step3: Perform MDS

17

slide-18
SLIDE 18

Geodesic: the shortest curve on a manifold that connects two points on the manifold

small circle great circle

18

Differential Geometry

Example (3D sphere)

slide-19
SLIDE 19

Euclidean distance needs not be a good measure between two points on a manifold Length of geodesic is more appropriate

19

Geodesic distance

slide-20
SLIDE 20

The Swiss-role Dataset

20

slide-21
SLIDE 21

Isomap

21

slide-22
SLIDE 22

22

slide-23
SLIDE 23

23

slide-24
SLIDE 24

ISOMAP Interpolation

24

slide-25
SLIDE 25

ISOMAP Interpolation

25

slide-26
SLIDE 26

ISOMAP Interpolation

26

slide-27
SLIDE 27

ISOMAP Summary

 Build graph from kNN or epsilon neighbors  Run MDS  Since MDS is slow, ISOMAP will be very slow.  Need estimate of k or epsilon.  Assumes data set is convex (no holes).

27

slide-28
SLIDE 28

Local Linear Embedding

28

Nonlinear dimensionality reduction by locally linear embedding. Sam Roweis & Lawrence Saul. Science, v.290 no 5500, Dec.22, 2000. pp.2323-- 2326.

slide-29
SLIDE 29

Local Linear Embedding

  • 1. select neighbors

(epsilon or kNN)

  • 2. reconstruct with linear weights

Assumption: manifold is approximately “linear” when viewed locally. Data:

29

slide-30
SLIDE 30

Local Linear Embedding

 Step 1.

Without the constraints the weights that minimize the reconstruction errors are invariant to rotation, rescaling and translation of the data points.

30

slide-31
SLIDE 31

Local Linear Embedding

 Step 2. Given the weights W, find the embedded points:

The same weights that reconstruct the data points in D dimensions should also reconstruct the points in d dimensions.

The weights characterize the intrinsic geometric properties

  • f each neighborhood.

31

slide-32
SLIDE 32

Locally Linear Embedding

Fit Locally , Think Globally

32

slide-33
SLIDE 33

33

slide-34
SLIDE 34

34

slide-35
SLIDE 35

35

slide-36
SLIDE 36

Maximum Variance Unfolding

36

  • K. Q. Weinberger and L. K. Saul.

Unsupervised learning of image manifolds by semidefinite programming. International Journal of Computer Vision, Volume 70 Issue 1, October 2006, Pages 77 - 90

slide-37
SLIDE 37

Maximum Variance Unfolding

37

Formally,

Build a graph from kNN or epsilon neighbors.

slide-38
SLIDE 38

Maximum Variance Unfolding

38

Consider the constraint From this, we have

slide-39
SLIDE 39

Maximum Variance Unfolding

39

Consider the cost function:

slide-40
SLIDE 40

Maximum Variance Unfolding

40

The final problem is a semi-definite problem (SDP) :

slide-41
SLIDE 41

Maximum Variance Unfolding

D= 76*101*3 d=3 N=400 images

41

slide-42
SLIDE 42

Maximum Variance Unfolding

Swiss roll “unfolded” by maximizing variance subject to constraints that preserve local distances and angles. The middle snap-shots show various feasible (but non-optimal) intermediate solutions.

42

slide-43
SLIDE 43

Maximum Variance Unfolding

43

slide-44
SLIDE 44

Laplacian Eigenmap

44

  • M. Belkin and P. Niyogi. “Laplacian eigenmaps for dimensionality

reduction and data representation,” Neural Comput.,15(6):1373–1396, 2003.

slide-45
SLIDE 45

45

Laplacian Eigenmap

Step 1. Build graph from kNN or epsilon neighbors Step 2. Choose weights: Special case:

slide-46
SLIDE 46

46

Laplacian Eigenmap

Step 3. Assume the graph is connected, otherwise proceed with Step 3 for each connected component: Lemma: Solve the eigenvector problem:

slide-47
SLIDE 47

47

Laplacian Eigenmap

Solve the eigenvector problem: The first m+1 smallest eigenvalues: The embedding:

slide-48
SLIDE 48

48

Laplacian Eigenmap (Explanation)

Let us embed the neighborhood graph to 1 dim first. A reasonable cost function is: subject to appropriate constraints to avoid y=0.

Lemma

Proof:

slide-49
SLIDE 49

49

Laplacian Eigenmap (Explanation)

Therefore, our minimization problem is Subject to:

Embedding the neighborhood graph to m dimension:

Subject to:

Solution:

slide-50
SLIDE 50

Variational Variational Inference for Bayesian Mixtures of Factor Analysers

50

Zoubin Ghahramani, Matthew J. Beal, NIPS 1999

slide-51
SLIDE 51

MANI Matlab demo

MANIfold learning demonstration GUI Contains a couple of methods and examples.

http://www.math.ucla.edu/~wittman/mani

Todd Wittman:

The following results are taken from Todd Wittman

51

slide-52
SLIDE 52

How do we compare the methods?

 Speed  Manifold Geometry  Non-convexity  Curvature  Corners  High-Dimensional Data: Can the method process image manifolds?  Sensitivity to Parameters

  • K Nearest Neighbors: Isomap, LLE, Hessian,

Laplacian, KNN Diffusion

  • Sigma: Diffusion Map, KNN Diffusion

 Noise  Non-uniform Sampling  Sparse Data  Clustering

52

slide-53
SLIDE 53

Testing Examples

 Swiss Roll  Swiss Hole  Punctured Sphere  Corner Planes  3D Clusters Twin Peaks Toroidal Helix Gaussian Occluded Disks

We’ll compare the speed and sensitivity to parameters throughout.

53

slide-54
SLIDE 54

Manifold Geometry

First, let’s try to unroll the Swiss Roll. We should see a plane.

54

slide-55
SLIDE 55

Hessian LLE is pretty slow, MDS is very slow, and ISOMAP is extremely slow. MDS and PCA don’t can’t unroll Swiss Roll, use no manifold information. LLE and Laplacian can’t handle this data. Diffusion Maps could not unroll Swiss Roll for any value of Sigma.

55

slide-56
SLIDE 56

Non-Convexity

Can we handle a data set with a hole? Swiss Hole: Can we still unroll the Swiss Roll when it has a hole in the middle?

56

slide-57
SLIDE 57

Only Hessian LLE can handle non-convexity. ISOMAP, LLE, and Laplacian find the hole but the set is distorted. 57

slide-58
SLIDE 58

Manifold Geometry

Twin Peaks: fold up the corners of a plane. LLE will have trouble because it introduces curvature to plane.

58

slide-59
SLIDE 59

PCA, LLE, and Hessian LLE distort the mapping the most.

59

slide-60
SLIDE 60

Curvature & Non-uniform Sampling

Gaussian: We can randomly sample a Gaussian distribution. We increase the curvature by decreasing the standard deviation. Coloring on the z-axis, we should map to concentric circles.

60

slide-61
SLIDE 61

For std = 1 (low curvature), MDS and PCA can project accurately. Laplacian Eigenmap cannot handle the change in sampling.

61

slide-62
SLIDE 62

For std = 0.4 (higher curvature), PCA projects from the side rather than top-down. Laplacian looks even worse.

62

slide-63
SLIDE 63

For std = 0.3 (high curvature), none of the methods can project correctly.

63

slide-64
SLIDE 64

Corners

Corner Planes: We bend a plane with a lift angle A. We want to bend it back down to a plane. If A > 90, we might see the data points written on top of each other.

A

64

slide-65
SLIDE 65

For angle A=75, we see some disortions in PCA and Laplacian.

65

slide-66
SLIDE 66

For A = 135, MDS, PCA, and Hessian LLE overwrite the data points. Diffusion Maps work very well for Sigma < 1. LLE handles corners surprisingly well.

66

slide-67
SLIDE 67

Clustering

A good mapping should preserve clusters in the original data set. 3D Clusters: Generate M non-overlapping clusters with random centers. Connect the clusters with a line.

67

slide-68
SLIDE 68

For M = 3 clusters, MDS and PCA can project correctly. Diffusion Maps work well with large Sigma. LLE compresses each cluster into a single point. Hessian LLE has trouble with the sparse connecting lines.

68

slide-69
SLIDE 69

For M=8 clusters, MDS and PCA can still recover. Diffusion Maps do quite well. LLE and ISOMAP are decent, but Hessian and Laplacian fail.

69

slide-70
SLIDE 70

Noise & Non-uniform Sampling

Can the method handle changes from dense to sparse regions? Toroidal Helix should be unraveled into a circle parametrized by t. We can change the sampling rate along the helix by changing the exponent R on the parameter t and we can add some noise.

70

slide-71
SLIDE 71

With no noise added, ISOMAP, LLE, Laplacian, and Diffusion Map are correct. MDS and PCA project to an asterisk. What’s up with Hessian and KNN Diffusion?

71

slide-72
SLIDE 72

Adde noise to the Helix sampling. LLE cannot recover the circle. ISOMAP emphasizes outliers more than the other methods.

72

slide-73
SLIDE 73

When the sampling rate is changed along the torus, Laplacian starts to mess up and Hessian is completely thrown off. Hessian LLE code crashed frequently on this example. Diffusion maps handle it quite well for carefully chosen Sigma=0.3.

73

slide-74
SLIDE 74

Sparse Data & Non-uniform Sampling

Of course, we want as much data as possible. But can the method handle sparse regions in the data? Punctured Sphere: the sampling is very sparse at the bottom and dense at the top.

74

slide-75
SLIDE 75

Only LLE and Laplacian get decent results. PCA projects the sphere from the side. MDS turns it inside-out. Hessian and Diffusion Maps get correct shape, but give too much emphasis to the sparse region at the bottom of the sphere.

75

slide-76
SLIDE 76

High-Dimensional Data

All of the examples so far have been 3D. But can the data handle high-dimensional data sets, like images? Disks: Create 20x20 images with a disk of fixed radius and random center. We should recover the centers of the circles.

76

slide-77
SLIDE 77

??? LLE Crashed

LLE crashed on high-dimensional data set. Number of images was not high enough, but ISOMAP did a very good job.

77

slide-78
SLIDE 78

Occluded Disks

We can add a second disk of radius R in the center of every image.

78

slide-79
SLIDE 79

??? LLE Crashed Hessian Crashed

Both LLE and Hessian crashed, possibly # points is too small. Laplacian failed completely. Is ISOMAP the best for high-dimensional data?

79

slide-80
SLIDE 80

Sensitivity to Parameters

When the number of points is small or the data geometry is complex, it is important to set K appropriately, neither too big nor small. But if the data set is dense enough, we expect K around 8 or 10 to suffice. For Diffusion Maps, the method is very sensitive to the Sigma in Gaussian kernel. Varies from example to example.

80

slide-81
SLIDE 81

X Diffusion Map Sigma depends on manifold. Helix Clusters Sigma = 0.2 Sigma = 10

81

slide-82
SLIDE 82

So what have you learned, Dorothy?

MDS PCA ISOMAP LLE Hessian Laplacian Diffusion Map KNN Diffusion Speed Very slow Extremely fast Extremely slow Fast Slow Fast Fast Fast Infers geometry? NO NO YES YES YES YES MAYBE MAYBE Handles non- convex? NO NO NO

MAYBE

YES MAYBE MAYBE MAYBE Handles non- uniform sampling? YES YES YES YES MAYBE NO YES YES Handles curvature? NO NO YES

MAYBE

YES YES YES YES Handles corners? NO NO YES YES NO YES YES YES Clusters? YES YES YES YES NO NO YES YES Handles noise? YES YES MAYBE NO YES YES YES YES Handles sparsity? YES YES YES YES NO

may crash

YES NO NO Sensitive to parameters? NO NO YES YES YES YES VERY VERY

82

slide-83
SLIDE 83

Some Notes on using MANI

 Hard to set K and Sigma just right.  MDS and ISOMAP are very slow.  Hessian LLE is pretty slow. Since Hessian needs a dense data set, this means it takes even longer when the # points is large.  Occluded Disks is 400-dimensional data, which takes a long time and a lot of data points to correctly map.  Matlab GUIs seem to run better on PC than Linux.

83

slide-84
SLIDE 84

84

Credits

  • M. Belkin,
  • P. Niyogi,

Todd Wittman

slide-85
SLIDE 85

Thanks for your attention 

85