Diffusion Maps and Coarse-Graining: A unified framework for - - PowerPoint PPT Presentation

diffusion maps and coarse graining
SMART_READER_LITE
LIVE PREVIEW

Diffusion Maps and Coarse-Graining: A unified framework for - - PowerPoint PPT Presentation

Diffusion Maps and Coarse-Graining: A unified framework for dimensionality reduction, graph partitioning and data set parameterization Authors: Stephane Lafon and Ann B. Lee (PAMI, Sept. 2006) Presented by Shihao Ji Duke University Machine


slide-1
SLIDE 1

Diffusion Maps and Coarse-Graining:

A unified framework for dimensionality reduction, graph partitioning and data set parameterization

Presented by Shihao Ji Duke University Machine Learning Group

  • Sept. 28, 2006

Authors: Stephane Lafon and Ann B. Lee (PAMI, Sept. 2006)

slide-2
SLIDE 2
  • Diffusion distances and Maps
  • Graph partitioning and subsampling
  • Numerical examples

Outline

slide-3
SLIDE 3

Diffusion distances

  • Let be a finite graph with n nodes, and

weight matrix W satisfies the following conditions: – symmetry: – positivity: i.e. Gaussian kernel

  • Markov random walk
  • t step random walk
slide-4
SLIDE 4

Diffusion distances (cont’d)

  • Definition

where is the unique stationary distribution of P.

slide-5
SLIDE 5

Diffusion Maps

  • The transition matrix P is adjoint to a symmetric matrix

thus, P and Ps share the same eigenvalues.

  • Since Ps is a symmetric matrix

eigenvalues: eigenvectors: form the orthonormal basis.

slide-6
SLIDE 6

Diffusion Maps (cont’d)

  • The left and right eigenvectors of P:
  • Biorthogonal spectral decomposition
  • Diffusion distances
slide-7
SLIDE 7

Diffusion Maps (cont’d)

  • Diffusion Maps
  • Diffusion distances
slide-8
SLIDE 8

Graph partitioning

  • Consider an arbitrary partition
slide-9
SLIDE 9

Graph partitioning (cont’d)

  • Definition (geometric centroid):
  • Theorem: for , we have

where and

slide-10
SLIDE 10

Graph partitioning (Cont’d)

  • This theorem tells us

1. If then and are approximate left and right eigenvectors of with approximate eigenvalue . 2. In order to maximize the quality of approximation, we need to minimize the following distortion in diffusion space:

  • This provides a rigorous justification for k-means

clustering in diffusion space.

slide-11
SLIDE 11

Numerical examples

  • Diffusion distance vs. Euclidean distance

The Swiss roll, and its quantization by k-means (k=4)

slide-12
SLIDE 12

Numerical examples (cont’d)

  • Robustness of the diffusion distance

Dijkstra’s algorithm Diffusion distance

slide-13
SLIDE 13

Numerical examples (cont’d)

Averaged on 1000 instances

slide-14
SLIDE 14

Messages

  • Diffusion maps provide a unified framework for

dimensionality reduction, graph partitioning and data set parameterization.

  • Coarse-graining gives a rigorous justification of k-

means clustering in diffusion space.

  • Diffusion distance is robust to noise and small

perturbations of the data.