1
play

1 ! k c a B Signal Processing on Graphs s e k i r t S g - PowerPoint PPT Presentation

1 ! k c a B Signal Processing on Graphs s e k i r t S g SpaRTaN-MacSeNet Spring School n i r e t l i F s s a P Pierre Vandergheynst - w o L Swiss Federal Institute of Technology April is Autism Awareness Month:


  1. 1

  2. ! k c a B Signal Processing on Graphs s e k i r t S g SpaRTaN-MacSeNet Spring School n i r e t l i F s s a P Pierre Vandergheynst - w o L Swiss Federal Institute of Technology April is Autism Awareness Month: https://www.autismspeaks.org/wordpress-tags/autism-awareness-month

  3. 3 Signal Processing on Graphs Social Networks Energy Networks Transportation Networks Biological Networks Irregular Data Domains

  4. 4

  5. 5 Some Typical Processing Problems Compression / Visualization Many interesting new contributions with a SP perspective Denoising [Coifman, Maggioni, Kolaczyk, Ortega, Ramchandran, Moura, Lu, Borgnat] or IP perspective [ElMoataz, Lezoray] Earth data source: Frederik Simons Semi-Supervised Learning See review in 2013 IEEE SP Mag Analysis / Information Extraction

  6. 6 Outline l Introduction: - Graphs and elements of spectral graph theory, with emphasis on functional calculs l Kernel Convolution: - Localization, filtering, smoothing and applications l An application to spectral clustering that unifies some of the themes you’ve heard of during the workshop: machine learning, compressive sensing, optimisation algorithms, graphs

  7. 7 Elements of Spectral Graph Theory Reference: F. Chung, Spectral Graph Theory

  8. 8 Definitions A graph G is given by a set of vertices and « relationships » between them encoded in edges G = ( V , E ) A set V of vertices of cardinality | V | = N A set E of edges: e ∈ E, e = ( u, v ) with u, v ∈ V e 0 = ( v, u ) and e 6 = e 0 Directed edge: e = ( u, v ) , e 0 = ( v, u ) and e = e 0 Undirected edge: e = ( u, v ) , A graph is undirected if it contains only undirected edges A weighted graph has an associated non-negative weight function: w : V × V → R + ( u, v ) / ∈ E ⇒ w ( u, v ) = 0

  9. 9 Matrix Formulation Connectivity captured via the (weighted) adjacency matrix W ( u, v ) = w ( u, v ) with obvious restriction for unweighted graphs W ( u, u ) = 0 no loops Let d ( u ) be the degree of u and D = diag(d) the degree matrix Graph Laplacians, Signals on Graphs L norm = D − 1 / 2 L D − 1 / 2 L = D − W Graph signal: f : V → R Laplacian as an operator on space of graph signals X � � L f ( u ) = w ( u, v ) f ( u ) − f ( v ) v ∼ u

  10. 10 Some di ff erential operators The Laplacian can be factorized as L = SS ∗ Explicit form of the incidence matrix (unweighted in this example): e =( u , v ) ( ) -1 u S = 1 v is a gradient S ∗ f ( u, v ) = f ( v ) − f ( u ) X X is a negative divergence g ( v 0 , u ) S g ( u ) = g ( u, v ) − ( u,v ) 2 E ( v 0 ,u ) 2 E

  11. 11 Properties of the Laplacian Laplacian is symmetric and has real eigenvalues � 2 � 0 X Dirichlet form Moreover: � h f, L f i = w ( u, v ) f ( u ) � f ( v ) u ∼ v positive semi-definite, non-negative eigenvalues Spectrum: 0 = λ 0 ≤ λ 1 ≤ . . . λ max G connected: λ 1 > 0 G has i +1 connected components λ i = 0 and λ i +1 > 0 Notation: h f, L g i = f t L g

  12. 12 Measuring Smoothness � 2 � 0 X � h f, L f i = f ( u ) � f ( v ) u ∼ v is a measure of « how smooth » f is on G Using our definition of gradient: r u f = { S ∗ f ( u, v ) , 8 v ⇠ u } sX Local variation kr u f k 2 = | S ∗ f ( u, v ) | 2 v ∼ u sX X X Total variation | f | T V = kr u f k 2 = | S ∗ f ( u, v ) | 2 v ∼ u u ∈ V u ∈ V

  13. 13 Notions of Global Regularity for Graph Discrete Calculus , Grady and Polimeni, 2010 � ∂ f Edge p � := w ( m, n ) [ f ( n ) − f ( m )] Derivative � ∂ e � m "⇢ ∂ f # � � Graph � O m f := Gradient � ∂ e � e ∈ E s.t. e =( m,n ) m # 1 " X 2 Local w ( m, n ) [ f ( n ) − f ( m )] 2 || O m f || 2 = Variation n ∈ N m 1 w ( m, n ) [ f ( n ) − f ( m )] 2 = f Quadratic X || O m f || 2 X 2 = T L f Form 2 m ∈ V ( m,n ) ∈ E

  14. 14 Smoothness of Graph Signals G 2 G 3 G 1 T L 3 f = 1 . 81 T L 1 f = 0 . 14 T L 2 f = 1 . 31 f f f ˆ ˆ ˆ ( ) ( ) ( ) f λ  f λ  f λ  λ  λ  λ 

  15. 15 Remark on Discrete Calculus Discrete operators on graphs form the basis of an interesting field aiming at bringing a PDE-like framework for computational analysis on graphs: l Leo Grady: Discrete Calculus l Olivier Lezoray, Abderrahim Elmoataz and co-workers: PDEs on graphs: - many methods from PDEs in image processing can be 
 transposed on arbitrary graphs - applications in vision (point clouds) but also machine learning 
 (inference with graph total variation)

  16. 16 Laplacian eigenvectors Spectral Theorem: Laplacian is PSD with eigen decomposition { ( λ ` , u ` ) } ` =0 , 1 ,...,N − 1 L = D − W L = U Λ U t That particular basis will play the role of the Fourier basis: Graph Fourier Transform, Coherence N ˆ X f ( i ) u ∗ f ( λ ` ) := h f , u ` i = ` ( i ) , i =1 h 1 h Graph Coherence p ` ,i | h u ` , δ i i | 2 µ := max , 1 N

  17. 17 Important remark on eigenvectors h 1 h p ` ,i | h u ` , δ i i | 2 µ := max , 1 N Optimal - Fourier case What does that mean ?? Eigenvectors of modified path graph

  18. 18 Examples: Cut and Clustering RatioCut( A, A ) := 1 C ( A, A ) + 1 C ( A, A ) X C ( A, B ) := W [ i, j ] 2 | A | 2 | A | i ∈ A,j ∈ B 8 q | A | / | A | if i ∈ A < f [ i ] = A ⊂ V RatioCut( A, A ) min q | A | / | A | if i ∈ A − : p k f k = | V | and h f, 1 i = 0 f t L f = | V | · RatioCut( A, A ) p f ∈ R | V | f t L f subject to k f k = | V | and h f, 1 i = 0 arg min Relaxed problem Looking for a smooth partition function

  19. 19

  20. 20 Examples: Cut and Clustering Spectral Clustering p f ∈ R | V | f t L f subject to k f k = | V | and h f, 1 i = 0 arg min By Rayleigh-Ritz, solution is second eigenvector u 1 Remarks: Natural extension to more than 2 sets Solution is real-valued and needs to be quantized. In general, k-MEANS is used. First k eigenvectors of sparse Laplacians via Lanczos, complexity driven by eigengap | λ k − λ k +1 | Spectral clustering := embedding + k-MEANS � � 8 i 2 V : i 7! u 0 ( i ) , . . . , u k − 1 ( i )

  21. 21 Graph Embedding/Laplacian Eigenmaps Goal: embed vertices in low dimensional space, discovering geometry ( x 1 , . . . x N ) 7! ( y 1 , . . . y N ) k < d x i ∈ R d y i ∈ R k Good embedding: nearby points mapped nearby, so smooth map y i = Φ ( x i )

  22. 22 Graph Embedding/Laplacian Eigenmaps Goal: embed vertices in low dimensional space, discovering geometry ( x 1 , . . . x N ) 7! ( y 1 , . . . y N ) k < d x i ∈ R d y i ∈ R k Good embedding: nearby points mapped nearby, so smooth map minimize variations/ X W [ i, j ]( y i − y j ) 2 maximize smoothness of embedding i,j Laplacian Eigenmaps y t L y arg min y L y = λ Dy y t Dy = 1 fix scale y t D 1 = 0

  23. 23 Laplacian Eigenmaps [Belkin, Niyogi, 2003]

  24. 24 Remark on Smoothness Linear / Sobolev case Smoothness, loosely defined, has been used to motivate various methods and algorithms. But in the discrete, finite dimensional case, asymptotic decay does not mean much f ( ` ) | 2 ≤ M � ` | ˆ X kr f k 2 2  M , f t L f  M ⇔ ` E K ( f )  kr f k 2 E K ( f ) = k f � P K ( f ) k 2 p λ K +1 √ M | ˆ f ( ` ) | ≤ √ � `

  25. 25 Smoothness of Graph Signals Revisited G 2 G 3 G 1 ˆ ˆ ˆ ( ) ( ) ( ) f λ  f λ  f λ  λ  λ  λ  T L 3 f = 1 . 81 T L 1 f = 0 . 14 T L 2 f = 1 . 31 f f f

  26. 26 Functional calculus It will be useful to manipulate functions of the Laplacian f ( L ) , f : R 7! R L k u ` = λ k polynomials ` u ` Symmetric matrices admit a (Borel) functional calculus Borel functional calculus for symmetric matrices X f ( λ ` ) u ` u t f ( L ) = ` ` ∈ S ( L ) Use spectral theorem on powers, get to polynomials From polynomial to continuous functions by Stone-Weierstrass Then Riesz-Markov (non-trivial !)

  27. 27 Example: Di ff usion on Graphs Consider the following « heat » di ff usion model ∂ f @ f ( ` , t ) = − � ` ˆ ˆ f ( ` , 0) := ˆ ˆ f ( ` , t ) f 0 ( ` ) ∂ t = − L f @ t f ( ` , t ) = e − t λ ` ˆ ˆ by functional calculus f = e − t L f 0 f 0 ( ` ) X X e − t � ` u ` ( i ) u ` ( j ) f 0 ( j ) f ( i ) = Explicitly: j ∈ V ` X e − t L = e − t � ` u ` u t X X e − t � ` u ` ( i ) = u ` ( j ) f 0 ( j ) ` ` ` j ∈ V X e − t L [ i, j ] = e − t � ` u ` ( i ) u ` ( j ) e − t � ` ˆ X = f 0 ( ` ) u ` ( i ) ` `

  28. 28 Example: Di ff usion on Graphs examples of heat kernel on graph f 0 ( j ) = δ k ( j ) e − t � ` ˆ X f ( i ) = f 0 ( ` ) u ` ( i ) ` X e − t � ` u ` ( k ) u ` ( i ) = `

  29. 29 Simple De-Noising Example Suppose a smooth signal f on a graph 1 kr f k 2 2  M , f t L f  M 0.8 0.6 √ 0.4 M | ˆ 0.2 f ( ` ) | ≤ √ � ` 0 − 0.2 − 0.4 − 0.6 − 0.8 − 1 Noisy Original 2 But you observe only a noisy version y 1.5 1 0.5 0 y ( i ) = f ( i ) + n ( i ) − 0.5 − 1 − 1.5 − 2

  30. 30 Simple De-Noising Example De-Noising by Regularization 2 s.t. f t L f  M k f � y k 2 argmin f L r f ∗ + τ τ 2 � f − y � 2 2( f ∗ − y ) = 0 T L r f. argmin 2 + f f Graph Fourier � � � � L r f ∗ ( ℓ ) + τ f ∗ ( ℓ ) − ˆ y ( ℓ ) = 0 , 2 ∀ ℓ ∈ { 0 , 1 , . . . , N − 1 } τ � “Low pass” filtering ! f ∗ ( ℓ ) = y ( ℓ ) ˆ τ + 2 λ r � Convolution with a kernel: ˆ f ( ` )ˆ g ( � ` ; ⌧ , r ) ⇒ g ( L ; ⌧ , r )

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend