identifying undirected network structure via semidefinite
play

Identifying Undirected Network Structure via Semidefinite Relaxation - PowerPoint PPT Presentation

Identifying Undirected Network Structure via Semidefinite Relaxation Rasoul Shafipour, Santiago Segarra , Antonio G. Marques and Gonzalo Mateos Institute for Data, Systems, and Society Massachusetts Institute of Technology segarra@mit.edu


  1. Identifying Undirected Network Structure via Semidefinite Relaxation Rasoul Shafipour, Santiago Segarra , Antonio G. Marques and Gonzalo Mateos Institute for Data, Systems, and Society Massachusetts Institute of Technology segarra@mit.edu http://www.mit.edu/~segarra/ ICASSP, April 20, 2018 Santiago Segarra 1 / 18

  2. Network Science analytics Online social media Internet Clean energy and grid analy,cs ◮ Desiderata: Process, analyze and learn from network data [Kolaczyk09] Santiago Segarra 2 / 18

  3. Network Science analytics Online social media Internet Clean energy and grid analy,cs ◮ Desiderata: Process, analyze and learn from network data [Kolaczyk09] ◮ Network as graph G : encode pairwise relationships ◮ Sometimes both G and data at the nodes are available ⇒ Leverage G to process network data ⇒ Graph Signal Processing Santiago Segarra 2 / 18

  4. Network Science analytics Online social media Internet Clean energy and grid analy,cs ◮ Desiderata: Process, analyze and learn from network data [Kolaczyk09] ◮ Network as graph G : encode pairwise relationships ◮ Sometimes both G and data at the nodes are available ⇒ Leverage G to process network data ⇒ Graph Signal Processing ◮ Sometimes we have access to network data but not to G itself ⇒ Leverage the relation between them to infer G from the data Santiago Segarra 2 / 18

  5. Graph signal processing (GSP) x 2 x 4 ◮ Undirected G with adjacency matrix A 2 4 ⇒ A ij = Proximity between i and j x 1 1 ◮ Define a signal x on top of the graph ⇒ x i = Signal value at node i 3 5 x 3 x 5 Santiago Segarra 3 / 18

  6. Graph signal processing (GSP) x 2 x 4 ◮ Undirected G with adjacency matrix A 2 4 ⇒ A ij = Proximity between i and j x 1 1 ◮ Define a signal x on top of the graph ⇒ x i = Signal value at node i 3 5 x 3 x 5 ◮ Associated with G is the graph-shift operator S = VΛV T ∈ M N ⇒ S ij = 0 for i � = j and ( i , j ) �∈ E (local structure in G ) ⇒ Ex: adjacency A and Laplacian L = D − A matrices Santiago Segarra 3 / 18

  7. Graph signal processing (GSP) x 2 x 4 ◮ Undirected G with adjacency matrix A 2 4 ⇒ A ij = Proximity between i and j x 1 1 ◮ Define a signal x on top of the graph ⇒ x i = Signal value at node i 3 5 x 3 x 5 ◮ Associated with G is the graph-shift operator S = VΛV T ∈ M N ⇒ S ij = 0 for i � = j and ( i , j ) �∈ E (local structure in G ) ⇒ Ex: adjacency A and Laplacian L = D − A matrices ◮ Graph filters H : R N → R N are maps between graph signals ⇒ Polynomial in S with coefficients h ∈ R L +1 ⇒ H := � L l =0 h l S l ◮ How to use GSP to infer the graph topology? Santiago Segarra 3 / 18

  8. Topology inference: Motivation and context ◮ Network topology inference from nodal observations [Kolaczyk09] ◮ Partial correlations and conditional dependence [Dempster74] ◮ Sparsity [Friedman07] and consistency [Meinshausen06] ◮ [Banerjee08], [Lake10], [Slawski15], [Karanikolas16] ◮ Key in neuroscience [Sporns10] ⇒ Functional net inferred from activity Santiago Segarra 4 / 18

  9. Topology inference: Motivation and context ◮ Network topology inference from nodal observations [Kolaczyk09] ◮ Partial correlations and conditional dependence [Dempster74] ◮ Sparsity [Friedman07] and consistency [Meinshausen06] ◮ [Banerjee08], [Lake10], [Slawski15], [Karanikolas16] ◮ Key in neuroscience [Sporns10] ⇒ Functional net inferred from activity ◮ Noteworthy GSP-based approaches ◮ Gaussian graphical models [Egilmez16] ◮ Smooth signals [Dong15], [Kalofolias16] ◮ Stationary signals [Pasdeloup15], [Segarra16] ◮ Directed graphs [Mei15], [Shen16] ◮ Low-rank excitation [Wai18] ◮ Our contribution: topology inference from non-stationary graph signals Santiago Segarra 4 / 18

  10. Problem formulation ◮ Underlying graph G with undirected unknown GSO S ◮ Observe signals { y i } K i =1 defined on the unknown graph Setup y 1 y 2 y 3 Santiago Segarra 5 / 18

  11. Problem formulation ◮ Underlying graph G with undirected unknown GSO S ◮ Observe signals { y i } K i =1 defined on the unknown graph Setup y 1 y 2 y 3 Problem statement Given observations { y i } K i =1 , determine the network S knowing that: { y i } K i =1 are outputs of a diffusion process on S . Santiago Segarra 5 / 18

  12. Problem formulation ◮ Consider an arbitrary linear network process on the GSO S ⇒ Every realization corresponds to a different input x i � L � � h l S l y i = x i = Hx i , i = 1 , . . . , K l =0 ◮ Goal : Recover S from the observation of K signals { y i } K i =1 ◮ Additional unknowns ⇒ The degree of the filter L ⇒ The filter coefficients { h l } L l =0 ⇒ The specific inputs x i ; but we know that x i ∼ N ( 0 , C x ) Santiago Segarra 6 / 18

  13. Blueprint of our solution % STEP%1:%Es5mate% { y i } K the%eigenvectors%of% i =1 % S % STEP%2:%Find% ˆ S eigenvalues%via% op5miza5on% A%priori%info%and% desirable%features%% Santiago Segarra 7 / 18

  14. Blueprint of our solution % STEP%1:%Es5mate% { y i } K the%eigenvectors%of% i =1 % ˆ V :%noisy% S % STEP%2:%Find% ˆ S eigenvalues%via% op5miza5on% Sparsity%and% A%priori%info%and% desirable%features%% GSO%feasibility% Santiago Segarra 7 / 18

  15. Step 1: Estimating the eigenvectors of S ◮ y is the output of a local diffusion process on the graph ∞ � N − 1 � � � h l S l y = α 0 ( I − α l S ) x = x := Hx l =1 l =0 ◮ Whenever the input x is white ⇒ graph stationary process on S [Marques17, Girault15, Perraudin17] Santiago Segarra 8 / 18

  16. Step 1: Estimating the eigenvectors of S ◮ y is the output of a local diffusion process on the graph ∞ � N − 1 � � � h l S l y = α 0 ( I − α l S ) x = x := Hx l =1 l =0 ◮ Whenever the input x is white ⇒ graph stationary process on S [Marques17, Girault15, Perraudin17] Stationary case ◮ The covariance C y of y shares V with S C y = H 2 = h 2 1 S 2 + ... 0 I + 2 h 0 h 1 S + h 2 C y ⇒ Diagonalize ⇒ Obtain ˆ i =1 as ˆ ◮ Estimate covariance from { y i } K V Santiago Segarra 8 / 18

  17. Non-stationary graph signals ◮ Q: What if the signal y = Hx is not stationary (i.e., x colored)? ⇒ Matrices S and C y no longer simultaneously diagonalizable since C y = HC x H Santiago Segarra 9 / 18

  18. Non-stationary graph signals ◮ Q: What if the signal y = Hx is not stationary (i.e., x colored)? ⇒ Matrices S and C y no longer simultaneously diagonalizable since C y = HC x H l =0 h l S l diagonalized by the eigenvectors V of S ◮ Key: still H = � L − 1 ⇒ Infer V by estimating the unknown diffusion (graph) filter H ⇒ Step 1 boils down to system identification + eigendecomposition ˆ % H % System% { y i } K ˆ Eigendecomposi5on% V Iden5fica5on% i =1 % % Santiago Segarra 9 / 18

  19. System identification Define C xyx := C 1 / 2 C y C 1 / 2 , with eigenvectors V xyx . If C x is non- x x singular then all admissible symmetric filters H are of the form H = C − 1 / 2 C 1 / 2 xyx V xyx diag( b ) V T xyx C − 1 / 2 , x x where b ∈ {− 1 , 1 } N is a binary (signed) vector. Santiago Segarra 10 / 18

  20. System identification Define C xyx := C 1 / 2 C y C 1 / 2 , with eigenvectors V xyx . If C x is non- x x singular then all admissible symmetric filters H are of the form H = C − 1 / 2 C 1 / 2 xyx V xyx diag( b ) V T xyx C − 1 / 2 , x x where b ∈ {− 1 , 1 } N is a binary (signed) vector. ◮ Even if we get C y exactly, H is not identifiable ⇒ Not surprising since we only have second moment info ◮ Consider having access to multiple input distributions { C x , m } M m =1 Santiago Segarra 10 / 18

  21. Multiple input processes ◮ Define A m := ( C − 1 / 2 x , m V xyx , m ) ⊙ ( C − 1 / 2 x , m C 1 / 2 xyx , m V xyx , m )  A 1 − A 2 0 · · · 0 0  0 A 2 − A 3 · · · 0 0   Ψ :=  . . . . .  ... . . . . .   . . . . .   0 0 0 · · · A M − 1 − A M ◮ b m ∈ {− 1 , 1 } N and b = [ b T M ] T , then Ψb ∗ = 0 1 , b T 2 , . . . , b T Santiago Segarra 11 / 18

  22. Multiple input processes ◮ Define A m := ( C − 1 / 2 x , m V xyx , m ) ⊙ ( C − 1 / 2 x , m C 1 / 2 xyx , m V xyx , m )  A 1 − A 2 0 · · · 0 0  0 A 2 − A 3 · · · 0 0   Ψ :=  . . . . .  ... . . . . .   . . . . .   0 0 0 · · · A M − 1 − A M ◮ b m ∈ {− 1 , 1 } N and b = [ b T M ] T , then Ψb ∗ = 0 1 , b T 2 , . . . , b T C y , m are available, we can estimate b ∗ as Whenever only estimates ˆ T ˆ b ∗ = b ∈{− 1 , 1 } NM b T ˆ ˆ argmin Ψ Ψb , obtaining our estimate for the filter H as M H = 1 ˆ x , m ˆ xyx , m ˆ V xyx , m diag(ˆ m )ˆ � C − 1 / 2 C 1 / 2 b ∗ V T xyx , m C − 1 / 2 x , m M m =1 Santiago Segarra 11 / 18

  23. Boolean quadratic program ◮ Our problem then reduces to solving the BQP T ˆ b ∗ = b ∈{− 1 , 1 } NM b T ˆ ˆ argmin Ψ Ψb T ˆ ◮ Define ˆ W = ˆ Ψ and B = bb T Ψ B � 0 tr( ˆ min WB ) s. to rank( B ) = 1 , B ii = 1 , i = 1 , . . . , NM ◮ Drop source of non-convexity to obtain the semi-definite relaxation B ∗ = argmin tr( ˆ WB ) s. to B ii = 1 , i = 1 , . . . , NM B � 0 Santiago Segarra 12 / 18

  24. Performance guarantee ◮ For l = 1 , . . . , L , draw z l ∼ N ( 0 , B ∗ ), round ˜ b l = sign ( z l ), to obtain l ∗ = argmin ˜ l ˆ W ˜ b T b l l =1 ,..., L Santiago Segarra 13 / 18

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend