nonconvex sparse graph learning under laplacian
play

Nonconvex Sparse Graph Learning under Laplacian-structured Graphical - PowerPoint PPT Presentation

Nonconvex Sparse Graph Learning under Laplacian-structured Graphical Model a talk by Jiaxi Ying, Jos Vincius de M. Cardoso, and Daniel P. Palomar The Hong Kong University of Science and Technology Thirty-fourth Conference on Neural


  1. Nonconvex Sparse Graph Learning under Laplacian-structured Graphical Model a talk by Jiaxi Ying, José Vinícius de M. Cardoso, and Daniel P. Palomar The Hong Kong University of Science and Technology Thirty-fourth Conference on Neural Information Processing Systems (NeurIPS 2020), Vancouver, Canada December, 2020

  2. Learning Sparse Undirected Connected Graphs data generating process: Laplacian-constrained Gaussian Markov Random Field (L-GMRF) with rank p − 1 its p × p precision matrix Θ is modeled as a combinatorial graph Laplacian state of the art (Egilmez et al. 2017) 1 , (Zhao et al. 2019) 2 : tr ( S Θ ) − log det ⋆ ( Θ + J ) + λ � Θ � 1 , off , minimize Θ � 0 (1) subject to Θ 1 = 0 , Θ ij = Θ ji ≤ 0 where J = 1 p 11 ⊤ , � Θ � 1 , off = � i > j | Θ ij | is the entrywise ℓ 1 -norm, and λ ≥ 0 1 HE Egilmez et al. Graph learning from data under Laplacian and structural constraints. IEEE Journal of Selected Topics in Signal Processing 11 (6), 825-841. 2 L Zhao et al. Optimization algorithms for graph laplacian estimation via ADMM and MM. IEEE Transactions on Signal Processing 67 (16), 4231-4244. 1/8

  3. Are sparse solutions recoverable via ℓ 1 -norm? TL;DR: they aren’t empirically: (a) ground-truth (b) λ = 0 (c) λ = 0 . 1 (d) λ = 10 2/8

  4. Are sparse solutions recoverable via ℓ 1 -norm? theoretically: Theorem Θ ∈ R p × p be the global minimum of (1) with p > 3 . Define s 1 = max k S kk and Let ˆ s 2 = min ij S ij . If the regularization parameter λ in (1) satisfies √ 2 )( p + 1 )( s 1 − s 2 ) , + ∞ ) , then the estimated graph weight ˆ W ij = − ˆ λ ∈ [( 2 + 2 Θ ij obeys 1 ˆ W ij ≥ ( s 1 − ( p + 1 ) s 2 + λ ) p > 0 , ∀ i � = j . Proof Please refer to our supplementary material 3/8

  5. Our framework for sparse graphs nonconvex formulation: minimize tr ( S L w ) − log det( L w + J ) + � i h λ ( w i ) (2) w ≥ 0 L is the Laplacian operator and h λ ( · ) is a nonconvex regularizer such as Minimax Concave Penalty (MCP) Smoothly Clipped Absolute Deviation (SCAD) 4/8

  6. Our framework for sparse graphs Algorithm 0: Connected sparse graph learning w ( 0 ) Data: Sample covariance S , λ > 0 , ˆ w ( k ) Result: Laplacian estimation: L ˆ 1 k ← 1 2 while stopping criteria not met do ⊲ update z ( k − 1 ) w ( k − 1 ) = h ′ λ (ˆ ) , for i = 1 , . . . , p ( p − 1 ) / 2 3 i i w ( k ) = arg min w ≥ 0 − log det( L w + J ) + tr ( S L w ) + � i z ( k − 1 ) ⊲ update ˆ w i 4 i ⊲ k ← k + 1 5 6 end 5/8

  7. Sneak peek on the results: synthetic data 6/8

  8. Sneak peek on the results: S&P 500 stocks (a) GLE-ADMM (benchmark) λ = 0 (b) NGL-MCP (proposed) λ = 0 . 5 7/8

  9. Reproducibility The code for the experiments can be found at https://github.com/mirca/sparseGraph Convex Research Group at HKUST: https://www.danielppalomar.com 8/8

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend