Nonconvex Sparse Graph Learning under Laplacian-structured Graphical - - PowerPoint PPT Presentation

nonconvex sparse graph learning under laplacian
SMART_READER_LITE
LIVE PREVIEW

Nonconvex Sparse Graph Learning under Laplacian-structured Graphical - - PowerPoint PPT Presentation

Nonconvex Sparse Graph Learning under Laplacian-structured Graphical Model a talk by Jiaxi Ying, Jos Vincius de M. Cardoso, and Daniel P. Palomar The Hong Kong University of Science and Technology Thirty-fourth Conference on Neural


slide-1
SLIDE 1

Nonconvex Sparse Graph Learning under Laplacian-structured Graphical Model

Thirty-fourth Conference on Neural Information Processing Systems (NeurIPS 2020), Vancouver, Canada

a talk by

Jiaxi Ying, José Vinícius de M. Cardoso, and Daniel P. Palomar The Hong Kong University of Science and Technology

December, 2020

slide-2
SLIDE 2

Learning Sparse Undirected Connected Graphs

1/8

data generating process: Laplacian-constrained Gaussian Markov Random Field (L-GMRF) with rank p − 1 its p × p precision matrix Θ is modeled as a combinatorial graph Laplacian state of the art (Egilmez et al. 2017)1, (Zhao et al. 2019)2: minimize

Θ0

tr(SΘ) − log det⋆ (Θ + J) + λΘ1,off, subject to Θ1 = 0, Θij = Θji ≤ 0 (1) where J = 1

p11⊤, Θ1,off = i>j |Θij| is the entrywise ℓ1-norm, and λ ≥ 0

1HE Egilmez et al. Graph learning from data under Laplacian and structural constraints. IEEE Journal

  • f Selected Topics in Signal Processing 11 (6), 825-841.

2L Zhao et al. Optimization algorithms for graph laplacian estimation via ADMM and MM. IEEE

Transactions on Signal Processing 67 (16), 4231-4244.

slide-3
SLIDE 3

Are sparse solutions recoverable via ℓ1-norm?

2/8

TL;DR: they aren’t empirically:

(a) ground-truth (b) λ = 0 (c) λ = 0.1 (d) λ = 10

slide-4
SLIDE 4

Are sparse solutions recoverable via ℓ1-norm?

3/8

theoretically: Theorem

Let ˆ Θ ∈ Rp×p be the global minimum of (1) with p > 3. Define s1 = maxk Skk and s2 = minij Sij. If the regularization parameter λ in (1) satisfies λ ∈ [(2 + 2 √ 2)(p + 1)(s1 − s2), +∞), then the estimated graph weight ˆ Wij = −ˆ Θij

  • beys

ˆ Wij ≥ 1 (s1 − (p + 1)s2 + λ)p > 0, ∀ i = j.

Proof

Please refer to our supplementary material

slide-5
SLIDE 5

Our framework for sparse graphs

4/8

nonconvex formulation: minimize

w≥0

tr(SLw) − log det(Lw + J) +

i hλ(wi)

(2) L is the Laplacian operator and hλ(·) is a nonconvex regularizer such as

Minimax Concave Penalty (MCP) Smoothly Clipped Absolute Deviation (SCAD)

slide-6
SLIDE 6

Our framework for sparse graphs

5/8

Algorithm 0: Connected sparse graph learning Data: Sample covariance S, λ > 0, ˆ w(0) Result: Laplacian estimation: L ˆ w(k)

1 k ← 1 2 while stopping criteria not met do 3

⊲ update z(k−1)

i

= h′

λ(ˆ

w(k−1)

i

), for i = 1, . . . , p(p − 1)/2

4

⊲ update ˆ w(k) = arg minw≥0 − log det(Lw + J) + tr(SLw) +

i z(k−1) i

wi

5

⊲ k ← k + 1

6 end

slide-7
SLIDE 7

Sneak peek on the results: synthetic data

6/8

slide-8
SLIDE 8

Sneak peek on the results: S&P 500 stocks

7/8

(a) GLE-ADMM (benchmark) λ = 0 (b) NGL-MCP (proposed) λ = 0.5

slide-9
SLIDE 9

Reproducibility

8/8

The code for the experiments can be found at https://github.com/mirca/sparseGraph Convex Research Group at HKUST: https://www.danielppalomar.com