Sparsifying sums of positive semidefinite matrices Cristiane Sato - - PowerPoint PPT Presentation

sparsifying sums of positive semidefinite matrices
SMART_READER_LITE
LIVE PREVIEW

Sparsifying sums of positive semidefinite matrices Cristiane Sato - - PowerPoint PPT Presentation

Sparsifying sums of positive semidefinite matrices Cristiane Sato Joint work with Nick Harvey 1 and Marcel Silva 2 Federal University of the ABC Region, Brazil Center of Mathematics, Computing and Cognition 1 University of British Columbia 2


slide-1
SLIDE 1

Sparsifying sums of positive semidefinite matrices

Cristiane Sato Joint work with Nick Harvey 1 and Marcel Silva 2 Federal University of the ABC Region, Brazil Center of Mathematics, Computing and Cognition

1University of British Columbia 2University of S˜

ao Paulo

slide-2
SLIDE 2

Cut Sparsifiers

Theorem (Karger ’94)

◮ weighted graph G = (V , E, w) where w : E → R+, ◮ ε > 0 small

There exists a subgraph H = (V , F, y) of G and y : F → R+ s.t. |F| = O(n ln n/ε2)

◮ The weight of every cut is approximately preserved

slide-3
SLIDE 3

Cut Sparsifiers

Theorem (Karger ’94)

◮ weighted graph G = (V , E, w) where w : E → R+, ◮ ε > 0 small

There exists a subgraph H = (V , F, y) of G and y : F → R+ s.t. |F| = O(n ln n/ε2)

◮ The weight of every cut is approximately preserved ◮ That is,

w(δG(S)) = (1 ± ε)y(δH(S)), ∀S ⊆ V

slide-4
SLIDE 4

Cut Sparsifiers

Theorem (Karger ’94)

◮ weighted graph G = (V , E, w) where w : E → R+, ◮ ε > 0 small

There exists a subgraph H = (V , F, y) of G and y : F → R+ s.t. |F| = O(n ln n/ε2)

◮ The weight of every cut is approximately preserved ◮ That is,

w(δG(S)) = (1 ± ε)y(δH(S)), ∀S ⊆ V

◮ Application: Faster algorithms by preprocessing the graph

slide-5
SLIDE 5

Cut Sparsifiers

Theorem (Karger ’94)

◮ weighted graph G = (V , E, w) where w : E → R+, ◮ ε > 0 small

There exists a subgraph H = (V , F, y) of G and y : F → R+ s.t. |F| = O(n ln n/ε2)

◮ The weight of every cut is approximately preserved ◮ That is,

w(δG(S)) = (1 ± ε)y(δH(S)), ∀S ⊆ V

◮ Application: Faster algorithms by preprocessing the graph ◮ How sparse can H be? ◮ Can we build H efficiently?

slide-6
SLIDE 6

Weighted Laplacians

◮ G = (V , E, w) a weighted graph, where w : E → R+ ◮ Laplacian of G is the V × V matrix LaplG s.t.

LaplG(i, i) = degree of i LaplG(i, j) = −wi,j if ij ∈ E

slide-7
SLIDE 7

Weighted Laplacians

◮ G = (V , E, w) a weighted graph, where w : E → R+ ◮ Laplacian of G is the V × V matrix LaplG s.t.

LaplG(i, i) = degree of i LaplG(i, j) = −wi,j if ij ∈ E

◮ LaplG is positive semidefinite

All eigenvalues are ≥ 0 Notation: LaplG 0

slide-8
SLIDE 8

Spectral sparsifiers

Theorem (Spielman, Teng ’04)

◮ G = (V , E, w) a weighted graph, where w : E → R+ ◮ ε > 0 small

There are new weights y : E → R+ s.t.

◮ has n polylog(n)/ε2 nonzero entries ◮ H := (V , E, y) satisfies

LaplG LaplH (1 + ε) LaplG

slide-9
SLIDE 9

Spectral sparsifiers

Theorem (Spielman, Teng ’04)

◮ G = (V , E, w) a weighted graph, where w : E → R+ ◮ ε > 0 small

There are new weights y : E → R+ s.t.

◮ has n polylog(n)/ε2 nonzero entries ◮ H := (V , E, y) satisfies

LaplG LaplH (1 + ε) LaplG Notation: A B ⇐ ⇒ B − A is positive semidefinite

◮ nearly-linear time solvers for symmetric, diagonally-dominant linear

systems (Spielman and Teng + Koutis, Miller, and Peng, 2004)

slide-10
SLIDE 10

Spectral sparsifiers

Theorem (Spielman, Teng ’04)

◮ G = (V , E, w) a weighted graph, where w : E → R+ ◮ ε > 0 small

There are new weights y : E → R+ s.t.

◮ has n polylog(n)/ε2 nonzero entries ◮ H := (V , E, y) satisfies

LaplG LaplH (1 + ε) LaplG y may be found in ˜ O(m) time Notation: A B ⇐ ⇒ B − A is positive semidefinite

◮ nearly-linear time solvers for symmetric, diagonally-dominant linear

systems (Spielman and Teng + Koutis, Miller, and Peng, 2004)

slide-11
SLIDE 11

Spectral sparsifiers are cut sparsifiers

◮ h the incidence vector of S ⊆ V ,

hv = 1 if v ∈ S 0, otherwise

slide-12
SLIDE 12

Spectral sparsifiers are cut sparsifiers

◮ h the incidence vector of S ⊆ V ,

hv = 1 if v ∈ S 0, otherwise

◮ hT LaplG h =

  • ij∈E

xij(hi − hj)2

slide-13
SLIDE 13

Spectral sparsifiers are cut sparsifiers

◮ h the incidence vector of S ⊆ V ,

hv = 1 if v ∈ S 0, otherwise

◮ hT LaplG h =

  • ij∈E

xij(hi − hj)2

◮ hT LaplG h is the w-weight of the cut δ(S), i.e., w(δ(S))

slide-14
SLIDE 14

Spectral sparsifiers are cut sparsifiers

◮ h the incidence vector of S ⊆ V ,

hv = 1 if v ∈ S 0, otherwise

◮ hT LaplG h =

  • ij∈E

xij(hi − hj)2

◮ hT LaplG h is the w-weight of the cut δ(S), i.e., w(δ(S)) ◮ hT LaplH h is the y-weight of the cut δ(S), i.e., y(δ(S))

slide-15
SLIDE 15

Spectral sparsifiers are cut sparsifiers

◮ h the incidence vector of S ⊆ V ,

hv = 1 if v ∈ S 0, otherwise

◮ hT LaplG h =

  • ij∈E

xij(hi − hj)2

◮ hT LaplG h is the w-weight of the cut δ(S), i.e., w(δ(S)) ◮ hT LaplH h is the y-weight of the cut δ(S), i.e., y(δ(S)) ◮ LaplH LaplG implies

hT(LaplH − LaplG)h ≥ 0 That is, hT LaplH h ≥ hT LaplG h

slide-16
SLIDE 16

Spectral sparsifiers are cut sparsifiers

◮ h the incidence vector of S ⊆ V ,

hv = 1 if v ∈ S 0, otherwise

◮ hT LaplG h =

  • ij∈E

xij(hi − hj)2

◮ hT LaplG h is the w-weight of the cut δ(S), i.e., w(δ(S)) ◮ hT LaplH h is the y-weight of the cut δ(S), i.e., y(δ(S)) ◮ LaplH LaplG implies

hT(LaplH − LaplG)h ≥ 0 That is, hT LaplH h ≥ hT LaplG h

◮ LaplH (1 + ε) LaplG implies hT LaplH h ≤ (1 + ε)hT LaplG h

slide-17
SLIDE 17

Laplacian matrix as a sum of matrices

◮ G = (V , E, w) a weighted graph, where w : E → R+ ◮ The Laplacian of G is the V × V matrix

LaplG :=

  • ij∈E

wij       i j i 1 −1 j −1 1      

slide-18
SLIDE 18

Laplacian matrix as a sum of matrices

◮ G = (V , E, w) a weighted graph, where w : E → R+ ◮ The Laplacian of G is the V × V matrix

LaplG :=

  • ij∈E

wij       i 1 j −1      

  • i

j 1 −1

slide-19
SLIDE 19

Laplacian matrix as a sum of matrices

◮ G = (V , E, w) a weighted graph, where w : E → R+ ◮ The Laplacian of G is the V × V matrix

LaplG :=

  • ij∈E

wij       i 1 j −1      

  • i

j 1 −1

  • ◮ LaplG is a sum of rank-one positive semidefinite matrices
slide-20
SLIDE 20

Sparsifiers of Sums of Rank-One PSD Matrices

Theorem (Batson, Spielman, Srivastava ’09)

◮ B1, . . . , Bm p.s.d. n × n matrices of rank one ◮ B := i Bi ◮ ε > 0 small

There are new weights y ∈ Rm

+ s.t. ◮ y has O(n/ε2) nonzero entries ◮ B i yiBi (1 + ε)B

y may be found in O(mn3/ε2) time

slide-21
SLIDE 21

Sparsifiers of Sums of Rank-One PSD Matrices

Theorem (Batson, Spielman, Srivastava ’09)

◮ B1, . . . , Bm p.s.d. n × n matrices of rank one ◮ B := i Bi ◮ ε > 0 small

There are new weights y ∈ Rm

+ s.t. ◮ y has O(n/ε2) nonzero entries ◮ B i yiBi (1 + ε)B

y may be found in O(mn3/ε2) time

◮ (Lee-Sun’15) Almost-linear time method

slide-22
SLIDE 22

Sparsifiers of Sums of Rank-One PSD Matrices

Theorem (Batson, Spielman, Srivastava ’09)

◮ B1, . . . , Bm p.s.d. n × n matrices of rank one ◮ B := i Bi ◮ ε > 0 small

There are new weights y ∈ Rm

+ s.t. ◮ y has O(n/ε2) nonzero entries ◮ B i yiBi (1 + ε)B

y may be found in O(mn3/ε2) time

slide-23
SLIDE 23

Sparsifiers of Sums of PSD Matrices

Theorem (de Carli Silva, Harvey, S., ’11)

◮ B1, . . . , Bm p.s.d. n × n matrices of any rank ◮ B := i Bi ◮ ε > 0 small

There are new weights y ∈ Rm

+ s.t. ◮ y has O(n/ε2) nonzero entries ◮ B i yiBi (1 + ε)B

y may be found in O(mn3/ε2) time

slide-24
SLIDE 24

Applications

◮ spectral sparsifiers of graphs with extra properties ◮ cut sparsifiers of uniform hypergraphs (specially 3-uniform) ◮ sparse solutions to semidefinite programs

slide-25
SLIDE 25

Sparsifiers with Costs

Theorem

◮ G = (V , E, w) a weighted graph, where w : E → R+ ◮ ε > 0 small

There are new weights y : E → R+ s.t.

◮ y has O(n/ε2) nonzero entries ◮ the reweighted graph H := (V , E, y) satisfies

LaplG LaplH (1 + ε) LaplG y may be found in O(mn3/ε2) time

slide-26
SLIDE 26

Sparsifiers with Costs

Theorem

◮ G = (V , E, w) a weighted graph, where w : E → R+ ◮ ε > 0 small ◮ “costs” c : E → R+

There are new weights y : E → R+ s.t.

◮ y has O(n/ε2) nonzero entries ◮ the reweighted graph H := (V , E, y) satisfies

LaplG LaplH (1 + ε) LaplG

◮ cTw ≤ cTy ≤ (1 + ε)cTw

y may be found in O(mn3/ε2) time

slide-27
SLIDE 27

Add extra info to Laplacian

  • ij∈E

wij       i j i 1 −1 j −1 1      

slide-28
SLIDE 28

Add extra info to Laplacian

  • ij∈E

      i j i wij −wij j −wij wij      

slide-29
SLIDE 29

Add extra info to Laplacian

  • ij∈E

        i j i wij −wij j −wij wij cij        

slide-30
SLIDE 30

Cut Sparsifiers of 3-Uniform Hypergraphs

Theorem

◮ G = (V , E, w) a weighted 3-uniform hypergraph, where w : E → R+ ◮ i.e., E ⊆

V

3

  • ◮ ε > 0 small

There are new weights y : E → R+ s.t.

◮ y has O(n/ε2) nonzero entries ◮ the reweighted hypergraph H := (V , E, y) satisfies

w(δG(S)) ≤ y(δH(S)) ≤ (1 + ε)w(δG(S)) ∀S ⊆ V y may be found in O(mn3/ε2) time

slide-31
SLIDE 31

Hypergraph Laplacians

  • ijk∈E

wijk           i j k i 2 −1 −1 j −1 2 −1 k −1 −1 2          

slide-32
SLIDE 32

Semidefinite Programs

Theorem

◮ B1, . . . , Bm p.s.d. n × n matrices

B sym n × n matrix

◮ c ∈ Rm + ◮ Semidefinite program (SDP)

min cTz

  • i

ziBi B z ∈ Rm

+ ◮ feasible solution z∗ ◮ ε ∈ (0, 1)

There exists a feasible solution ˜ z with at most O(n/ε2) nonzero entries and cT˜ z ≤ (1 + ε)cTz∗.

slide-33
SLIDE 33

Future directions

◮ Find more applications of the arbitrary-rank sparsification result ◮ Improve running times

slide-34
SLIDE 34

Future directions

◮ Find more applications of the arbitrary-rank sparsification result ◮ Improve running times ◮ Positive Semidefiniteness Assumption

For each n > 0, there exist B1, . . . , Bm with m = Ω(n2) and B :=

i Bi p.d. such that

for every ε ∈ (0, 1) and y ∈ Rm

+ with (1 − ε)B i yiBi,

every entry of y is nonzero

slide-35
SLIDE 35

Pseudoinverse

We may assume that

m

  • i=1

Bi = I by applying Moore-Penrose pseudoinverse

slide-36
SLIDE 36

O(n log n/ε2) versions

◮ Ahlswede–Winter Theorem ◮ Can be de-randomized using pessimistic estimators

slide-37
SLIDE 37

The approach

  • 1. Start with A = 0
  • 2. In each iteration choose a matrix Bi and compute a weight α.

Set A = A + αBi

slide-38
SLIDE 38

The approach

  • 1. Start with A = 0
  • 2. In each iteration choose a matrix Bi and compute a weight α.

Set A = A + αBi Batson, Spielman, Srivastava: Bi = vv T is a rank-one matrix. Use Sherman-Morrison formula Tr(M − αvv T)−1 = Tr(M−1) + αv TM−2v 1 − v TM−1v

slide-39
SLIDE 39

The approach

  • 1. Start with A = 0
  • 2. In each iteration choose a matrix Bi and compute a weight α.

Set A = A + αBi Batson, Spielman, Srivastava: Bi = vv T is a rank-one matrix. Use Sherman-Morrison formula Tr(M − αvv T)−1 = Tr(M−1) + αv TM−2v 1 − v TM−1v Ours: Bi = VV T is an arbitrary-rank matrix. Sherman-Morrison-Woodbury formula: Tr(M−αVV T)−1 = Tr

  • M−1

+Tr

  • αM−1V (I −αV TM−1V )−1V TM−1
slide-40
SLIDE 40

Upper barrier

◮ u (upper bound for eigenvalues) ◮ Barrier function

Φu(A) =

n

  • i=1

1 u − λi(A) = Tr (uI − A)−1

slide-41
SLIDE 41

Upper barrier

◮ u (upper bound for eigenvalues) ◮ Barrier function

Φu(A) =

n

  • i=1

1 u − λi(A) = Tr (uI − A)−1

◮ matrices A and B 0, we want to control the λmax(A + αB) ◮ Given δu > 0,

suppose we want λmax(A + αB) ≤ u + δU := u′. What conditions on α guarantee that?

slide-42
SLIDE 42

Upper barrier

◮ u (upper bound for eigenvalues) ◮ Barrier function

Φu(A) =

n

  • i=1

1 u − λi(A) = Tr (uI − A)−1

◮ matrices A and B 0, we want to control the λmax(A + αB) ◮ Given δu > 0,

suppose we want λmax(A + αB) ≤ u + δU := u′. What conditions on α guarantee that?

◮ We have

1/α ≥ UA(B) implies Φu′(A + αB) ≤ Φu(A). and λmax(A + αB) < u′ where M := u′I − A UA(B) := M−2, B Φu(A) − Φu′(A) + M−1, B

slide-43
SLIDE 43

Lower barrier

◮ ℓ (lower bound for eigenvalues) ◮ Barrier function

Φℓ(A) =

n

  • i=1

1 λi(A) − ℓ = Tr (A − ℓI)−1

◮ Given δℓ > 0,

suppose we want λmin(A + αB) ≥ ℓ + δℓ := ℓ′. What conditions on α guarantee that?

◮ We have

1/α ≤ LA(B) implies Φℓ′(A + αB) ≤ Φℓ(A). and λmin(A + αB) > ℓ′ where N := A − ℓ′I LA(B) := N−2, B Φℓ′(A) − Φℓ(A) + N−1, B

slide-44
SLIDE 44

Overview

◮ A(0) := 0 and y(0) := 0.

Parameters u0, ℓ0, δL, δU to be chosen T := 4n/ε2. Define the barrier functions Φu0(A) and Φℓ0(A).

slide-45
SLIDE 45

Overview

◮ A(0) := 0 and y(0) := 0.

Parameters u0, ℓ0, δL, δU to be chosen T := 4n/ε2. Define the barrier functions Φu0(A) and Φℓ0(A).

◮ For t = 1, . . . , T

◮ ut := ut−1 + δU and ℓt := ℓt−1 + δL. ◮ Find a matrix Bj and a value α > 0 such that

Φut (A(t − 1) + αBj) ≤ Φut−1(A(t − 1)) Φℓt (A(t − 1) + αBj) ≤ Φℓt−1(A(t − 1)). and so A(t − 1) + αBj ∈ [ℓt, ut]

◮ A(t) := A(t − 1) + αBj

y(t) := y(t − 1) + αej.

slide-46
SLIDE 46

End of the algorithm

λmax(A(T)) λmin(A(T)) ≤ u0 + δUT ℓ0 + δLT ≤ 1 + ε 1 − ε with T = 4n/ε2 δL := 1 εL := ε 2 ℓ0 := − n εL δU := 2 + ε 2 − ε εU := ε 2δU u0 := n εU .

slide-47
SLIDE 47

Overview

◮ A(0) := 0 and y(0) := 0.

Parameters u0, ℓ0, δL, δU to be chosen T := 4n/ε2. Define the barrier functions Φu0(A) and Φℓ0(A).

◮ For t = 1, . . . , T

◮ ut := ut−1 + δU and ℓt := ℓt−1 + δL. ◮ Find a matrix Bj and a value α > 0 such that

Φut (A(t − 1) + αBj) ≤ Φut−1(A(t − 1)) Φℓt (A(t − 1) + αBj) ≤ Φℓt−1(A(t − 1)). and so A(t − 1) + αBj ∈ [ℓt, ut]

◮ A(t) := A(t − 1) + αBj

y(t) := y(t − 1) + αej.

slide-48
SLIDE 48

Satisfying both barriers at the same time

◮ Averaging argument

  • i

UA(Bi) ≤ 1/δU + εU < 1/δL + εL ≤

  • i

LA(Bi)

slide-49
SLIDE 49

Satisfying both barriers at the same time

◮ Averaging argument

  • i

UA(Bi) ≤ 1/δU + εU < 1/δL + εL ≤

  • i

LA(Bi)

◮ ∃ i s.t. UA(Bi) ≤ LA(Bi) ◮ We can choose α such that

UA(Bi) ≤ 1/α ≤ LA(Bi) as needed for the algorithm

slide-50
SLIDE 50

Satisfying both barriers at the same time

◮ Averaging argument

  • i

UA(Bi) ≤ 1/δU + εU < 1/δL + εL ≤

  • i

LA(Bi)

◮ ∃ i s.t. UA(Bi) ≤ LA(Bi) ◮ We can choose α such that

UA(Bi) ≤ 1/α ≤ LA(Bi) as needed for the algorithm

◮ Compute UA(Bi) and LA(Bi)