Lecture 4: Goemans-Williamson Algorithm for MAX-CUT Lecture Outline - - PowerPoint PPT Presentation

β–Ά
lecture 4 goemans williamson algorithm for max cut
SMART_READER_LITE
LIVE PREVIEW

Lecture 4: Goemans-Williamson Algorithm for MAX-CUT Lecture Outline - - PowerPoint PPT Presentation

Lecture 4: Goemans-Williamson Algorithm for MAX-CUT Lecture Outline Part I: Analyzing semidefinite programs Part II: Analyzing Goemans-Williamson Part III: Tight examples for Goemans-Williamson Part IV: Impressiveness of


slide-1
SLIDE 1

Lecture 4: Goemans-Williamson Algorithm for MAX-CUT

slide-2
SLIDE 2

Lecture Outline

  • Part I: Analyzing semidefinite programs
  • Part II: Analyzing Goemans-Williamson
  • Part III: Tight examples for Goemans-Williamson
  • Part IV: Impressiveness of Goemans-Williamson

and open problems

slide-3
SLIDE 3

Part I: Analyzing semidefinite programs

slide-4
SLIDE 4

Goemans-Williamson Program

  • Recall Goemans-Williamson program: Maximize

σ𝑗,π‘˜:𝑗<π‘˜, 𝑗,π‘˜ ∈𝐹(𝐻)

1βˆ’π‘π‘—π‘˜ 2

subject to M ≽ 0 where 𝑁 ≽ 0 and βˆ€π‘—, 𝑁𝑗𝑗 = 1

  • Theorem: Goemans-Williamson gives a .878

approximation for MAX-CUT

  • How do we analyze Goemans-Williamson and
  • ther semidefinite programs?
slide-5
SLIDE 5

Vector Solutions

  • Want: matrix 𝑁 such that π‘π‘—π‘˜ = π‘¦π‘—π‘¦π‘˜ where

𝑦𝑗 are the problem variables.

  • Semidefinite program: Assigns a vector 𝑀𝑗 to

each 𝑦𝑗, gives the matrix 𝑁 where π‘π‘—π‘˜ = 𝑀𝑗 β‹… π‘€π‘˜

  • Note: This is a relaxation of the problem. To
  • btain an actual solution, we need a rounding

algorithm to round this vector solution into an actual solution.

slide-6
SLIDE 6

Vector Solution Justification

  • Theorem: 𝑁 ≽ 0 if and only if there are vectors

𝑀𝑗 such that π‘π‘—π‘˜ = 𝑀𝑗 β‹… π‘€π‘˜

  • Example: 𝑁 =

1 βˆ’1 1 βˆ’1 2 βˆ’1 1 βˆ’1 2 , 𝑀1 = 1,0,0 𝑀2 = βˆ’1,1,0 𝑀3 = 1,0,1

  • One way to see this: take a β€œsquare root” of 𝑁
  • Second way to see this: Cholesky decomposition
slide-7
SLIDE 7

Square Root of a PSD Matrix

  • If there are vectors {𝑀𝑗} such that π‘π‘—π‘˜ = 𝑀𝑗 β‹… π‘€π‘˜,

take π‘Š to be the matrix with rows 𝑀1, β‹― , π‘€π‘œ. 𝑁 = π‘Šπ‘Šπ‘ˆ ≽ 0

  • Conversely, if 𝑁 ≽ 0 then 𝑁 = σ𝑗=1

π‘œ

πœ‡π‘—π‘£π‘—π‘£π‘—

π‘ˆ

where πœ‡π‘— β‰₯ 0 for all 𝑗. Taking π‘Š to be the matrix with columns πœ‡π‘—π‘£π‘—, π‘Šπ‘Šπ‘ˆ = 𝑁. Taking 𝑀𝑗 to be the ith row of π‘Š, π‘π‘—π‘˜ = 𝑀𝑗 β‹… π‘€π‘˜

slide-8
SLIDE 8

Cholesky Decomposition

  • Cholesky decomposition: 𝑁 = π·π·π‘ˆ where 𝐷 is

a lower triangular matrix.

  • 𝑀𝑗 = σ𝑏 𝐷𝑗𝑏𝑓𝑏 is the ith row of 𝐷
  • We can find the entries of 𝐷 one by one.
slide-9
SLIDE 9

Cholesky Decomposition Example

  • Example: 𝑁 =

1 βˆ’1 1 βˆ’1 2 βˆ’1 1 βˆ’1 2

  • 𝑀1 = 1,0,0
  • Need 𝐷21 = βˆ’1 so that 𝑀2 β‹… 𝑀1 = βˆ’1. 𝑀2 =

βˆ’1, 𝐷22, 0

  • Taking 𝐷22 = 1, 𝑀2 β‹… 𝑀2 = 2. 𝑀2 = βˆ’1,1,0
  • Need 𝐷31 = 1 and 𝐷32 = 0 so that 𝑀3 β‹… 𝑀1 =

1, 𝑀3 β‹… 𝑀2 = βˆ’1. 𝑀3 = 1,0, 𝐷33 .

  • Taking 𝐷33 = 1, 𝑀3 β‹… 𝑀3 = 1. 𝑀3 = 1,0,1
slide-10
SLIDE 10

Cholesky Decomposition Example

  • 1

βˆ’1 1 βˆ’1 2 βˆ’1 1 βˆ’1 2 = 1 βˆ’1 1 1 1 1 βˆ’1 1 1 1

  • 𝑀1 =

1 , 𝑀2 = βˆ’1 1 , 𝑀3 = 1 1

slide-11
SLIDE 11

Cholesky Decomposition Formulas

  • βˆ€π‘— < 𝑙, take 𝐷𝑙𝑗 = π‘π‘—π‘™βˆ’Οƒπ‘=1

π‘—βˆ’1 𝐷𝑙𝑏𝐷𝑗𝑏

𝑑𝑗𝑗

  • Take 𝐷𝑙𝑗 = 0 if 𝑁𝑗𝑙 βˆ’ σ𝑏=1

π‘—βˆ’1 𝐷𝑙𝑏𝐷𝑗𝑏 = 𝐷𝑗𝑗 = 0

  • Note that 𝑀𝑙 β‹… 𝑀𝑗 = σ𝑏=1

π‘—βˆ’1 𝐷𝑙𝑏 𝐷𝑗𝑏 + 𝐷𝑙𝑗𝐷𝑗𝑗 = 𝑁𝑗𝑙

  • βˆ€π‘™, take 𝐷𝑙𝑙 =

𝑁𝑙𝑙 βˆ’ σ𝑏=1

π‘™βˆ’1 𝐷𝑙𝑏 2

  • These formulas are the basis for the Cholesky-

Banachiewicz algorithm and the Cholesky-Crout algorithm (these algorithms only differ in the

  • rder the entries are evaluated)
slide-12
SLIDE 12

Cholesky Decomposition Failure

1. βˆ€π‘— < 𝑙, 𝐷𝑙𝑗 =

π‘π‘—π‘™βˆ’Οƒπ‘=1

π‘—βˆ’1 𝐷𝑙𝑏𝐷𝑗𝑏

𝐷𝑗𝑗

2. βˆ€π‘™, 𝐷𝑙𝑙 = 𝑁𝑙𝑙 βˆ’ σ𝑏=1

π‘™βˆ’1 𝐷𝑙𝑏 2

  • If the Cholesky decomposition succeeds, it gives

us vectors 𝑀𝑗 such that π‘π‘—π‘˜ = 𝑀𝑗 β‹… π‘€π‘˜

  • The formulas can fail in two ways:

1. 𝑁𝑙𝑙 βˆ’ σ𝑏=1

π‘™βˆ’1 𝐷𝑙𝑏 2 < 0 for some 𝑙

2. 𝐷𝑗𝑗 = 0 and 𝑁𝑗𝑙 βˆ’ σ𝑏=1

π‘—βˆ’1 𝐷𝑙𝑏𝐷𝑗𝑏 β‰  0 for some 𝑗, 𝑙

  • Failure implies 𝑁 is not PSD (see problem set)
slide-13
SLIDE 13

Part II: Analyzing Goemans- Williamson

slide-14
SLIDE 14

Vectors for Goemans-Williamson

  • Goemans-Williamson: Maximize

σ𝑗,π‘˜:𝑗<π‘˜, 𝑗,π‘˜ ∈𝐹(𝐻)

1βˆ’π‘π‘—π‘˜ 2

subject to M ≽ 0 where 𝑁 ≽ 0 and βˆ€π‘—, 𝑁𝑗𝑗 = 1

  • Semidefinite program gives us vectors 𝑀𝑗

where 𝑀𝑗 β‹… π‘€π‘˜ = π‘π‘—π‘˜

4 3 1 5 2 𝑀1 𝑀4 𝑀2 𝑀5 𝑀3

G

slide-15
SLIDE 15

Rounding Vectors

  • Beautiful idea: Map each vector 𝑀𝑗 to Β±1 by

taking a random vector π‘₯ and setting 𝑦𝑗 = 1 if π‘₯ β‹… 𝑀𝑗 > 0 and setting 𝑦𝑗 = βˆ’1 if π‘₯ β‹… 𝑀𝑗 < 0

  • Example:

𝑀1 𝑀4 𝑀2 𝑀5 𝑀3 π‘₯ 𝑦1 = 𝑦4 = 1, 𝑦2 = 𝑦3 = 𝑦5 = βˆ’1 4 3 1 5 2

G

slide-16
SLIDE 16

Expected Cut Value

  • Consider 𝐹 σ𝑗,π‘˜:𝑗<π‘˜, 𝑗,π‘˜ ∈𝐹 𝐻

1βˆ’π‘¦π‘—π‘¦π‘˜ 2

  • For each 𝑗, π‘˜ such that 𝑗 < π‘˜, 𝑗, π‘˜ ∈ 𝐹(𝐻),

𝐹

1βˆ’π‘¦π‘—π‘¦π‘˜ 2

= Θ

𝜌 where Θ ∈ [0, 𝜌] is the angle

between 𝑀𝑗 and π‘€π‘˜

  • On the other hand

1βˆ’π‘π‘—π‘˜ 2

=

1βˆ’π‘‘π‘π‘‘Ξ˜ 2

slide-17
SLIDE 17

Approximation Factor

  • Goemens-Williamson gives a cut with expected

value at least min

Θ

Θ 𝜌 1βˆ’π‘‘π‘π‘‘Ξ˜ 2

σ𝑗,π‘˜:𝑗<π‘˜, 𝑗,π‘˜ ∈𝐹 𝐻

1βˆ’πΉπ‘—π‘˜ 2

  • The first term is β‰ˆ .878 at Ξ˜π‘‘π‘ π‘—π‘’ β‰ˆ 134Β°

σ𝑗,π‘˜:𝑗<π‘˜, 𝑗,π‘˜ ∈𝐹 𝐻

1βˆ’πΉπ‘—π‘˜ 2

is an upper bound on the max cut size, so we have a .878 approximation.

slide-18
SLIDE 18

Part III: Tight Examples

slide-19
SLIDE 19

Showing Tightness

  • How can we show this analysis is tight?
  • We give two examples where we obtain a cut of

value β‰ˆ .878 σ𝑗,π‘˜:𝑗<π‘˜, 𝑗,π‘˜ ∈𝐹 𝐻

1βˆ’πΉπ‘—π‘˜ 2

  • In one example, σ𝑗,π‘˜:𝑗<π‘˜, 𝑗,π‘˜ ∈𝐹 𝐻

1βˆ’πΉπ‘—π‘˜ 2

is the value of the maximum cut. In the other example, .878 σ𝑗,π‘˜:𝑗<π‘˜, 𝑗,π‘˜ ∈𝐹 𝐻

1βˆ’πΉπ‘—π‘˜ 2

is the value

  • f the maximum cut.
slide-20
SLIDE 20

Example 1: Hypercube

  • Have one vertex for each point 𝑦𝑗 ∈ {Β±1}π‘œ
  • We have an edge between 𝑦𝑗 and π‘¦π‘˜ in 𝐻 if

cosβˆ’1

π‘¦π‘—β‹…π‘¦π‘˜ π‘œ

βˆ’ Ξ˜π‘‘π‘ π‘—π‘’ < πœ€ for an arbitrarily small πœ€ > 0

  • Goemans-Williamson value β‰ˆ

1βˆ’cos Ξ˜π‘‘π‘ π‘—π‘’ 2

𝐹(𝐻)

  • This is achieved by the coordinate cuts.
  • Goemans-Williamson rounds to a random cut

which gives value β‰ˆ Ξ˜π‘‘π‘ π‘—π‘’

𝜌 𝐹(𝐻)

slide-21
SLIDE 21

Example 2: Sphere

  • Take a large number of random points {𝑦𝑗} on

the unit sphere

  • We have an edge between 𝑦𝑗 and π‘¦π‘˜ in 𝐻 if

cosβˆ’1 𝑦𝑗 β‹… π‘¦π‘˜ βˆ’ Ξ˜π‘‘π‘ π‘—π‘’ < πœ€ for an arbitrarily small πœ€ > 0

  • Goemans-Williamson value β‰ˆ 1βˆ’cos Ξ˜π‘‘π‘ π‘—π‘’

2

𝐹(𝐻)

  • A random hyperplane cut gives value β‰ˆ

Ξ˜π‘‘π‘ π‘—π‘’ 𝜌 𝐹(𝐻) and this is essentially optimal.

slide-22
SLIDE 22

Proof requirements

  • How can we prove the above examples behave

as claimed?

  • For the hypercube, have to upper bound the

value of the Goemans-Williamson program.

  • This can be done by determining the

eigenvalues of the hypercube graph and using this to analyze the dual (see problem set)

  • For the sphere, have to prove that no cut does

better than a random hyperplane cut (this is hard, see Feige-Schechtman [FS02])

slide-23
SLIDE 23

Part IV: Impressiveness of Goemans- Williamson and Open Problems

slide-24
SLIDE 24

Failure of Linear Programming

  • Trivial algorithm: Randomly guess which side of

the cut each vertex is on.

  • Gives approximation factor

1 2

  • Linear programming doesn’t do any better, not

even polynomial sized linear programming extensions [CLRS13]!

slide-25
SLIDE 25

Hardness of beating GW

  • Only know NP-hardness for a 16

17 approximation

[HΓ₯s01], [TSSW00]

  • Unique-Games hard to beat Goemans-

Williamson on MAX-CUT [KKMO07]

slide-26
SLIDE 26

Open problems

  • Can we find a subexponential time algorithm

beating Goemans-Williamson on max cut?

  • Can we prove constant degree SOS lower

bounds for obtaining a better approximation than Goemans-Williamson?

slide-27
SLIDE 27

References

  • [CLRS13] S. Chan, J. Lee, P. Raghavendra, and D. Steurer. Approximate constraint

satisfaction requires large lp relaxations. FOCS 2013.

  • [FS02] U. Feige and G. Schechtman. On the optimality of the random hyperplane

rounding technique for max cut. Random Structures & Algorithms - Probabilistic methods in combinatorial optimization, 20 (3), p. 403 – 440. 2002

  • [GW95] M. X. Goemans and D. P. Williamson. Improved Approximation Algorithms

for Maximum Cut and Satisfiability Problems Using Semidefinite Programming. J. ACM, 42(6):1115-1145, 1995.

  • [HΓ₯s01] J. HΓ₯stad. Some optimal inapproximability results. JACM 48: p.798-869,

2001.

  • [KKMO07] S. Khot, G. Kindler, E. Mossell, R. O’Donnell. Optimal Inapproximability

Results for MAX-CUT and Other 2-Variable CSPs? SIAM Journal on Computing, 37 (1): p. 319-357, 2007.

  • [TSSW00] L. Trevisan, G. Sorkin, M. Sudan, and D. Williamson. Gadgets,

approximation, and linear programming. SIAM Journal on Computing, 29(6): p. 2074-2097, 2000.