Introduction to LP and SDP Hierarchies Madhur Tulsiani Princeton - - PowerPoint PPT Presentation

introduction to lp and sdp hierarchies
SMART_READER_LITE
LIVE PREVIEW

Introduction to LP and SDP Hierarchies Madhur Tulsiani Princeton - - PowerPoint PPT Presentation

Introduction to LP and SDP Hierarchies Madhur Tulsiani Princeton University Convex Relaxations for Combinatorial Optimization Toy Problem minimize: x + y subject to: x + 2 y 1 2 x + y 1 x , y { 0 , 1 } Convex Relaxations for


slide-1
SLIDE 1

Introduction to LP and SDP Hierarchies

Madhur Tulsiani

Princeton University

slide-2
SLIDE 2

Convex Relaxations for Combinatorial Optimization

Toy Problem minimize: x + y subject to: x + 2y ≥ 1 2x + y ≥ 1 x, y ∈ {0, 1}

slide-3
SLIDE 3

Convex Relaxations for Combinatorial Optimization

Toy Problem minimize: x + y subject to: x + 2y ≥ 1 2x + y ≥ 1 x, y ∈ {0, 1}

00 01 10 11

slide-4
SLIDE 4

Convex Relaxations for Combinatorial Optimization

Toy Problem minimize: x + y subject to: x + 2y ≥ 1 2x + y ≥ 1 x, y ∈ {0, 1}

00 01 10 11

slide-5
SLIDE 5

Convex Relaxations for Combinatorial Optimization

Toy Problem minimize: x + y subject to: x + 2y ≥ 1 2x + y ≥ 1 x, y ∈ [0, 1]

00 01 10 11

slide-6
SLIDE 6

Convex Relaxations for Combinatorial Optimization

Toy Problem minimize: x + y subject to: x + 2y ≥ 1 2x + y ≥ 1 x, y ∈ [0, 1]

00 01 10 11 ( 1

3, 1 3)

slide-7
SLIDE 7

Convex Relaxations for Combinatorial Optimization

Toy Problem minimize: x + y subject to: x + 2y ≥ 1 2x + y ≥ 1 x, y ∈ [0, 1]

00 01 10 11 ( 1

3, 1 3)

Large number of approximation algorithms derived precisely as above. Analysis consists of understanding extra solutions introduced by the relaxation.

slide-8
SLIDE 8

Convex Relaxations for Combinatorial Optimization

Toy Problem minimize: x + y subject to: x + 2y ≥ 1 2x + y ≥ 1 x, y ∈ [0, 1]

00 01 10 11 ( 1

3, 1 3)

Large number of approximation algorithms derived precisely as above. Analysis consists of understanding extra solutions introduced by the relaxation. Integrality Gap = Combinatorial Optimum Optimum of Relaxation = 1 2/3 = 3 2

slide-9
SLIDE 9

Generating “tighter” relaxations

Would like to make our relaxations less relaxed. Various hierarchies give increasingly powerful programs at different levels (rounds), starting from a basic relaxation. 00 01 10 11

slide-10
SLIDE 10

Generating “tighter” relaxations

Would like to make our relaxations less relaxed. Various hierarchies give increasingly powerful programs at different levels (rounds), starting from a basic relaxation. 00 01 10 11

slide-11
SLIDE 11

Generating “tighter” relaxations

Would like to make our relaxations less relaxed. Various hierarchies give increasingly powerful programs at different levels (rounds), starting from a basic relaxation. 00 01 10 11

slide-12
SLIDE 12

Generating “tighter” relaxations

Would like to make our relaxations less relaxed. Various hierarchies give increasingly powerful programs at different levels (rounds), starting from a basic relaxation. Powerful computational model capturing most known LP/SDP algorithms within constant number of levels. Does approximation get better a higher levels? 00 01 10 11

slide-13
SLIDE 13

LP/SDP Hierarchies

Various hierarchies studied in the Operations Research literature:

Lovász-Schrijver (LS, LS+) Sherali-Adams Lasserre

slide-14
SLIDE 14

LP/SDP Hierarchies

Various hierarchies studied in the Operations Research literature:

Lovász-Schrijver (LS, LS+) Sherali-Adams Lasserre

LS(1) LS(2)

. . .

SA(1) SA(2)

. . .

LS(1)

+

LS(2)

+

. . .

Las(1) Las(2)

. . .

slide-15
SLIDE 15

LP/SDP Hierarchies

Various hierarchies studied in the Operations Research literature:

Lovász-Schrijver (LS, LS+) Sherali-Adams Lasserre

LS(3) SA(3) Las(3) LS(1) LS(2)

. . .

SA(1) SA(2)

. . .

LS(1)

+

LS(2)

+

. . .

Las(1) Las(2)

. . .

slide-16
SLIDE 16

LP/SDP Hierarchies

Various hierarchies studied in the Operations Research literature:

Lovász-Schrijver (LS, LS+) Sherali-Adams Lasserre

LS(3) SA(3) Las(3) LS(1) LS(2)

. . .

SA(1) SA(2)

. . .

LS(1)

+

LS(2)

+

. . .

Las(1) Las(2)

. . . Can optimize over r th level in time nO(r). nth level is tight.

slide-17
SLIDE 17

Example: Souping up the Independent Set relaxation

maximize:

  • u

xu subject to: xu + xv ≤ 1 ∀ (u, v) ∈ E xu ∈ [0, 1] K

slide-18
SLIDE 18

Example: Souping up the Independent Set relaxation

maximize:

  • u

xu subject to: xu + xv ≤ 1 ∀ (u, v) ∈ E xu ∈ [0, 1]

  • u∈K

xu ≤ 1 K

slide-19
SLIDE 19

Example: Souping up the Independent Set relaxation

maximize:

  • u

xu subject to: xu + xv ≤ 1 ∀ (u, v) ∈ E xu ∈ [0, 1]

  • u∈K

xu ≤ 1 K

  • Implied by one level of LS+ hierarchy.
  • Polytime algorithm for Independent Set on perfect graphs [GLS

81].

slide-20
SLIDE 20

What Hierarchies want

Example: Maximum Independent Set for graph G = (V, E) minimize

  • u

xu subject to xu + xv ≤ 1 ∀ (u, v) ∈ E xu ∈ [0, 1] Hope: x1, . . . , xn is convex combination of 0/1 solutions.

slide-21
SLIDE 21

What Hierarchies want

Example: Maximum Independent Set for graph G = (V, E) minimize

  • u

xu subject to xu + xv ≤ 1 ∀ (u, v) ∈ E xu ∈ [0, 1] Hope: x1, . . . , xn is convex combination of 0/1 solutions.

1/3 1/3 1/3

=

1 3× 1

+

1 3× 1

+

1 3× 1

slide-22
SLIDE 22

What Hierarchies want

Example: Maximum Independent Set for graph G = (V, E) minimize

  • u

xu subject to xu + xv ≤ 1 ∀ (u, v) ∈ E xu ∈ [0, 1] Hope: x1, . . . , xn is marginal of distribution over 0/1 solutions.

1/3 1/3 1/3

=

1 3× 1

+

1 3× 1

+

1 3× 1

slide-23
SLIDE 23

What Hierarchies want

Example: Maximum Independent Set for graph G = (V, E) minimize

  • u

xu subject to xu + xv ≤ 1 ∀ (u, v) ∈ E xu ∈ [0, 1] Hope: x1, . . . , xn is marginal of distribution over 0/1 solutions.

1/3 1/3 1/3

=

1 3× 1

+

1 3× 1

+

1 3× 1

Hierarchies add variables for conditional/joint probabilities.

slide-24
SLIDE 24

The Sherali-Adams Hierarchy

slide-25
SLIDE 25

The Sherali-Adams Hierarchy

Start with a 0/1 integer linear program. Hope: Fractional (x1, . . . , xn) = E [(z1, . . . , zn)] for integral (z1, . . . , zn)

slide-26
SLIDE 26

The Sherali-Adams Hierarchy

Start with a 0/1 integer linear program. Hope: Fractional (x1, . . . , xn) = E [(z1, . . . , zn)] for integral (z1, . . . , zn) Add “big variables” XS for |S| ≤ r (think XS = E

  • i∈S zi
  • = P [All vars in S are 1])
slide-27
SLIDE 27

The Sherali-Adams Hierarchy

Start with a 0/1 integer linear program. Hope: Fractional (x1, . . . , xn) = E [(z1, . . . , zn)] for integral (z1, . . . , zn) Add “big variables” XS for |S| ≤ r (think XS = E

  • i∈S zi
  • = P [All vars in S are 1])

Constraints:

slide-28
SLIDE 28

The Sherali-Adams Hierarchy

Start with a 0/1 integer linear program. Hope: Fractional (x1, . . . , xn) = E [(z1, . . . , zn)] for integral (z1, . . . , zn) Add “big variables” XS for |S| ≤ r (think XS = E

  • i∈S zi
  • = P [All vars in S are 1])

Constraints:

  • i

aizi ≤ b E

  • i

aizi

  • · z5z7(1 − z9)

E [b · z5z7(1 − z9)]

slide-29
SLIDE 29

The Sherali-Adams Hierarchy

Start with a 0/1 integer linear program. Hope: Fractional (x1, . . . , xn) = E [(z1, . . . , zn)] for integral (z1, . . . , zn) Add “big variables” XS for |S| ≤ r (think XS = E

  • i∈S zi
  • = P [All vars in S are 1])

Constraints:

  • i

aizi ≤ b E

  • i

aizi

  • · z5z7(1 − z9)

E [b · z5z7(1 − z9)]

  • i

ai · (X{i,5,7} − X{i,5,7,9}) ≤ b · (X{5,7} − X{5,7,9}) LP on nr variables.

slide-30
SLIDE 30

Sherali-Adams ≈ Locally Consistent Distributions

slide-31
SLIDE 31

Sherali-Adams ≈ Locally Consistent Distributions

Using 0 ≤ z1 ≤ 1, 0 ≤ z2 ≤ 1 ≤ X{1,2} ≤ 1 ≤ X{1} − X{1,2} ≤ 1 ≤ X{2} − X{1,2} ≤ 1 ≤ 1 − X{1} − X{2} + X{1,2} ≤ 1

slide-32
SLIDE 32

Sherali-Adams ≈ Locally Consistent Distributions

Using 0 ≤ z1 ≤ 1, 0 ≤ z2 ≤ 1 ≤ X{1,2} ≤ 1 ≤ X{1} − X{1,2} ≤ 1 ≤ X{2} − X{1,2} ≤ 1 ≤ 1 − X{1} − X{2} + X{1,2} ≤ 1 X{1}, X{2}, X{1,2} define a distribution D({1, 2}) over {0, 1}2.

slide-33
SLIDE 33

Sherali-Adams ≈ Locally Consistent Distributions

Using 0 ≤ z1 ≤ 1, 0 ≤ z2 ≤ 1 ≤ X{1,2} ≤ 1 ≤ X{1} − X{1,2} ≤ 1 ≤ X{2} − X{1,2} ≤ 1 ≤ 1 − X{1} − X{2} + X{1,2} ≤ 1 X{1}, X{2}, X{1,2} define a distribution D({1, 2}) over {0, 1}2. D({1, 2, 3}) and D({1, 2, 4}) must agree with D({1, 2}).

slide-34
SLIDE 34

Sherali-Adams ≈ Locally Consistent Distributions

Using 0 ≤ z1 ≤ 1, 0 ≤ z2 ≤ 1 ≤ X{1,2} ≤ 1 ≤ X{1} − X{1,2} ≤ 1 ≤ X{2} − X{1,2} ≤ 1 ≤ 1 − X{1} − X{2} + X{1,2} ≤ 1 X{1}, X{2}, X{1,2} define a distribution D({1, 2}) over {0, 1}2. D({1, 2, 3}) and D({1, 2, 4}) must agree with D({1, 2}). SA(r) = ⇒ LCD(r). If each constraint has at most k vars, LCD(r+k) = ⇒ SA(r)

slide-35
SLIDE 35

The Lasserre Hierarchy

slide-36
SLIDE 36

The Lasserre Hierarchy

Start with a 0/1 integer quadratic program.

slide-37
SLIDE 37

The Lasserre Hierarchy

Start with a 0/1 integer quadratic program. Think “big" variables ZS =

i∈S zi.

slide-38
SLIDE 38

The Lasserre Hierarchy

Start with a 0/1 integer quadratic program. Think “big" variables ZS =

i∈S zi.

Associated psd matrix Y (moment matrix)

YS1,S2 = E

  • ZS1 · ZS2
  • = E

 

  • i∈S1∪S2

zi  

slide-39
SLIDE 39

The Lasserre Hierarchy

Start with a 0/1 integer quadratic program. Think “big" variables ZS =

i∈S zi.

Associated psd matrix Y (moment matrix)

YS1,S2 = E

  • ZS1 · ZS2
  • = E

 

  • i∈S1∪S2

zi  

slide-40
SLIDE 40

The Lasserre Hierarchy

Start with a 0/1 integer quadratic program. Think “big" variables ZS =

i∈S zi.

Associated psd matrix Y (moment matrix)

YS1,S2 = E

  • ZS1 · ZS2
  • = E

 

  • i∈S1∪S2

zi   = P[All vars in S1 ∪ S2 are 1]

slide-41
SLIDE 41

The Lasserre Hierarchy

Start with a 0/1 integer quadratic program. Think “big" variables ZS =

i∈S zi.

Associated psd matrix Y (moment matrix)

YS1,S2 = E

  • ZS1 · ZS2
  • = E

 

  • i∈S1∪S2

zi   = P[All vars in S1 ∪ S2 are 1]

(Y 0) + original constraints + consistency constraints.

slide-42
SLIDE 42

The Lasserre hierarchy (constraints)

Y is psd. (i.e. find vectors US satisfying YS1,S2 =

  • US1, US2
  • )
slide-43
SLIDE 43

The Lasserre hierarchy (constraints)

Y is psd. (i.e. find vectors US satisfying YS1,S2 =

  • US1, US2
  • )

YS1,S2 only depends on S1 ∪ S2. (YS1,S2 = P[All vars in S1 ∪ S2 are 1])

slide-44
SLIDE 44

The Lasserre hierarchy (constraints)

Y is psd. (i.e. find vectors US satisfying YS1,S2 =

  • US1, US2
  • )

YS1,S2 only depends on S1 ∪ S2. (YS1,S2 = P[All vars in S1 ∪ S2 are 1]) Original quadratic constraints as inner products.

SDP for Independent Set

maximize

  • i∈V
  • U{i}
  • 2

subject to

  • U{i}, U{j}
  • = 0

∀ (i, j) ∈ E

  • US1, US2
  • =
  • US3, US4
  • ∀ S1 ∪ S2 = S3 ∪ S4
  • US1, US2
  • ∈ [0, 1]

∀S1, S2

slide-45
SLIDE 45

The “Mixed” hierarchy

Motivated by [Raghavendra 08]. Used by [CS 08] for Hypergraph Independent Set. Captures what we actually know how to use about Lasserre solutions.

slide-46
SLIDE 46

The “Mixed” hierarchy

Motivated by [Raghavendra 08]. Used by [CS 08] for Hypergraph Independent Set. Captures what we actually know how to use about Lasserre solutions. Level r has Variables XS for |S| ≤ r and all Sherali-Adams constraints. Vectors U0, U1, . . . , Un satisfying Ui, Uj = X{i,j}, U0, Ui = X{i} and |U0| = 1.

slide-47
SLIDE 47

Hands-on: Deriving some constraints

slide-48
SLIDE 48

The triangle inequality

|Ui − Uj|2 + |Uj − Uk|2 ≥ |Ui − Uk|2 is equivalent to Ui − Uj, Uk − Uj ≥ 0

slide-49
SLIDE 49

The triangle inequality

|Ui − Uj|2 + |Uj − Uk|2 ≥ |Ui − Uk|2 is equivalent to Ui − Uj, Uk − Uj ≥ 0 Mix(3) = ⇒ ∃ distribution on zi, zj, zk such that E[zi · zj] = Ui, Uj (and so on).

slide-50
SLIDE 50

The triangle inequality

|Ui − Uj|2 + |Uj − Uk|2 ≥ |Ui − Uk|2 is equivalent to Ui − Uj, Uk − Uj ≥ 0 Mix(3) = ⇒ ∃ distribution on zi, zj, zk such that E[zi · zj] = Ui, Uj (and so on). For all integer solutions (zi − zj) · (zk − zj) ≥ 0.

slide-51
SLIDE 51

The triangle inequality

|Ui − Uj|2 + |Uj − Uk|2 ≥ |Ui − Uk|2 is equivalent to Ui − Uj, Uk − Uj ≥ 0 Mix(3) = ⇒ ∃ distribution on zi, zj, zk such that E[zi · zj] = Ui, Uj (and so on). For all integer solutions (zi − zj) · (zk − zj) ≥ 0. ∴ Ui − Uj, Uk − Uj = E [(zi − zj) · (zk − zj)] ≥ 0

slide-52
SLIDE 52

“Clique constraints” for Independent Set

For every clique K in a graph, adding the constraint

  • i∈K

xi ≤ 1 makes the independent set LP tight for perfect graphs.

slide-53
SLIDE 53

“Clique constraints” for Independent Set

For every clique K in a graph, adding the constraint

  • i∈K

xi ≤ 1 makes the independent set LP tight for perfect graphs. Too many constraints, but all implied by one level of the mixed hierarchy.

slide-54
SLIDE 54

“Clique constraints” for Independent Set

For every clique K in a graph, adding the constraint

  • i∈K

xi ≤ 1 makes the independent set LP tight for perfect graphs. Too many constraints, but all implied by one level of the mixed hierarchy. For i, j ∈ K, Ui, Uj = 0. Also, ∀i U0, Ui = |Ui|2 = xi. By Pythagoras,

  • i∈K
  • U0, Ui

|Ui| 2 ≤ |U0|2 = 1 = ⇒

  • i∈B

x2

i

xi ≤ 1. Derived by Lovász using the ϑ-function.

slide-55
SLIDE 55

The Lovász-Schrijver Hierarchy

slide-56
SLIDE 56

The Lovász-Schrijver Hierarchy

Start with a 0/1 integer program and a relaxation P. Define tigher relaxation LS(P).

slide-57
SLIDE 57

The Lovász-Schrijver Hierarchy

Start with a 0/1 integer program and a relaxation P. Define tigher relaxation LS(P). Hope: Fractional (x1, . . . , xn) = E [(z1, . . . , zn)] for integral (z1, . . . , zn)

slide-58
SLIDE 58

The Lovász-Schrijver Hierarchy

Start with a 0/1 integer program and a relaxation P. Define tigher relaxation LS(P). Hope: Fractional (x1, . . . , xn) = E [(z1, . . . , zn)] for integral (z1, . . . , zn) Restriction: x = (x1, . . . , xn) ∈ LS(P) if ∃Y satisfying (think Yij = E [zizj] = P [zi ∧ zj]) Y = Y T Yii = xi ∀i Yi xi ∈ P, x − Yi 1 − xi ∈ P ∀i Y 0

slide-59
SLIDE 59

The Lovász-Schrijver Hierarchy

Start with a 0/1 integer program and a relaxation P. Define tigher relaxation LS(P). Hope: Fractional (x1, . . . , xn) = E [(z1, . . . , zn)] for integral (z1, . . . , zn) Restriction: x = (x1, . . . , xn) ∈ LS(P) if ∃Y satisfying (think Yij = E [zizj] = P [zi ∧ zj]) Y = Y T Yii = xi ∀i Yi xi ∈ P, x − Yi 1 − xi ∈ P ∀i Y 0 Above is an LP (SDP) in n2 + n variables.

slide-60
SLIDE 60

Lovász-Schrijver in action

r th level optimizes over distributions conditioned on r variables.

1/3 1/3 1/3 1/3 1/3 1/3

slide-61
SLIDE 61

Lovász-Schrijver in action

r th level optimizes over distributions conditioned on r variables.

1/3 1/3 1/3 1/3 1/3 1/3

slide-62
SLIDE 62

Lovász-Schrijver in action

r th level optimizes over distributions conditioned on r variables.

1/3 1/3 1/3 1/3 1/3 1/3 1/2 1/2 1/2 1/2 1/2 1

2/3× 1/3×

slide-63
SLIDE 63

Lovász-Schrijver in action

r th level optimizes over distributions conditioned on r variables.

1/3 1/3 1/3 1/3 1/3 1/3 1/2 1/2 1/2 1/2 1/2 1

2/3× 1/3×

slide-64
SLIDE 64

Lovász-Schrijver in action

r th level optimizes over distributions conditioned on r variables.

1/3 1/3 1/3 1/3 1/3 1/3 1/2 1/2 1/2 1/2 1/2 1

2/3× 1/3× 1/2×

1

slide-65
SLIDE 65

Lovász-Schrijver in action

r th level optimizes over distributions conditioned on r variables.

1/3 1/3 1/3 1/3 1/3 1/3 1/2 1/2 1/2 1/2 1/2 1

2/3× 1/3× 1/2×

1

1/2×

? ? ? ? ?

slide-66
SLIDE 66

And if you just woke up . . .

slide-67
SLIDE 67

And if you just woke up . . .

LS(1) LS(2)

. . .

SA(1) SA(2)

. . .

LS(1)

+

LS(2)

+

. . .

Las(1) Las(2)

. . .

Mix(1) Mix(2)

. . .

slide-68
SLIDE 68

And if you just woke up . . .

LS(1) LS(2)

. . .

Yi xi , x − Yi 1 − xi

SA(1) SA(2)

. . .

XS

LS(1)

+

LS(2)

+

. . .

Y 0

Las(1) Las(2)

. . .

US

Mix(1) Mix(2)

. . .

SA + Ui

slide-69
SLIDE 69

Algorithmic Applications

Many known LP/SDP relaxations captured by 2-3 levels. [Chlamtac 07]: Explicitly used level-3 Lasserre SDP for graph coloring. [CS 08]: Algorithms using Mixed and Lasserre hierarchies for hypergraph independent set (guarantee improves with more levels). [KKMN 10]: Hierarchies yield a PTAS for Knapsack. [BRS 11, GS 11]: Algorithms for Unique Games using nǫ levels of Lassere.

slide-70
SLIDE 70

Lower bound techniques

Expansion in CSP instances (Proof Complexity)

slide-71
SLIDE 71

Lower bound techniques

Expansion in CSP instances (Proof Complexity) Reductions

slide-72
SLIDE 72

Lower bound techniques

Expansion in CSP instances (Proof Complexity) Reductions [ABLT 06, STT 07, dlVKM 07, CMM 09]: Distributions from local probabilistic processes. [Charikar 02, GMPT 07, BCGM 10]: Polynomial tensoring. [RS 09, KS 09]: Higher level distributions from level-1 vectors.

slide-73
SLIDE 73

Integrality Gaps for Expanding CSPs

slide-74
SLIDE 74

CSP Expansion

MAX k-CSP: m constraints on k-tuples of (n) boolean variables. Satisfy maximum. e.g. MAX 3-XOR (linear equations mod 2) z1 + z2 + z3 = 0 z3 + z4 + z5 = 1 · · ·

slide-75
SLIDE 75

CSP Expansion

MAX k-CSP: m constraints on k-tuples of (n) boolean variables. Satisfy maximum. e.g. MAX 3-XOR (linear equations mod 2) z1 + z2 + z3 = 0 z3 + z4 + z5 = 1 · · · Expansion: Every set S of constraints involves at least β|S| variables (for |S| < αm).

slide-76
SLIDE 76

CSP Expansion

MAX k-CSP: m constraints on k-tuples of (n) boolean variables. Satisfy maximum. e.g. MAX 3-XOR (linear equations mod 2) z1 + z2 + z3 = 0 z3 + z4 + z5 = 1 · · · Expansion: Every set S of constraints involves at least β|S| variables (for |S| < αm).

Cm

. . .

C1 zn

. . .

z1

slide-77
SLIDE 77

CSP Expansion

MAX k-CSP: m constraints on k-tuples of (n) boolean variables. Satisfy maximum. e.g. MAX 3-XOR (linear equations mod 2) z1 + z2 + z3 = 0 z3 + z4 + z5 = 1 · · · Expansion: Every set S of constraints involves at least β|S| variables (for |S| < αm).

Cm

. . .

C1 zn

. . .

z1

slide-78
SLIDE 78

CSP Expansion

MAX k-CSP: m constraints on k-tuples of (n) boolean variables. Satisfy maximum. e.g. MAX 3-XOR (linear equations mod 2) z1 + z2 + z3 = 0 z3 + z4 + z5 = 1 · · · Expansion: Every set S of constraints involves at least β|S| variables (for |S| < αm).

Cm

. . .

C1 zn

. . .

z1

In fact, γ|S| variables appearing in only one constraint in S.

slide-79
SLIDE 79

CSP Expansion

MAX k-CSP: m constraints on k-tuples of (n) boolean variables. Satisfy maximum. e.g. MAX 3-XOR (linear equations mod 2) z1 + z2 + z3 = 0 z3 + z4 + z5 = 1 · · · Expansion: Every set S of constraints involves at least β|S| variables (for |S| < αm).

Cm

. . .

C1 zn

. . .

z1

In fact, γ|S| variables appearing in only one constraint in S. Used extensively in proof complexity e.g. [BW01], [BGHMP03]. For LS+ by [AAT04].

slide-80
SLIDE 80

Sherali-Adams LP for CSPs

Variables: X(S,α) for |S| ≤ t, partial assignments α ∈ {0, 1}S maximize

m

  • i=1
  • α∈{0,1}Ti

Ci(α)·X(Ti ,α) subject to X(S∪{i},α◦0) + X(S∪{i},α◦1) = X(S,α) ∀i / ∈ S X(S,α) ≥ 0 X(∅,∅) = 1

slide-81
SLIDE 81

Sherali-Adams LP for CSPs

Variables: X(S,α) for |S| ≤ t, partial assignments α ∈ {0, 1}S maximize

m

  • i=1
  • α∈{0,1}Ti

Ci(α)·X(Ti ,α) subject to X(S∪{i},α◦0) + X(S∪{i},α◦1) = X(S,α) ∀i / ∈ S X(S,α) ≥ 0 X(∅,∅) = 1

X(S,α) ∼ P[Vars in S assigned according to α]

slide-82
SLIDE 82

Sherali-Adams LP for CSPs

Variables: X(S,α) for |S| ≤ t, partial assignments α ∈ {0, 1}S maximize

m

  • i=1
  • α∈{0,1}Ti

Ci(α)·X(Ti ,α) subject to X(S∪{i},α◦0) + X(S∪{i},α◦1) = X(S,α) ∀i / ∈ S X(S,α) ≥ 0 X(∅,∅) = 1

X(S,α) ∼ P[Vars in S assigned according to α] Need distributions D(S) such that D(S1), D(S2) agree on S1 ∩ S2.

slide-83
SLIDE 83

Sherali-Adams LP for CSPs

Variables: X(S,α) for |S| ≤ t, partial assignments α ∈ {0, 1}S maximize

m

  • i=1
  • α∈{0,1}Ti

Ci(α)·X(Ti ,α) subject to X(S∪{i},α◦0) + X(S∪{i},α◦1) = X(S,α) ∀i / ∈ S X(S,α) ≥ 0 X(∅,∅) = 1

X(S,α) ∼ P[Vars in S assigned according to α] Need distributions D(S) such that D(S1), D(S2) agree on S1 ∩ S2. Distributions should “locally look like" supported on satisfying assignments.

slide-84
SLIDE 84

Local Satisfiability

C1 C2 C3 z1 z2 z3 z4 z5 z6

  • Take γ = 0.9
  • Can show any three 3-XOR constraints are

simultaneously satisfiable.

slide-85
SLIDE 85

Local Satisfiability

C1 C2 C3 z1 z2 z3 z4 z5 z6

  • Take γ = 0.9
  • Can show any three 3-XOR constraints are

simultaneously satisfiable. Ez1...z6 [C1(z1, z2, z3) · C2(z3, z4, z5) · C3(z4, z5, z6)]

slide-86
SLIDE 86

Local Satisfiability

C1 C2 C3 z1 z2 z3 z4 z5 z6

  • Take γ = 0.9
  • Can show any three 3-XOR constraints are

simultaneously satisfiable. Ez1...z6 [C1(z1, z2, z3) · C2(z3, z4, z5) · C3(z4, z5, z6)] = Ez2...z6 [C2(z3, z4, z5) · C3(z4, z5, z6) · Ez1 [C1(z1, z2, z3)]]

slide-87
SLIDE 87

Local Satisfiability

C1 C2 C3 z1 z2 z3 z4 z5 z6

  • Take γ = 0.9
  • Can show any three 3-XOR constraints are

simultaneously satisfiable. Ez1...z6 [C1(z1, z2, z3) · C2(z3, z4, z5) · C3(z4, z5, z6)] = Ez2...z6 [C2(z3, z4, z5) · C3(z4, z5, z6) · Ez1 [C1(z1, z2, z3)]] = Ez4,z5,z6 [C3(z4, z5, z6) · Ez3 [C2(z3, z4, z5)] · (1/2)]

slide-88
SLIDE 88

Local Satisfiability

C1 C2 C3 z1 z2 z3 z4 z5 z6

  • Take γ = 0.9
  • Can show any three 3-XOR constraints are

simultaneously satisfiable. Ez1...z6 [C1(z1, z2, z3) · C2(z3, z4, z5) · C3(z4, z5, z6)] = Ez2...z6 [C2(z3, z4, z5) · C3(z4, z5, z6) · Ez1 [C1(z1, z2, z3)]] = Ez4,z5,z6 [C3(z4, z5, z6) · Ez3 [C2(z3, z4, z5)] · (1/2)] = 1/8

slide-89
SLIDE 89

Local Satisfiability

C1 C2 C3 z1 z2 z3 z4 z5 z6

  • Take γ = 0.9
  • Can show any three 3-XOR constraints are

simultaneously satisfiable.

  • Can take γ ≈ (k − 2) and any αn constraints.
  • Just require E[C(z1, . . . , zk)] over any k − 2

vars to be constant. Ez1...z6 [C1(z1, z2, z3) · C2(z3, z4, z5) · C3(z4, z5, z6)] = Ez2...z6 [C2(z3, z4, z5) · C3(z4, z5, z6) · Ez1 [C1(z1, z2, z3)]] = Ez4,z5,z6 [C3(z4, z5, z6) · Ez3 [C2(z3, z4, z5)] · (1/2)] = 1/8

slide-90
SLIDE 90

Obtaining integrality gaps for CSPs [BGMT 09]

Cm

. . .

C1 zn

. . .

z1 Want to define distribution D(S) for set S of variables.

slide-91
SLIDE 91

Obtaining integrality gaps for CSPs [BGMT 09]

Cm

. . .

C1 zn

. . .

z1 Want to define distribution D(S) for set S of variables.

slide-92
SLIDE 92

Obtaining integrality gaps for CSPs [BGMT 09]

Cm

. . .

C1 zn

. . .

z1 Want to define distribution D(S) for set S of variables.

slide-93
SLIDE 93

Obtaining integrality gaps for CSPs [BGMT 09]

Cm

. . .

C1 zn

. . .

z1 Want to define distribution D(S) for set S of variables. Find set of constraints C such that G − C − S remains expanding. D(S) = uniform over assignments satisfying C

slide-94
SLIDE 94

Obtaining integrality gaps for CSPs [BGMT 09]

Cm

. . .

C1 zn

. . .

z1 Want to define distribution D(S) for set S of variables. Find set of constraints C such that G − C − S remains expanding. D(S) = uniform over assignments satisfying C Remaining constraints “independent" of this assignment. Gives optimal integrality gaps for Ω(n) levels in the mixed hierarchy.

slide-95
SLIDE 95

Vectors for Linear CSPs

slide-96
SLIDE 96

A “new look” Lasserre

Start with a {−1, 1} quadratic integer program. (z1, . . . , zn) → ((−1)z1, . . . , (−1)zn)

slide-97
SLIDE 97

A “new look” Lasserre

Start with a {−1, 1} quadratic integer program. (z1, . . . , zn) → ((−1)z1, . . . , (−1)zn) Define big variables ˜ ZS =

i∈S(−1)zi.

slide-98
SLIDE 98

A “new look” Lasserre

Start with a {−1, 1} quadratic integer program. (z1, . . . , zn) → ((−1)z1, . . . , (−1)zn) Define big variables ˜ ZS =

i∈S(−1)zi.

Consider the psd matrix ˜ Y ˜ YS1,S2 = E

  • ˜

ZS1 · ˜ ZS2

  • = E

 

  • i∈S1∆S2

(−1)zi  

slide-99
SLIDE 99

A “new look” Lasserre

Start with a {−1, 1} quadratic integer program. (z1, . . . , zn) → ((−1)z1, . . . , (−1)zn) Define big variables ˜ ZS =

i∈S(−1)zi.

Consider the psd matrix ˜ Y ˜ YS1,S2 = E

  • ˜

ZS1 · ˜ ZS2

  • = E

 

  • i∈S1∆S2

(−1)zi   Write program for inner products of vectors WS s.t. ˜ YS1,S2 = WS1, WS2

slide-100
SLIDE 100

Gaps for 3-XOR

SDP for MAX 3-XOR

maximize

  • Ci ≡(zi1 +zi2 +zi3 =bi )

1 + (−1)bi W{i1,i2,i3}, W∅

  • 2

subject to

  • WS1, WS2
  • =
  • WS3, WS4
  • ∀ S1∆S2 = S3∆S4

|WS| = 1 ∀S, |S| ≤ r

slide-101
SLIDE 101

Gaps for 3-XOR

SDP for MAX 3-XOR

maximize

  • Ci ≡(zi1 +zi2 +zi3 =bi )

1 + (−1)bi W{i1,i2,i3}, W∅

  • 2

subject to

  • WS1, WS2
  • =
  • WS3, WS4
  • ∀ S1∆S2 = S3∆S4

|WS| = 1 ∀S, |S| ≤ r

[Schoenebeck’08]: If width 2r resolution does not derive contradiction, then SDP value =1 after r levels of Lasserre.

slide-102
SLIDE 102

Gaps for 3-XOR

SDP for MAX 3-XOR

maximize

  • Ci ≡(zi1 +zi2 +zi3 =bi )

1 + (−1)bi W{i1,i2,i3}, W∅

  • 2

subject to

  • WS1, WS2
  • =
  • WS3, WS4
  • ∀ S1∆S2 = S3∆S4

|WS| = 1 ∀S, |S| ≤ r

[Schoenebeck’08]: If width 2r resolution does not derive contradiction, then SDP value =1 after r levels of Lasserre. Expansion guarantees there are no width 2r contradictions.

slide-103
SLIDE 103

Gaps for 3-XOR

SDP for MAX 3-XOR

maximize

  • Ci ≡(zi1 +zi2 +zi3 =bi )

1 + (−1)bi W{i1,i2,i3}, W∅

  • 2

subject to

  • WS1, WS2
  • =
  • WS3, WS4
  • ∀ S1∆S2 = S3∆S4

|WS| = 1 ∀S, |S| ≤ r

[Schoenebeck’08]: If width 2r resolution does not derive contradiction, then SDP value =1 after r levels of Lasserre. Expansion guarantees there are no width 2r contradictions. Used by [FO 06], [STT 07] for LS+ hierarchy.

slide-104
SLIDE 104

Schonebeck’s construction

slide-105
SLIDE 105

Schonebeck’s construction

z1 + z2 + z3 = 1 mod 2 = ⇒ (−1)z1+z2 = −(−1)z3 = ⇒ W{1,2} = −W{3}

slide-106
SLIDE 106

Schonebeck’s construction

z1 + z2 + z3 = 1 mod 2 = ⇒ (−1)z1+z2 = −(−1)z3 = ⇒ W{1,2} = −W{3} Equations of width 2r divide |S| ≤ r into equivalence classes. Choose

  • rthogonal eC for each class C.
slide-107
SLIDE 107

Schonebeck’s construction

z1 + z2 + z3 = 1 mod 2 = ⇒ (−1)z1+z2 = −(−1)z3 = ⇒ W{1,2} = −W{3} Equations of width 2r divide |S| ≤ r into equivalence classes. Choose

  • rthogonal eC for each class C.

No contradictions ensure each S ∈ C can be uniquely assigned ±eC.

WS1 = eC1 WS2 = −eC2 WS3 = eC2

slide-108
SLIDE 108

Schonebeck’s construction

z1 + z2 + z3 = 1 mod 2 = ⇒ (−1)z1+z2 = −(−1)z3 = ⇒ W{1,2} = −W{3} Equations of width 2r divide |S| ≤ r into equivalence classes. Choose

  • rthogonal eC for each class C.

No contradictions ensure each S ∈ C can be uniquely assigned ±eC. Relies heavily on constraints being linear equations.

WS1 = eC1 WS2 = −eC2 WS3 = eC2

slide-109
SLIDE 109

Reductions

slide-110
SLIDE 110

Spreading the hardness around (Reductions) [T]

If problem A reduces to B, can we say Integrality Gap for A = ⇒ Integrality Gap for B?

slide-111
SLIDE 111

Spreading the hardness around (Reductions) [T]

If problem A reduces to B, can we say Integrality Gap for A = ⇒ Integrality Gap for B? Reductions are (often) local algorithms.

slide-112
SLIDE 112

Spreading the hardness around (Reductions) [T]

If problem A reduces to B, can we say Integrality Gap for A = ⇒ Integrality Gap for B? Reductions are (often) local algorithms. Reduction from integer program A to integer program B. Each variable z′

i of B is a boolean function of few (say 5) variables

zi1, . . . , zi5 of A.

slide-113
SLIDE 113

Spreading the hardness around (Reductions) [T]

If problem A reduces to B, can we say Integrality Gap for A = ⇒ Integrality Gap for B? Reductions are (often) local algorithms. Reduction from integer program A to integer program B. Each variable z′

i of B is a boolean function of few (say 5) variables

zi1, . . . , zi5 of A. To show: If A has good vector solution, so does B.

slide-114
SLIDE 114

Spreading the hardness around (Reductions) [T]

If problem A reduces to B, can we say Integrality Gap for A = ⇒ Integrality Gap for B? Reductions are (often) local algorithms. Reduction from integer program A to integer program B. Each variable z′

i of B is a boolean function of few (say 5) variables

zi1, . . . , zi5 of A. To show: If A has good vector solution, so does B. Question posed in [AAT 04]. First done by [KV 05] from Unique Games to Sparsest Cut.

slide-115
SLIDE 115

Integrality Gaps for Independent Set

FGLSS: Reduction from MAX k-CSP to Independent Set in graph GΦ. z1 + z2 + z3 = 1

001 010 100 111

z3 + z4 + z5 = 0

110 011 000 101

slide-116
SLIDE 116

Integrality Gaps for Independent Set

FGLSS: Reduction from MAX k-CSP to Independent Set in graph GΦ. z1 + z2 + z3 = 1

001 010 100 111

z3 + z4 + z5 = 0

110 011 000 101

Need vectors for subsets of vertices in the GΦ.

slide-117
SLIDE 117

Integrality Gaps for Independent Set

FGLSS: Reduction from MAX k-CSP to Independent Set in graph GΦ. z1 + z2 + z3 = 1

001 010 100 111

z3 + z4 + z5 = 0

110 011 000 101

Need vectors for subsets of vertices in the GΦ. Every vertex (or set of vertices) in GΦ is an indicator function!

slide-118
SLIDE 118

Integrality Gaps for Independent Set

FGLSS: Reduction from MAX k-CSP to Independent Set in graph GΦ. z1 + z2 + z3 = 1

001 010 100 111

z3 + z4 + z5 = 0

110 011 000 101

Need vectors for subsets of vertices in the GΦ. Every vertex (or set of vertices) in GΦ is an indicator function!

U{(z1,z2,z3)=(0,0,1)} = 1 8 (W∅ + W{1} + W{2} − W{3} + W{1,2} − W{2,3} − W{1,3} − W{1,2,3})

slide-119
SLIDE 119

Graph Products

v1 w1

×

v2 w2

×

v3 w3

slide-120
SLIDE 120

Graph Products

v1 w1

×

v2 w2

×

v3 w3

slide-121
SLIDE 121

Graph Products

v1 w1

×

v2 w2

×

v3 w3 U{(v1,v2,v3)} = ?

slide-122
SLIDE 122

Graph Products

v1 w1

×

v2 w2

×

v3 w3 U{(v1,v2,v3)} = U{v1} ⊗ U{v2} ⊗ U{v3}

slide-123
SLIDE 123

Graph Products

v1 w1

×

v2 w2

×

v3 w3 U{(v1,v2,v3)} = U{v1} ⊗ U{v2} ⊗ U{v3} Similar transformation for sets (project to each copy of G).

slide-124
SLIDE 124

Graph Products

v1 w1

×

v2 w2

×

v3 w3 U{(v1,v2,v3)} = U{v1} ⊗ U{v2} ⊗ U{v3} Similar transformation for sets (project to each copy of G). Intuition: Independent set in product graph is product of independent sets in G.

slide-125
SLIDE 125

Graph Products

v1 w1

×

v2 w2

×

v3 w3 U{(v1,v2,v3)} = U{v1} ⊗ U{v2} ⊗ U{v3} Similar transformation for sets (project to each copy of G). Intuition: Independent set in product graph is product of independent sets in G. Together give a gap of

n 2O(√

log n log log n) .

slide-126
SLIDE 126

A few problems

slide-127
SLIDE 127

Problem 1: Lasserre Gaps

Show an integrality gap of 2 − ǫ for Vertex Cover, even for O(1) levels of the Lasserre hierarchy. Obtain integrality gaps Unique Games (and Small-Set Expansion) Gaps for O((log log n)1/4) levels of mixed hierarchy were

  • btained by [RS 09] and [KS 09].

Extension to Lasserre?

slide-128
SLIDE 128

Problem 2: Generalize Schoenebeck’s technique

slide-129
SLIDE 129

Problem 2: Generalize Schoenebeck’s technique

Technique seems specialized for linear equations. Breaks down even if there are few local contradictions (which doesn’t rule out a gap).

slide-130
SLIDE 130

Problem 2: Generalize Schoenebeck’s technique

Technique seems specialized for linear equations. Breaks down even if there are few local contradictions (which doesn’t rule out a gap). We know mixed hierarchy gaps for other CSPs: know local distributions, but not vectors.

slide-131
SLIDE 131

Problem 2: Generalize Schoenebeck’s technique

Technique seems specialized for linear equations. Breaks down even if there are few local contradictions (which doesn’t rule out a gap). We know mixed hierarchy gaps for other CSPs: know local distributions, but not vectors. What extra constraints do vectors capture?

slide-132
SLIDE 132

Thank You Questions?