From Weak to Strong LP Gaps for all CSPs Mrinalkanti Ghosh joint - - PowerPoint PPT Presentation

from weak to strong lp gaps for all csps
SMART_READER_LITE
LIVE PREVIEW

From Weak to Strong LP Gaps for all CSPs Mrinalkanti Ghosh joint - - PowerPoint PPT Presentation

From Weak to Strong LP Gaps for all CSPs Mrinalkanti Ghosh joint work with: Madhur Tulsiani MAX k-CSP - n variables - m constraints MAX k-CSP - n variables taking boolean values. - m constraints: each is a k-ary boolean predicate. - Satisfy


slide-1
SLIDE 1

From Weak to Strong LP Gaps for all CSPs

Mrinalkanti Ghosh

joint work with:

Madhur Tulsiani

slide-2
SLIDE 2

MAX k-CSP

  • n variables
  • m constraints
slide-3
SLIDE 3

MAX k-CSP

  • n variables taking boolean values.
  • m constraints: each is a k-ary boolean predicate.
  • Satisfy as many as possible.
slide-4
SLIDE 4

MAX k-CSP

  • n variables taking boolean values.
  • m constraints: each is a k-ary boolean predicate.
  • Satisfy as many as possible.

Max-3-SAT x1 ∨ x22 ∨ x19 x3 ∨ x9 ∨ x23 x5 ∨ x7 ∨ x9 . . .

slide-5
SLIDE 5

MAX k-CSP

  • n variables taking boolean values.
  • m constraints: each is a k-ary boolean predicate.
  • Satisfy as many as possible.

Max-3-SAT x1 ∨ x22 ∨ x19 x3 ∨ x9 ∨ x23 x5 ∨ x7 ∨ x9 . . . Max-Cut x6 x5 x7 x3 x4 x2 x1 x1 = x2 x2 = x5 x3 = x4 . . .

slide-6
SLIDE 6

MAX k-CSP

  • n variables taking boolean values.
  • m constraints: each is a k-ary boolean predicate.
  • Satisfy as many as possible.

Max-3-SAT x1 ∨ x22 ∨ x19 x3 ∨ x9 ∨ x23 x5 ∨ x7 ∨ x9 . . . Max-Cut x6 x5 x7 x3 x4 x2 x1 x1 = x2 x2 = x5 x3 = x4 . . . Approximation Problem: Approximate the fraction of constraints simultaneously satisfiable.

slide-7
SLIDE 7

MAX k-CSP

  • n variables taking values in some finite domains.
  • m constraints: each is a non-negative k-ary function.
  • Satisfy as many as possible.

Max-3-SAT x1 ∨ x22 ∨ x19 x3 ∨ x9 ∨ x23 x5 ∨ x7 ∨ x9 . . . Max-Cut x6 x5 x7 x3 x4 x2 x1 x1 = x2 x2 = x5 x3 = x4 . . . Approximation Problem: Approximate the fraction of constraints simultaneously satisfiable.

slide-8
SLIDE 8

CSPs and Relaxations

MAX k-CSP (f): for i-th constraint, let SCi := (xi1, · · · , xik). Then: Ci ≡ f (xi1 + bi1, · · · , xik + bik) ≡

  • α∈{0,1}

SCi

f (α + bCi) · x(SCi ,α) , with x(SCi ,α) = indicator of assignment of α to SCi.

slide-9
SLIDE 9

CSPs and Relaxations

MAX k-CSP (f): for i-th constraint, let SCi := (xi1, · · · , xik). Then: Ci ≡ f (xi1 + bi1, · · · , xik + bik) ≡

  • α∈{0,1}

SCi

f (α + bCi) · x(SCi ,α) , with x(SCi ,α) = indicator of assignment of α to SCi.

  • α∈{0,1}SC

α(i)=b

x(SC,α) = x(i,b) ∀C ∈ Φ, i ∈ SC, b ∈ {0, 1}

  • b∈{0,1}

x(i,b) = 1 ∀i ∈ [n] x(S,α) ≥ 0

slide-10
SLIDE 10

CSPs and Relaxations

MAX k-CSP (f): for i-th constraint, let SCi := (xi1, · · · , xik). Then: Ci ≡ f (xi1 + bi1, · · · , xik + bik) ≡

  • α∈{0,1}

SCi

f (α + bCi) · x(SCi ,α) , with x(SCi ,α) = indicator of assignment of α to SCi. maximize EC∈Φ

  • α∈{0,1}SC f (α + bC) · x(SC,α)
  • α∈{0,1}SC

α(i)=b

x(SC,α) = x(i,b) ∀C ∈ Φ, i ∈ SC, b ∈ {0, 1}

  • b∈{0,1}

x(i,b) = 1 ∀i ∈ [n] x(S,α) ≥ 0

slide-11
SLIDE 11

CSPs and Relaxations

MAX k-CSP (f): for i-th constraint, let SCi := (xi1, · · · , xik). Then: Ci ≡ f (xi1 + bi1, · · · , xik + bik) ≡

  • α∈{0,1}

SCi

f (α + bCi) · x(SCi ,α) , with x(SCi ,α) = indicator of assignment of α to SCi. maximize EC∈Φ

  • α∈{0,1}SC f (α + bC) · x(SC,α)
  • α∈{0,1}SC

α(i)=b

x(SC,α) = x(i,b) ∀C ∈ Φ, i ∈ SC, b ∈ {0, 1}

  • b∈{0,1}

x(i,b) = 1 ∀i ∈ [n] x(S,α) ≥ 0

#constraints = Θ

  • m · 2k
slide-12
SLIDE 12

Extended Formulation and Sherali-Adams Relaxation

  • Extended Formulation: Defined

by a feasible polytope P, and a way

  • f encoding instances Φ as a

(linear) objective function wΦ.

slide-13
SLIDE 13

Extended Formulation and Sherali-Adams Relaxation

  • Extended Formulation: Defined

by a feasible polytope P, and a way

  • f encoding instances Φ as a

(linear) objective function wΦ.

  • Optimize objective wΦ, x

(depending on Φ) over P.

slide-14
SLIDE 14

Extended Formulation and Sherali-Adams Relaxation

Image from [Fiorini-Rothvoss-Tiwari-11]

  • Extended Formulation: Defined

by a feasible polytope P, and a way

  • f encoding instances Φ as a

(linear) objective function wΦ.

  • Optimize objective wΦ, x

(depending on Φ) over P.

  • Introduce additional variables y.

Optimize over polytope P = {x | ∃y Ex + Fy = g, y ≥ 0} . Size equals #variables + #constraints.

slide-15
SLIDE 15

Extended Formulation and Sherali-Adams Relaxation

Image from [Fiorini-Rothvoss-Tiwari-11]

  • Extended Formulation: Defined

by a feasible polytope P, and a way

  • f encoding instances Φ as a

(linear) objective function wΦ.

  • Sherali-Adams: A Sherali-Adams
  • f level t is an Extended

Formulation with #variables =

n

t

· 2t.

slide-16
SLIDE 16

Extended Formulation and Sherali-Adams Relaxation

Image from [Fiorini-Rothvoss-Tiwari-11]

  • Extended Formulation: Defined

by a feasible polytope P, and a way

  • f encoding instances Φ as a

(linear) objective function wΦ.

  • Sherali-Adams: A Sherali-Adams
  • f level t is an Extended

Formulation with #variables =

n

t

· 2t.

  • Variables: x(S,α), |S| ≤ t,

α ∈ {0, 1}S.

slide-17
SLIDE 17

Extended Formulation and Sherali-Adams Relaxation

EF: SA: S T DS DT DS∩T

  • Extended Formulation: Defined

by a feasible polytope P, and a way

  • f encoding instances Φ as a

(linear) objective function wΦ.

  • Sherali-Adams: A Sherali-Adams
  • f level t is an Extended

Formulation with #variables =

n

t

· 2t.

  • Feasible point in SA(t): Family

{DS}|S|≤t of consistent distribution with DS a distribution on {0, 1}S.

slide-18
SLIDE 18

Extended Formulation and Sherali-Adams Relaxation

EF: SA: S T DS DT DS∩T Basic: C1 C2

  • Extended Formulation: Defined

by a feasible polytope P, and a way

  • f encoding instances Φ as a

(linear) objective function wΦ.

  • Sherali-Adams: A Sherali-Adams
  • f level t is an Extended

Formulation with #variables =

n

t

· 2t.

  • Feasible point in SA(t): Family

{DS}|S|≤t of consistent distribution with DS a distribution on {0, 1}S.

  • Similarly, for Basic LP solution.
slide-19
SLIDE 19

Result

slide-20
SLIDE 20

Result

slide-21
SLIDE 21

Result

Main Theorem: For all CSPs, if Ba- sic LP has integral- ity gap

  • f

(c, s) then for all ε > 0, there exist large enough instance(s) with integrality gap

  • f (c − ε, s + ε) for

SA( Oε(log n)).

slide-22
SLIDE 22

Result

With [Kothari- Meka-Raghavendra- 17]: For all CSPs, if Basic LP has (c, s) gap, then so does any LP Extended For- mulation

  • f

size n

O(log n).

Ignoring ε losses.

slide-23
SLIDE 23

Hard Instance

Basic: C1 C2 SA: S T DS DT DS∩T

slide-24
SLIDE 24

Hard Instance

Basic: C1 C2 SA: S T DS DT DS∩T Use the hard instance Φ0 of the basic relaxation as template to build the new hard instance on n variables and m = ∆ · n constraints.

slide-25
SLIDE 25

Hard Instance

#variables = n and #constraints = m = ∆ · n.

slide-26
SLIDE 26

Hard Instance

#variables = n and #constraints = m = ∆ · n. x1 Φ0 x2 Φ0 x3 Φ0 x4 Φ0 x5 Φ0 x6 Φ0 x7 Φ0 x8 Φ0 x9 Φ0 n/9 n/9 n/9 n/9 n/9 n/9 n/9 n/9 n/9

b1 b2 b3 b4 b5 b6 b7 b8 b9 - For each variable in Φ0, create

bucket with large number of variables.

slide-27
SLIDE 27

Hard Instance

#variables = n and #constraints = m = ∆ · n. x1 Φ0 x2 Φ0 x3 Φ0 x4 Φ0 x5 Φ0 x6 Φ0 x7 Φ0 x8 Φ0 x9 Φ0 n/9 n/9 n/9 n/9 n/9 n/9 n/9 n/9 n/9

b1 b2 b3 b4 b5 b6 b7 b8 b9 - For each variable in Φ0, create

bucket with large number of variables.

  • Independently, sample each

constraint as:

slide-28
SLIDE 28

Hard Instance

#variables = n and #constraints = m = ∆ · n. x1 Φ0 x2 Φ0 x3 Φ0 x4 Φ0 x5 Φ0 x6 Φ0 x7 Φ0 x8 Φ0 x9 Φ0 n/9 n/9 n/9 n/9 n/9 n/9 n/9 n/9 n/9

b1 b2 b3 b4 b5 b6 b7 b8 b9 - For each variable in Φ0, create

bucket with large number of variables.

  • Independently, sample each

constraint as:

Sample constraint C from Φ0.

slide-29
SLIDE 29

Hard Instance

#variables = n and #constraints = m = ∆ · n. x1 Φ0 x2 Φ0 x3 Φ0 x4 Φ0 x5 Φ0 x6 Φ0 x7 Φ0 x8 Φ0 x9 Φ0 n/9 n/9 n/9 n/9 n/9 n/9 n/9 n/9 n/9

b1 b2 b3 b4 b5 b6 b7 b8 b9 - For each variable in Φ0, create

bucket with large number of variables.

  • Independently, sample each

constraint as:

Sample constraint C from Φ0. For each variable x in SC, choose yx ∈ Bx, u.a.r.

slide-30
SLIDE 30

Hard Instance

#variables = n and #constraints = m = ∆ · n. x1 Φ0 x2 Φ0 x3 Φ0 x4 Φ0 x5 Φ0 x6 Φ0 x7 Φ0 x8 Φ0 x9 Φ0 n/9 n/9 n/9 n/9 n/9 n/9 n/9 n/9 n/9

b1 b2 b3 b4 b5 b6 b7 b8 b9 - For each variable in Φ0, create

bucket with large number of variables.

  • Independently, sample each

constraint as:

Sample constraint C from Φ0. For each variable x in SC, choose yx ∈ Bx, u.a.r. Put the constraint C on the variables {yx}x∈SC .

slide-31
SLIDE 31

Hard Instance

#variables = n and #constraints = m = ∆ · n. x1 Φ0 x2 Φ0 x3 Φ0 x4 Φ0 x5 Φ0 x6 Φ0 x7 Φ0 x8 Φ0 x9 Φ0 n/9 n/9 n/9 n/9 n/9 n/9 n/9 n/9 n/9

b1 b2 b3 b4 b5 b6 b7 b8 b9 - For each variable in Φ0, create

bucket with large number of variables.

  • Independently, sample each

constraint as:

Sample constraint C from Φ0. For each variable x in SC, choose yx ∈ Bx, u.a.r. Put the constraint C on the variables {yx}x∈SC .

W.h.p., the instance hypergraph generated has o(n) cycles of length at most η log n for η > 0.

slide-32
SLIDE 32

Hard Instance

#variables = n and #constraints = m = ∆ · n. x1 Φ0 x2 Φ0 x3 Φ0 x4 Φ0 x5 Φ0 x6 Φ0 x7 Φ0 x8 Φ0 x9 Φ0 n/9 n/9 n/9 n/9 n/9 n/9 n/9 n/9 n/9

b1 b2 b3 b4 b5 b6 b7 b8 b9 - For each variable in Φ0, create

bucket with large number of variables.

  • Independently, sample each

constraint as:

Sample constraint C from Φ0. For each variable x in SC, choose yx ∈ Bx, u.a.r. Put the constraint C on the variables {yx}x∈SC .

W.h.p., the instance hypergraph generated has o(n) cycles of length at most η log n for η > 0. Remove one constraint from every small cycle and get an instance of girth η log n.

slide-33
SLIDE 33

Overview - Completeness

Instance: Consistent Distributions: S T DS DT DS∩T

slide-34
SLIDE 34

Overview - Completeness

Instance: Consistent Distributions: S T DS DT DS∩T Step 2: Construction of consistent distribution – Conditioning and propagating.

slide-35
SLIDE 35

Overview - Completeness

Instance: Consistent Distributions: S T DS DT DS∩T Step 1: Consistent Low-Diameter Decompositions. Step 2: Construction of consistent distribution – Conditioning and propagating.

slide-36
SLIDE 36

Step 1: Requirements

A family of distributions, {CS}|S|≤t

slide-37
SLIDE 37

Step 1: Requirements

A family of distributions, {CS}|S|≤t CS: a distribution supported on partitions of S into low-diameter (not necessarily connected) components in the hypergraph.

slide-38
SLIDE 38

Step 1: Requirements

A family of distributions, {CS}|S|≤t CS: a distribution supported on partitions of S into low-diameter (not necessarily connected) components in the

  • hypergraph. Target diameter = girth/100.
slide-39
SLIDE 39

Step 1: Requirements

A family of distributions, {CS}|S|≤t CS: a distribution supported on partitions of S into low-diameter (not necessarily connected) components in the

  • hypergraph. Target diameter = girth/100.

Minimize the quantity: the probability of a hyperedge being

  • cut. Target = ε.
slide-40
SLIDE 40

Step 1: Requirements

A family of distributions, {CS}|S|≤t CS: a distribution supported on partitions of S into low-diameter (not necessarily connected) components in the

  • hypergraph. Target diameter = girth/100.

Consistency:

Figure: S ⊂ T

Minimize the quantity: the probability of a hyperedge being

  • cut. Target = ε.
slide-41
SLIDE 41

Step 1: Requirements

A family of distributions, {CS}|S|≤t CS: a distribution supported on partitions of S into low-diameter (not necessarily connected) components in the

  • hypergraph. Target diameter = girth/100.

Consistency:

Figure: S ⊂ T

Minimize the quantity: the probability of a hyperedge being

  • cut. Target = ε.
slide-42
SLIDE 42

Step 1: Requirements

A family of distributions, {CS}|S|≤t CS: a distribution supported on partitions of S into low-diameter (not necessarily connected) components in the

  • hypergraph. Target diameter = girth/100.

Consistency:

Figure: S ⊂ T

Minimize the quantity: the probability of a hyperedge being

  • cut. Target = ε.
slide-43
SLIDE 43

Step 1: Requirements

A family of distributions, {CS}|S|≤t CS: a distribution supported on partitions of S into low-diameter (not necessarily connected) components in the

  • hypergraph. Target diameter = girth/100.

Consistency:

Figure: S ⊂ T

S T DS DT DS∩T Minimize the quantity: the probability of a hyperedge being

  • cut. Target = ε.
slide-44
SLIDE 44

Step 2: Conditioning and Propagation

Assume: c = 1 Construction of DS:

  • Sample a partition P of S from

CS.

slide-45
SLIDE 45

Step 2: Conditioning and Propagation

Assume: c = 1 Construction of DS:

  • Sample a partition P of S from

CS.

  • For each cell T of P, construct

tree TS by connecting all shortest

  • paths. Root the tree arbitrarily.
slide-46
SLIDE 46

Step 2: Conditioning and Propagation

Assume: c = 1 C1 C2 Construction of DS:

  • Sample a partition P of S from

CS.

  • For each cell T of P, construct

tree TS by connecting all shortest

  • paths. Root the tree arbitrarily.
  • Independently, for each TS

condition and propagate assignments in TS using the local distribution from basic relaxation.

slide-47
SLIDE 47

Step 2: Conditioning and Propagation

Assume: c = 1 C1 C2 Construction of DS:

  • Sample a partition P of S from

CS.

  • For each cell T of P, construct

tree TS by connecting all shortest

  • paths. Root the tree arbitrarily.
  • Independently, for each TS

condition and propagate assignments in TS using the local distribution from basic relaxation.

slide-48
SLIDE 48

Step 2: Conditioning and Propagation

Assume: c = 1 C1 C2 Construction of DS:

  • Sample a partition P of S from

CS.

  • For each cell T of P, construct

tree TS by connecting all shortest

  • paths. Root the tree arbitrarily.
  • Independently, for each TS

condition and propagate assignments in TS using the local distribution from basic relaxation.

slide-49
SLIDE 49

Step 2: Conditioning and Propagation

Assume: c = 1 Construction of DS:

  • Sample a partition P of S from

CS.

  • For each cell T of P, construct

tree TS by connecting all shortest

  • paths. Root the tree arbitrarily.
  • Independently, for each TS

condition and propagate assignments in TS using the local distribution from basic relaxation.

  • For cell T, retain only the

assignments to variables in T.

slide-50
SLIDE 50

Step 2: Conditioning and Propagation

Assume: c = 1 The cut constraints may not be satisfied. Construction of DS:

  • Sample a partition P of S from

CS.

  • For each cell T of P, construct

tree TS by connecting all shortest

  • paths. Root the tree arbitrarily.
  • Independently, for each TS

condition and propagate assignments in TS using the local distribution from basic relaxation.

  • For cell T, retain only the

assignments to variables in T.

slide-51
SLIDE 51

Step 2: Conditioning and Propagation

Assume: c = 1 The cut constraints may not be satisfied. The dis- tribution for any tree is in- dependent of the choice of root. Construction of DS:

  • Sample a partition P of S from

CS.

  • For each cell T of P, construct

tree TS by connecting all shortest

  • paths. Root the tree arbitrarily.
  • Independently, for each TS

condition and propagate assignments in TS using the local distribution from basic relaxation.

  • For cell T, retain only the

assignments to variables in T.

slide-52
SLIDE 52

Step 2: Conditioning and Propagation

Assume: c = 1 The cut constraints may not be satisfied. The dis- tribution for any tree is in- dependent of the choice of root. Construction of DS:

  • Sample a partition P of S from

CS.

  • For each cell T of P, construct

tree TS by connecting all shortest

  • paths. Root the tree arbitrarily.
  • Independently, for each TS

condition and propagate assignments in TS using the local distribution from basic relaxation.

  • For cell T, retain only the

assignments to variables in T. High girth + consistent low-diameter decomposition ⇒ Consistent Distribution.

slide-53
SLIDE 53

Construction of Step 1

Charikar-Makarychev-Makarychev-09: Can define a metric on the hypergraph (that grows with hypergraph distance) so that restriction on any small set is isometrically embeddable on sphere.

slide-54
SLIDE 54

Construction of Step 1

Charikar-Makarychev-Makarychev-09: Can define a metric on the hypergraph (that grows with hypergraph distance) so that restriction on any small set is isometrically embeddable on sphere. Charikar et al. 1998: There exists a rotation invariant, oblivious decompo- sition of sphere into low diameter com- ponents.

slide-55
SLIDE 55

Construction of Step 1

Charikar-Makarychev-Makarychev-09: Can define a metric on the hypergraph (that grows with hypergraph distance) so that restriction on any small set is isometrically embeddable on sphere. Charikar et al. 1998: There exists a rotation invariant, oblivious decompo- sition of sphere into low diameter com- ponents.

slide-56
SLIDE 56

Construction of Step 1

Charikar-Makarychev-Makarychev-09: Can define a metric on the hypergraph (that grows with hypergraph distance) so that restriction on any small set is isometrically embeddable on sphere. Charikar et al. 1998: There exists a rotation invariant, oblivious decompo- sition of sphere into low diameter com- ponents.

slide-57
SLIDE 57

Construction of Step 1

Charikar-Makarychev-Makarychev-09: Can define a metric on the hypergraph (that grows with hypergraph distance) so that restriction on any small set is isometrically embeddable on sphere. Charikar et al. 1998: There exists a rotation invariant, oblivious decompo- sition of sphere into low diameter com- ponents.

slide-58
SLIDE 58

Construction of Step 1

Charikar-Makarychev-Makarychev-09: Can define a metric on the hypergraph (that grows with hypergraph distance) so that restriction on any small set is isometrically embeddable on sphere. Charikar et al. 1998: There exists a rotation invariant, oblivious decompo- sition of sphere into low diameter com- ponents.

slide-59
SLIDE 59

Construction of Step 1

Charikar-Makarychev-Makarychev-09: Can define a metric on the hypergraph (that grows with hypergraph distance) so that restriction on any small set is isometrically embeddable on sphere. Charikar et al. 1998: There exists a rotation invariant, oblivious decompo- sition of sphere into low diameter com- ponents. The probability of cutting a hyperedge dictates the size of the sets we can han- dle.

slide-60
SLIDE 60

Conclusion

  • We prove a dichotomy result for all CSPs for linear

programming relaxations.

slide-61
SLIDE 61

Conclusion

  • We prove a dichotomy result for all CSPs for linear

programming relaxations.

  • The result can also be interpreted as reducing the problem of

showing hardness to a possibly easier task.

slide-62
SLIDE 62

Conclusion

  • We prove a dichotomy result for all CSPs for linear

programming relaxations.

  • The result can also be interpreted as reducing the problem of

showing hardness to a possibly easier task. Q: Can the number of levels of SA be improved?

slide-63
SLIDE 63

Conclusion

  • We prove a dichotomy result for all CSPs for linear

programming relaxations.

  • The result can also be interpreted as reducing the problem of

showing hardness to a possibly easier task. Q: Can the number of levels of SA be improved? Q: What can be said for the case of SDP hierarchies?

slide-64
SLIDE 64

Conclusion

  • We prove a dichotomy result for all CSPs for linear

programming relaxations.

  • The result can also be interpreted as reducing the problem of

showing hardness to a possibly easier task. Q: Can the number of levels of SA be improved? Q: What can be said for the case of SDP hierarchies? Questions?

slide-65
SLIDE 65

Other Dichotomy Results

[Raghavendra-08]: Assuming Unique Games Conjecture, either a basic SDP achieves a (c, s)-approximation for a CSP or it is NP-hard to do so (for th

slide-66
SLIDE 66

Other Dichotomy Results

[Raghavendra-08]: Assuming Unique Games Conjecture, either a basic SDP achieves a (c, s)-approximation for a CSP or it is NP-hard to do so (for th [Raghavendra-Steurer-09]: (For Unique Games) If a basic SDP has gap of (c, s) then so does (log log n)

1 4 -levels of mixed

relaxation.

slide-67
SLIDE 67

Other Dichotomy Results

[Raghavendra-08]: Assuming Unique Games Conjecture, either a basic SDP achieves a (c, s)-approximation for a CSP or it is NP-hard to do so (for th [Raghavendra-Steurer-09]: (For Unique Games) If a basic SDP has gap of (c, s) then so does (log log n)

1 4 -levels of mixed

relaxation. This result If basic LP relaxation has a gap of (c, s), then so does O(log n)-level SA.