Lecture 7: Arora Rao Vazirani Lecture Outline Part I: Semidefinite - - PowerPoint PPT Presentation

โ–ถ
lecture 7 arora rao vazirani lecture outline
SMART_READER_LITE
LIVE PREVIEW

Lecture 7: Arora Rao Vazirani Lecture Outline Part I: Semidefinite - - PowerPoint PPT Presentation

Lecture 7: Arora Rao Vazirani Lecture Outline Part I: Semidefinite Programming Relaxation for Sparsest Cut Part II: Combining Approaches Part III: Arora-Rao-Vazirani Analysis Overview Part IV: Analyzing Matchings of Close Points


slide-1
SLIDE 1

Lecture 7: Arora Rao Vazirani

slide-2
SLIDE 2

Lecture Outline

  • Part I: Semidefinite Programming Relaxation for

Sparsest Cut

  • Part II: Combining Approaches
  • Part III: Arora-Rao-Vazirani Analysis Overview
  • Part IV: Analyzing Matchings of Close Points
  • Part V: Reduction to the Well-Separated Case
  • Part VI: Open Problems
slide-3
SLIDE 3

Part I: Semidefinite Programming Relaxation for Sparsest Cut

slide-4
SLIDE 4
  • Reformulation: Want to minimize

ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜, ๐‘—,๐‘˜ โˆˆ๐น(๐ป) ๐‘ฆ๐‘˜ โˆ’ ๐‘ฆ๐‘—

2 over all cut

pseudo-metrics normalized so that ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜ ๐‘ฆ๐‘˜ โˆ’ ๐‘ฆ๐‘—

2 = 1

  • More precisely, take ๐‘’2 ๐‘—, ๐‘˜ = ๐‘ฆ๐‘˜ โˆ’ ๐‘ฆ๐‘—

2 and

minimize ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜, ๐‘—,๐‘˜ โˆˆ๐น(๐ป) ๐‘’2(๐‘—, ๐‘˜) subject to:

1. โˆƒ๐‘‘: โˆ€๐‘—, xi โˆˆ {โˆ’๐‘‘, +๐‘‘} 2. ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜ ๐‘’2(๐‘—, ๐‘˜) = 1

Problem Reformulation

slide-5
SLIDE 5
  • Reformulation: Minimize

ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜, ๐‘—,๐‘˜ โˆˆ๐น(๐ป)(๐‘ฆ๐‘—

2 โˆ’ 2๐‘ฆ๐‘—๐‘ฆ๐‘˜ + ๐‘ฆ๐‘˜ 2) subject to:

1. โˆƒ๐‘‘: โˆ€๐‘—, xi โˆˆ {โˆ’๐‘‘, +๐‘‘} 2. ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜ (๐‘ฆ๐‘—

2 โˆ’ 2๐‘ฆ๐‘—๐‘ฆ๐‘˜ + ๐‘ฆ๐‘˜ 2) = 1

  • Relaxation: Minimize

ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜, ๐‘—,๐‘˜ โˆˆ๐น(๐ป)(๐‘๐‘—๐‘—โˆ’2๐‘๐‘—๐‘˜ + ๐‘

๐‘˜๐‘˜) subject to:

1. โˆ€๐‘—, ๐‘˜, ๐‘๐‘—๐‘— = ๐‘

๐‘˜๐‘˜

2. ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜ (๐‘๐‘—๐‘—โˆ’2๐‘๐‘—๐‘˜ + ๐‘

๐‘˜๐‘˜) = 1

3. ๐‘ โ‰ฝ 0

Problem Relaxation

slide-6
SLIDE 6
  • Consider the cycle of length ๐‘œ. The semidefinite

program can place the cycle on the unit circle and assign each ๐‘ฆ๐‘— the corresponding vector ๐‘ค๐‘—.

Bad Example: The Cycle

4 3 1 5 2 ๐‘ค1 ๐‘ค2 ๐‘ค3 ๐‘ค4 ๐‘ค5

G

slide-7
SLIDE 7
  • ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜(๐‘’2(๐‘—, ๐‘˜)) = ฮ˜(๐‘œ2)
  • ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜, ๐‘—,๐‘˜ โˆˆ๐น(๐ป)(๐‘’2(๐‘—, ๐‘˜)) = ฮ˜(๐‘œ โ‹… 1/๐‘œ2)
  • Gives sparsity ฮ˜(1/๐‘œ3), true value is ฮ˜(1/n2)
  • Gap is ฮฉ(๐‘œ), which is horrible!

Bad Example: The Cycle

4 3 1 5 2 ๐‘ค1 ๐‘ค2 ๐‘ค3 ๐‘ค4 ๐‘ค5

G

slide-8
SLIDE 8

Part II: Combining Approaches

slide-9
SLIDE 9
  • Why did the semidefinite program do so much

worse than the linear program?

  • Missing: Triangle inequalities

๐‘’2 ๐‘—, ๐‘™ โ‰ค ๐‘’2(๐‘—, ๐‘˜) + ๐‘’2(๐‘˜, ๐‘™)

  • What happens if we add the triangle inequalities

to the semidefinite program?

Adding the Triangle Inequalities

slide-10
SLIDE 10
  • Let ฮ˜ be the angle between ๐‘ค๐‘— โˆ’ ๐‘ค๐‘˜ and ๐‘ค๐‘™ โˆ’ ๐‘ค๐‘˜
  • ๐‘ค๐‘™ โˆ’ ๐‘ค๐‘—

2 =

๐‘ค๐‘˜ โˆ’ ๐‘ค๐‘—

2 + ๐‘ค๐‘™ โˆ’ ๐‘ค๐‘˜ 2 if ฮ˜ = ๐œŒ 2

  • ๐‘ค๐‘™ โˆ’ ๐‘ค๐‘—

2 >

๐‘ค๐‘˜ โˆ’ ๐‘ค๐‘—

2 + ๐‘ค๐‘™ โˆ’ ๐‘ค๐‘˜ 2 if ฮ˜ > ๐œŒ 2

  • ๐‘ค๐‘™ โˆ’ ๐‘ค๐‘—

2 <

๐‘ค๐‘˜ โˆ’ ๐‘ค๐‘—

2 + ๐‘ค๐‘™ โˆ’ ๐‘ค๐‘˜ 2 if ฮ˜ < ๐œŒ 2

  • Triangle inequalities โฌ„ no obtuse angles

Geometric Picture

๐‘ค๐‘— ๐‘ค๐‘˜ ๐‘ค๐‘™ ฮ˜

slide-11
SLIDE 11
  • Putting ๐‘œ > 4 vectors in a circle violates triangle

inequality, so the semidefinite program no longer behaves badly on the cycle. In fact, it gets very close to the right answer.

Fixing Cycle Example

๐‘ค๐‘— ๐‘ค๐‘˜ ๐‘ค๐‘™

ฮ˜

slide-12
SLIDE 12

Goemans-Linial Relaxation

  • Semidefinite program (proposed by Goemans

and Lineal): Minimize ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜: ๐‘—,๐‘˜ โˆˆ๐น(๐ป)(๐‘๐‘—๐‘— โˆ’ 2๐‘๐‘—๐‘˜ + ๐‘

๐‘˜๐‘˜) subject to:

1.

โˆ€๐‘—, ๐‘˜, M๐‘—๐‘— = ๐‘

๐‘˜๐‘˜

2. โˆ€๐‘—, ๐‘˜, ๐‘™, ๐‘’2(๐‘—, ๐‘™) โ‰ค ๐‘’2(๐‘—, ๐‘˜) + ๐‘’2(๐‘˜, ๐‘™) where ๐‘’2(๐‘—, ๐‘˜) = ๐‘๐‘—๐‘— โˆ’ 2๐‘๐‘—๐‘˜ + ๐‘

๐‘˜๐‘˜

3. ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜ ๐‘๐‘—๐‘— โˆ’ 2๐‘๐‘—๐‘˜ + ๐‘

๐‘˜๐‘˜ = 1

4. ๐‘ โ‰ฝ 0

slide-13
SLIDE 13

Arora-Rao-Vazirani Theorem

  • Theorem [ARV]: The Goemans-Linial

relaxation for sparsest cut gives an ๐‘ƒ ๐‘š๐‘๐‘•๐‘œ -approximation and has a polynomial time rounding algorithm.

slide-14
SLIDE 14
  • Also called metrics of negative type
  • Definition: A metric is an ๐‘€2

2 metric if it is possible

to assign a vector ๐‘ค๐‘ฆ to every point ๐‘ฆ such that ๐‘’ ๐‘ฆ, ๐‘ง = ๐‘ค๐‘ง โˆ’ ๐‘ค๐‘ฆ

2.

  • Last time: General metrics can be embedded into

๐‘€1 with ๐‘ƒ log ๐‘œ distortion.

  • Theorem [ALN08]: Any ๐‘€2

2 metric embeds into ๐‘€1

with ๐‘ƒ ๐‘š๐‘๐‘•๐‘œ(๐‘š๐‘๐‘•๐‘š๐‘๐‘•๐‘œ) distortion.

  • [ARV] analyzes the algorithm more directly

๐‘€2

2 Metric Spaces

slide-15
SLIDE 15
  • Degree 4 SOS captures the triangle inequality: if

๐‘ฆ๐‘—

2 = ๐‘ฆ๐‘˜ 2 = ๐‘ฆ๐‘™ 2 then

๐‘ฆ๐‘—

2 ๐‘ฆ๐‘™ โˆ’ ๐‘ฆ๐‘— 2 โ‰ค ๐‘ฆ๐‘— 2 ๐‘ฆ๐‘˜ โˆ’ ๐‘ฆ๐‘— 2 + ๐‘ฆ๐‘— 2 ๐‘ฆ๐‘™ โˆ’ ๐‘ฆ๐‘˜ 2

โฌ„2๐‘ฆ๐‘—

2 ๐‘ฆ๐‘— 2 โˆ’ ๐‘ฆ๐‘—๐‘ฆ๐‘™ โ‰ค 2๐‘ฆ๐‘— 2(2๐‘ฆ๐‘— 2 โˆ’ ๐‘ฆ๐‘—๐‘ฆ๐‘˜ โˆ’ ๐‘ฆ๐‘—๐‘ฆ๐‘˜)

  • Proof:

๐‘ฆ๐‘— โˆ’ ๐‘ฆ๐‘˜

2 ๐‘ฆ๐‘˜ โˆ’ ๐‘ฆ๐‘™ 2 = 4(๐‘ฆ๐‘— 2 โˆ’ ๐‘ฆ๐‘—๐‘ฆ๐‘˜)(๐‘ฆ๐‘— 2 โˆ’ ๐‘ฆ๐‘˜๐‘ฆ๐‘™)

= 4๐‘ฆ๐‘—

2 ๐‘ฆ๐‘— 2 โˆ’ ๐‘ฆ๐‘—๐‘ฆ๐‘˜ โˆ’ ๐‘ฆ๐‘˜๐‘ฆ๐‘™ + ๐‘ฆ๐‘—๐‘ฆ๐‘™ โ‰ฅ 0

  • Thus, degree 4 SOS captures the Goemans-Linial

relaxation

Goemans-Linial Relaxation and SOS

slide-16
SLIDE 16

Part III: Arora-Rao-Vazirani Analysis Overview

slide-17
SLIDE 17
  • Semidefinite program gives us one vector ๐‘ค๐‘— for

each vertex ๐‘—.

  • We first consider the case when these vectors

are spread out.

  • Definition: We say that a set of ๐‘œ vectors {๐‘ค๐‘—} is

well-spread if it can be scaled so that:

1. โˆ€๐‘—, ๐‘ค๐‘— โ‰ค 1 2.

1 ๐‘œ2 ฯƒ๐‘—<๐‘˜ ๐‘’๐‘—๐‘˜ 2 is ฮฉ(1) (the average squared distance

between vectors is constant)

  • We will assume we are using this scaling.

Well-Spread Case

slide-18
SLIDE 18
  • Theorem: Given a set of ๐‘œ vectors {๐‘ค๐‘—} which

are well-spread and obey the triangle inequality, there exist well-separated subsets ๐‘Œ and ๐‘ of these vectors of linear size. In other words, there exist ๐‘Œ, ๐‘ such that:

1. ๐‘Œ and ๐‘ are ฮ” far apart (i.e. โˆ€๐‘ค๐‘— โˆˆ ๐‘Œ, ๐‘ค๐‘˜ โˆˆ ๐‘, ๐‘’๐‘—๐‘˜

2 โ‰ฅ ฮ”) where ฮ” is ฮฉ 1 ๐‘š๐‘๐‘•๐‘œ

2. |๐‘Œ| and |๐‘| are both ฮฉ(๐‘œ)

Structure Theorem

slide-19
SLIDE 19
  • Idea: If we have well-separated subsets ๐‘Œ, ๐‘,

take a random cut of the form (๐‘‡๐‘ , าง ๐‘‡๐‘ ) where ๐‘‡๐‘  = {๐‘—: ๐‘’2 ๐‘ค๐‘—, ๐‘Œ = min

๐‘˜:๐‘ค๐‘˜โˆˆ๐‘ ๐‘’๐‘—๐‘˜ 2 โ‰ค ๐‘ } and ๐‘  โˆˆ 0, ฮ”

  • All ๐‘—, ๐‘˜ โˆˆ ๐น(๐ป) contribute at most

๐‘’๐‘—๐‘˜

2

ฮ” to the

expected number of edges cut and ๐‘’๐‘—๐‘˜

2 to

ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜, ๐‘—,๐‘˜ โˆˆ๐น(๐ป) ๐‘’๐‘—๐‘˜

2 (the number of edges the

SDP โ€œthinksโ€ are cut)

Finding a Sparse Cut

slide-20
SLIDE 20
  • Since ๐‘Œ, ๐‘ have size ฮฉ(๐‘œ) and are always on
  • pposite sides of the cut, we always have that

๐‘‡๐‘  โ‹… | าง ๐‘‡๐‘ | is ฮ˜ ๐‘œ2 . This matches ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜ ๐‘’๐‘—๐‘˜

2 up

to a constant factor. (this is why we need ๐‘Œ and ๐‘ to have linear size!)

  • Thus, the expected ratio of the sparsity to the

SDP value is at most

1 ฮ” = O

๐‘š๐‘๐‘•๐‘œ , as needed.

Finding a Sparse Cut Continued

slide-21
SLIDE 21
  • Take the hypercube โˆ’

1 log2 ๐‘œ , 1 log2 ๐‘œ log2 ๐‘œ

  • X = ๐‘ฆ: ฯƒ๐‘— ๐‘ฆ๐‘— โ‰ค โˆ’1 and Y = ๐‘ง: ฯƒ๐‘— ๐‘ฆ๐‘— โ‰ฅ 1

have the following properties:

1. ๐‘Œ and ๐‘ have linear size 2. โˆ€๐‘ฆ โˆˆ ๐‘Œ, ๐‘ง โˆˆ ๐‘, ๐‘ฆ, ๐‘ง differ in โ‰ฅ 2 log2 ๐‘œ

  • coordinates. Thus, ๐‘’2 ๐‘ฆ, ๐‘ง โ‰ฅ

2 log2 ๐‘œ log2 ๐‘œ = 2 log2 ๐‘œ

Tight Example: Hypercube

slide-22
SLIDE 22
  • Let ๐‘’ be the dimension such that โˆ€๐‘—, ๐‘ค๐‘— โˆˆ โ„๐‘’.
  • Algorithm (Parameters ๐œ > 0, ฮ”, ๐‘’)
  • 1. Choose a random ๐‘ฃ โˆˆ โ„๐‘’.
  • 2. Find a value ๐‘ such that there are ฮฉ(๐‘œ) vectors ๐‘ค๐‘—

with ๐‘ค๐‘— โ‹… ๐‘ฃ โ‰ค ๐‘ and ฮฉ(๐‘œ) vectors ๐‘ค๐‘˜ with ๐‘ค๐‘˜ โ‹… ๐‘ฃ โ‰ฅ ๐‘ +

๐œ ๐‘’. Let ๐‘Œโ€ฒ and ๐‘โ€ฒ be these two sets of vectors

  • 3. As long as there is a pair ๐‘ฆ โˆˆ ๐‘Œโ€ฒ, ๐‘ง โˆˆ ๐‘โ€ฒ such that

๐‘’ ๐‘ฆ, ๐‘ง < ฮ”, delete ๐‘ฆ from ๐‘Œโ€ฒ and ๐‘ง from ๐‘โ€ฒ. The resulting sets will be the desired ๐‘Œ, ๐‘.

  • Need to show: P[๐‘Œ, ๐‘ have size ฮฉ(๐‘œ)] is ฮฉ(1)

Finding Well-Separated Sets

slide-23
SLIDE 23
  • Will first explain why step 1,2 succeed with

probability 2๐œ€ > 0.

  • Will then show that the probability step 3

deletes a linear number of points is โ‰ค ๐œ€

  • Together, this implies that the entire algorithm

succeeds with probability at least ๐œ€ > 0.

Finding Well-Separated Sets

slide-24
SLIDE 24
  • What happens if we project a vector ๐‘ค of length

๐‘š in a random direction in โ„๐‘’?

  • Without loss of generality, assume ๐‘ค = ๐‘“1
  • To pick a random unit vector in โ„๐‘’, choose

each coordinate according to ๐‘‚ 0,

1 ๐‘’ (the

normal distribution with mean 0 and standard deviation

1 ๐‘’), then rescale.

  • If ๐‘’ is not too small, w.h.p. very little rescaling

will be needed.

Behavior of Gaussian Projections

slide-25
SLIDE 25
  • What happens if we project a vector of length ๐‘š

in a random direction in โ„๐‘’?

  • Resulting value has a distribution which is โ‰ˆ

normal distribution of mean 0, standard deviation

1 ๐‘’ (difference comes from the

rescaling step)

Behavior of Gaussian Projections

slide-26
SLIDE 26
  • If we take a random ๐‘ฃ โˆˆ โ„๐‘’, with probability

ฮฉ(1), ฯƒ๐‘—<๐‘˜ (๐‘ค๐‘˜ โˆ’ ๐‘ค๐‘—) โ‹… ๐‘ฃ is ฮฉ

๐‘œ2 ๐‘’

  • Note: this can fail with non-negligible

probability, consider the case when โˆ€๐‘—, ๐‘ค๐‘— = ยฑ๐‘ค. If ๐‘ฃ is orthogonal to ๐‘ค then everything is projected to 0.

  • For arbitrarily small ๐œ— > 0, with very high

probability, |๐‘ค๐‘— โ‹… ๐‘ฃ| is ๐‘ƒ

1 ๐‘’ for 1 โˆ’ ๐œ— ๐‘œ of the

๐‘— โˆˆ [1, ๐‘œ]

Success of Steps 1,2

slide-27
SLIDE 27
  • Together, these facts imply that if we choose a

random unit vector ๐‘ฃ, with probability ฮฉ(1), there exist ๐‘Œโ€ฒ, ๐‘โ€ฒ, ๐‘1, ๐‘2 such that

1. ๐‘Œโ€ฒ, ๐‘โ€ฒ have size ฮฉ(๐‘œ) 2. โˆ€๐‘ฆ โˆˆ ๐‘Œโ€ฒ, ๐‘ฃ โ‹… ๐‘ฆ โ‰ค ๐‘1 3. โˆ€๐‘ง โˆˆ ๐‘โ€ฒ, ๐‘ฃ โ‹… ๐‘ง โ‰ฅ ๐‘2 4. ๐‘2 โˆ’ ๐‘1 is ฮฉ(1)

Success of Steps 1,2

slide-28
SLIDE 28
  • We need to show that the probability step 3

eliminates

๐‘›๐‘—๐‘œ{ ๐‘Œ ,|๐‘|} 2

pairs of points is at most ๐œ€

  • We also need to show how the general case can

be reduced to the well-spread case.

Remaining Steps

slide-29
SLIDE 29

Part IV: Analyzing Matchings of Close Points

slide-30
SLIDE 30

Matching Covers

  • If part 3 of the algorithm causes it to fail with

probability ๐œ€, then for ๐œ€ fraction of the directions ๐‘ฃ there is a matching ๐‘๐‘ฃ of points of size ๐‘‘โ€ฒ๐‘œ such that for each pair (๐‘ค๐‘—, ๐‘ค๐‘˜) in the matching:

1. d2 ๐‘ค๐‘—, ๐‘ค๐‘˜ โ‰ค ฮ” 2. ๐‘ค๐‘˜ โˆ’ ๐‘ค๐‘— โ‹… ๐‘ฃ โ‰ฅ

2๐œ ๐‘’

where ๐œ€, ๐‘‘โ€ฒ, ๐œ > 0 are constants

  • Note: Corresponds to Definition 4 in [ARV]
  • Define the matching graph ๐‘ to be ๐‘ =โˆช๐‘ฃ ๐‘๐‘ฃ
slide-31
SLIDE 31
  • Assume that ๐‘’ ๐‘ค๐‘—, ๐‘ค๐‘˜ โ‰ค

ฮ” for some ๐‘ค๐‘—, ๐‘ค๐‘˜

  • P

๐‘ค๐‘˜ โˆ’ ๐‘ค๐‘— โ‹… ๐‘ฃ โ‰ฅ 2๐œ

๐‘’ โˆผ ๐‘“ โˆ’

4๐œ2 ๐‘’2(๐‘ค๐‘—,๐‘ค๐‘˜) โ‰ค ๐‘“โˆ’4๐œ2 ฮ”

  • If ฮ” is a sufficiently small constant times

1 ๐‘š๐‘๐‘•๐‘œ ,

with high probability there are no pairs of close points at all between ๐‘Œโ€ฒ and ๐‘โ€ฒ!

Analyzing ฮ” = ฮฉ

1 ๐‘š๐‘๐‘•๐‘œ

slide-32
SLIDE 32
  • When the algorithm fails in step 3, this gives us

pairs of points (๐‘ค๐‘—, ๐‘ค๐‘˜) which are edges of the matching graph ๐‘, implying that ๐‘’2 ๐‘ค๐‘—, ๐‘ค๐‘˜ โ‰ค ฮ” and ๐‘ค๐‘˜ โˆ’ ๐‘ค๐‘— โ‹… ๐‘ฃ โ‰ฅ

2๐œ ๐‘’

  • We will use this to find pairs of points (๐‘ค๐‘—, ๐‘ค๐‘˜)

which are ๐‘™ steps apart in the matching graph where ๐‘ค๐‘˜ โˆ’ ๐‘ค๐‘— โ‹… ๐‘ฃ โ‰ฅ

๐‘™๐œ ๐‘’

Key Idea for Larger ฮ”

slide-33
SLIDE 33
  • We will find pairs of points (๐‘ค๐‘—, ๐‘ค๐‘˜) which are ๐‘™

steps apart in the matching graph where ๐‘ค๐‘˜ โˆ’ ๐‘ค๐‘— โ‹… ๐‘ฃ โ‰ฅ ๐‘™๐œ

๐‘’

  • Using triangle inequality, ๐‘’2 ๐‘ค๐‘—, ๐‘ค๐‘˜ โ‰ค ๐‘™ฮ”
  • P

๐‘ค๐‘˜ โˆ’ ๐‘ค๐‘— โ‹… ๐‘ฃ โ‰ฅ ๐‘™๐œ

๐‘’ โˆผ ๐‘“ โˆ’

๐‘™2๐œ2 ๐‘’2(๐‘ค๐‘—,๐‘ค๐‘˜) โ‰ค ๐‘“โˆ’๐‘™๐œ2 ฮ”

  • For ฮ” = ฮฉ

1 log ๐‘œ , if we can apply this with ๐‘™ =

ฮฉ log ๐‘œ , we again obtain a contradiction.

Key Idea for Larger ฮ” Continued

slide-34
SLIDE 34
  • Lemma: If a graph ๐ป has average degree ๐‘’, we

can find a non-empty subgraph of ๐ป which has minimal degree

๐‘’ 4.

  • Proof: Iteratively delete vertices which have

degree โ‰ค ๐‘’

  • 4. The total number of edges deleted

is at most ๐‘œ๐‘’

4 . However, 2|๐น ๐ป | โ‰ฅ ๐‘œ๐‘’, so there

must be โ‰ฅ ๐‘œ๐‘’

4 edges remaining.

Average Degree to Minimal Degree

slide-35
SLIDE 35
  • Average probability that a vertex is matched is

at least ๐‘‘โ€ฒ๐œ€

  • Can apply a similar idea and delete any vertex

which is matched with probability โ‰ค ๐‘‘โ€ฒ๐œ€

4

  • By similar logic, at least half the edges are

preserved.

  • This implies that there are at least ๐‘‘โ€ฒ๐‘œ vertices

remaining (otherwise more than half of every matching of โ‰ฅ ๐‘‘โ€ฒ๐‘œ edges is deleted)

  • Note: Corresponds to Lemma 4 of [ARV09]

Minimal Probability Guarantee

slide-36
SLIDE 36
  • Corollary: There is a set of vertices ๐‘Œ of size โ‰ฅ

๐‘‘โ€ฒ๐‘œ such that โˆ€๐‘ฆ โˆˆ ๐‘Œ, ๐‘„ ๐‘ฆ is matched with an ๐‘ฆโ€ฒ โˆˆ ๐‘Œ โ‰ฅ ๐œ€โ€ฒ where ๐œ€โ€ฒ = ๐‘‘โ€ฒ๐œ€

4

Minimal Probability Guarantee

slide-37
SLIDE 37
  • How can we find pairs of points whose

projected distance is larger and larger by taking steps in the matching graph?

  • Letโ€™s assume we have a very convenient

inductive setup.

Building Up Projection Distances

slide-38
SLIDE 38
  • Have a set of points ๐‘Œ of size โ‰ฅ ๐‘‘โ€ฒ๐‘œ

โˆ€๐‘ฆ โˆˆ ๐‘Œ, ๐‘„ ๐‘ฆ is matched with an ๐‘ฆโ€ฒ โˆˆ ๐‘Œ โ‰ฅ ๐œ€โ€ฒ

  • Inductive setup: Assume we also have a subset

๐‘Ž โІ ๐‘Œ of points of size ๐œ|๐‘Œ| such that

โˆ€๐‘จ โˆˆ ๐‘Ž, ๐‘„ โˆƒ๐‘จโ€ฒ โˆˆ ๐‘Œ: ๐‘’๐‘ ๐‘จ, ๐‘จโ€ฒ โ‰ค ๐‘™, ๐‘จ โˆ’ ๐‘จโ€ฒ โ‹… ๐‘ฃ โ‰ฅ ๐‘™๐œ ๐‘’ โ‰ฅ 1 โˆ’ ๐œ€โ€ฒ 4

where ๐‘’๐‘(๐‘จ, ๐‘จโ€ฒ) is the number of steps required to reach ๐‘จโ€ฒ from ๐‘จ in the matching graph

  • Note: This corresponds to Definitions 6,8 of

[ARV]

Setup

slide-39
SLIDE 39
  • ๐‘Œ is a set of points where every ๐‘ฆ โˆˆ ๐‘Œ is

matched to another ๐‘ฆโ€ฒ โˆˆ ๐‘Œ for โ‰ฅ ๐œ€โ€ฒ fraction of the directions

  • Have a subset ๐‘Ž โІ ๐‘Œ of size โ‰ฅ ๐œ|๐‘Œ| where each

๐‘จ โˆˆ ๐‘Ž is โ€œcoveredโ€ in โ‰ฅ 1 โˆ’

๐œ€โ€ฒ 4 fraction of the

directions by points which are โ‰ค ๐‘™ steps away in the matching graph whose projected distance is โ‰ฅ

๐‘™๐œ ๐‘’

Setup Rephrased

slide-40
SLIDE 40

Composition Step

๐‘ฃ ๐‘จ ๐‘จโ€ฒ ๐‘ฆโ€ฒ

  • r

๐‘ฆโ€ฒ ๐‘จ ๐‘จโ€ฒ

slide-41
SLIDE 41
  • Given a direction ๐‘ฃ, for each point ๐‘จ โˆˆ ๐‘Ž:
  • 1. Check if ๐‘จ is matched in ๐‘๐‘ฃ = ๐‘โˆ’๐‘ฃ
  • 2. If so, let ๐‘ฆโ€ฒ be the point ๐‘จ is matched with.

๐‘จ โˆ’ ๐‘ฆโ€ฒ โ‹… ๐‘ฃ โ‰ฅ

2๐œ ๐‘’

  • 3. If ๐‘จ โˆ’ ๐‘ฆโ€ฒ โ‹… ๐‘ฃ > 0, check if ๐‘จ is covered in direction

๐‘ฃ. If ๐‘จ โˆ’ ๐‘ฆโ€ฒ โ‹… ๐‘ฃ < 0 check if ๐‘จ is covered in direction โˆ’๐‘ฃ. With probability โ‰ฅ 1 โˆ’

๐œ€โ€ฒ 2 , ๐‘จ is

covered in both directions. Let ๐‘จโ€ฒ = covering point.

  • 4. Observe that ๐‘จโ€ฒ โˆ’ ๐‘ฆโ€ฒ โ‹… ๐‘ฃ โ‰ฅ

๐‘™๐œ+2๐œ ๐‘’

and ๐‘’๐‘ ๐‘ฆโ€ฒ, ๐‘จโ€ฒ โ‰ค ๐‘™ + 1

Composition Step

slide-42
SLIDE 42
  • Have that the density of the new covering edges

is at least

๐œ๐œ€โ€ฒ 2 .

  • Following the same kind of logic we used to go

from average to minimal degree, can find a subset ๐‘Žโ€ฒ โІ ๐‘Œ of size โ‰ฅ

๐œ๐œ€โ€ฒ 8 |๐‘Œ| where every

vertex ๐‘จโ€ฒ โˆˆ ๐‘Žโ€ฒ is covered in โ‰ฅ ๐œ๐œ€โ€ฒ

8 of the

directions.

  • Note: Corresponds to Lemma 11 of [ARV]

Composition Step

slide-43
SLIDE 43
  • How can we recover the inductive hypothesis?
  • Can boost the covering probability to almost 1

with a small loss in the projection length!

  • Corollary 12 of [ARV] rephrased: If the covering

vectors have length at most

ฯƒ 16 log

16 ฯ„ฮดโ€ฒ +8 log 8 ฮดโ€ฒ

then if z is covered with probability

๐œ๐œ€โ€ฒ 8 with projection

length

๐‘™๐œ+2๐œ ๐‘’ , it is covered with probability 1 โˆ’

๐œ€โ€ฒ/4 with projection length (๐‘™+1)๐œ

๐‘’

Boosting Lemma

slide-44
SLIDE 44
  • If we apply this directly:

โ€“ ๐œ โˆผ ๐œ€โ€ฒ โˆ’๐‘™ โ€“ Need covering vectors to have length ๐‘ƒ

1 ๐‘š๐‘๐‘•๐œ

= ๐‘ƒ

1 ๐‘™

โ€“ Guaranteed to have length โ‰ค ๐‘™ฮ” โ€“ We can take ๐‘™ = ฮฉ(ฮ”โˆ’1

2). We want

๐‘™ ฮ” to be a large

constant times log ๐‘œ , which means we can take ฮ” = ฮฉ( ๐‘š๐‘๐‘•๐‘œ โˆ’2/3)

Bound on ๐‘™ and ฮ”

slide-45
SLIDE 45
  • To reach ๐‘™ = ฮฉ

๐‘š๐‘๐‘•๐‘œ , a more careful argument is needed, see [ARV].

  • Note: We should not expect ๐‘™ to be any higher

than O ๐‘š๐‘๐‘•๐‘œ . Recalling that the projection length with ๐‘™ steps is

๐‘™๐œ ๐‘’, if ๐‘’ = ฮ˜(๐‘š๐‘๐‘•๐‘œ)

(matching the hypercube example) and ๐‘™ is ๐œ• ๐‘š๐‘๐‘•๐‘œ then this is ๐œ•(1), which is too large!

Reaching ๐‘™ = ฮฉ ๐‘š๐‘๐‘•๐‘œ

slide-46
SLIDE 46

Part V: Reduction to the Well- Separated Case

slide-47
SLIDE 47
  • Take the scaling where ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜ ๐‘’2(๐‘—, ๐‘˜) =

๐‘œ 2

(i.e. the average squared distance between pairs

  • f points is 1)
  • One of the following two cases holds:
  • 1. There exists a point ๐‘ฆ0 such that

๐‘œ 10 other points

are within squared distance

1 10 of ๐‘ฆ0

  • 2. For all points ๐‘ฆ, less than

๐‘œ 10 other points are within

squared distance

1 10 of ๐‘ฆ

Two Cases

slide-48
SLIDE 48
  • Assume there exists a point ๐‘ฆ0 such that

๐‘œ 10 other

points are within squared distance

1 10 of ๐‘ฆ0

  • Let ๐‘Œ = {x: d2 x, x0 โ‰ค

1 10}

  • Key idea: Take the Frรฉchet embedding with

respect to ๐‘Œ!

  • In particular, take

๐‘’๐‘Œ ๐‘ง, ๐‘จ = |๐‘’2 ๐‘ง, ๐‘Œ โˆ’ ๐‘’2 ๐‘จ, ๐‘Œ |

Case #1

slide-49
SLIDE 49
  • We will show that

ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜, ๐‘—,๐‘˜ โˆˆ๐น(๐ป) ๐‘’๐‘Œ ๐‘—,๐‘˜ ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜ ๐‘’๐‘Œ ๐‘—,๐‘˜

is ๐‘ƒ

ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜, ๐‘—,๐‘˜ โˆˆ๐น ๐ป ๐‘’2 ๐‘—,๐‘˜ ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜ ๐‘’2 ๐‘—,๐‘˜

  • ๐‘’๐‘Œ is an ๐‘€1 metric, so this gives an ๐‘ƒ(1)-

approximation!

  • First note that ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜, ๐‘—,๐‘˜ โˆˆ๐น(๐ป) ๐‘’๐‘Œ ๐‘—, ๐‘˜ is less

than or equal to ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜, ๐‘—,๐‘˜ โˆˆ๐น(๐ป) ๐‘’2 ๐‘—, ๐‘˜

  • We just need to show that ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜ ๐‘’๐‘Œ ๐‘—, ๐‘˜ is

ฮฉ(๐‘œ2)

Case #1 Continued

slide-50
SLIDE 50
  • Proposition: The average squared distance of

points outside of ๐‘Œ from ๐‘Œ is at least

1 5

  • Proof: If this were not the case then the average

squared distance between points would be < 1 as for all ๐‘ง, z, ๐‘’2 ๐‘ง, ๐‘จ โ‰ค ๐‘’2 ๐‘ง, ๐‘Œ + ๐‘’2 ๐‘จ, ๐‘Œ + 1 5

  • Corollary: ฯƒ๐‘—,๐‘˜:๐‘—<๐‘˜ ๐‘’๐‘Œ (๐‘—, ๐‘˜) is ฮ˜ ๐‘œ2 . To show

this, it is sufficient to consider the pairs where exactly one of ๐‘—, ๐‘˜ are in ๐‘Œ.

Case #1 Continued

slide-51
SLIDE 51
  • Assume that for all points ๐‘ฆ, there are fewer than

๐‘œ 10 other points which are within squared

distance 1

10 of ๐‘ฆ

  • Proposition: There is a point ๐‘ฆ0 such that at least

๐‘œ 2 other points are within distance 2 of ๐‘ฆ0

  • Proof: If this were not the case then the average

distance between points would be > 1.

  • Let ๐‘Œ be the set of points within distance 2 of ๐‘ฆ0.

Case #2

slide-52
SLIDE 52
  • Key idea: Subtract ๐‘ฆ0 from all vectors!
  • After this translation:

โ€“ All points in ๐‘Œ have length โ‰ค 2 โ€“ For all points ๐‘ฆ โˆˆ ๐‘Œ, there are at least

n 2 โˆ’ n 10 = 2๐‘œ 5

points in ๐‘Œ which have squared distance more than

1 10 from ๐‘ฆ. Thus, the average squared distance

between points in ๐‘Œ is ฮฉ(1)

  • Restricting to ๐‘Œ and scaling down by a factor of

2, we are now in the well-spread case

Case #2 Continued

slide-53
SLIDE 53

Part VI: Open Problems

slide-54
SLIDE 54
  • Lower Bounds have been shown for this

semidefinite program

  • Khot and Vishnoi [KV05] proved the first super-

constant lower bound.

  • For weighted graphs, Naor and Young [NY17]

showed an ฮฉ ๐‘š๐‘๐‘•๐‘œ lower bound (which is tight up to a ๐‘š๐‘๐‘•๐‘š๐‘๐‘•๐‘œ factor).

  • However, these lower bounds donโ€™t apply even

to degree 4 SOS!

Lower Bounds

slide-55
SLIDE 55
  • Is this also true for unweighted graphs?
  • Does degree 4 SOS or higher degree SOS give

further improvements? Can we show a superconstant lower bound for a constant number of rounds of SOS?

Open Questions

slide-56
SLIDE 56

References

  • [ALN08] S. Arora, J. R. Lee, A. Naor. Euclidean distortion and the sparsest cut. J.
  • Amer. Math. Soc. 21 (1), p. 1โ€“21. 2008
  • [ARV] S. Arora, S. Rao, U. Vazirani. Expander Flows, Geometric Embeddings and

Graph Partitioning. https://www.cs.princeton.edu/~arora/pubs/arvfull.pdf

  • [KV05] S. Khot and N. Vishnoi. The unique games conjecture, integrality gap for cut

problems and embeddability of negative type metrics into ๐‘€1. FOCS 2005

  • [NY17] A. Naor and R. Young. The integrality gap of the Goemans--Linial SDP

relaxation for Sparsest Cut is at least a constant multiple of ๐‘š๐‘๐‘•๐‘œ. STOC 2017