Probabilistic Analysis of Optimization Problems on Sparse Random - - PowerPoint PPT Presentation

probabilistic analysis of optimization problems on sparse
SMART_READER_LITE
LIVE PREVIEW

Probabilistic Analysis of Optimization Problems on Sparse Random - - PowerPoint PPT Presentation

Probabilistic Analysis of Optimization Problems on Sparse Random Shortest Path Metrics Stefan Klootwijk Joint work with Bodo Manthey September 2020 Optimization in practice Large scale optimization problems are hard to solve within


slide-1
SLIDE 1

Probabilistic Analysis of Optimization Problems on Sparse Random Shortest Path Metrics

Stefan Klootwijk

Joint work with Bodo Manthey

September 2020

slide-2
SLIDE 2

Optimization in practice

◮ Large scale optimization problems are hard to solve within reasonable time. ◮ Often heuristics are used to provide (non-optimal) solutions. ◮ Big gap between theoretical and actual performance! Some examples of worst case approximation ratios:

Greedy for Minimum-weight Perfect Matching: O(nlog2(3/2)) ≈ O(n0.58) Nearest Neighbor (greedy) for TSP: O(log(n)) 2-Opt (local search) for TSP: O(√n) etc.

Random Shortest Path Metrics 2

slide-3
SLIDE 3

Optimization in practice

◮ Large scale optimization problems are hard to solve within reasonable time. ◮ Often heuristics are used to provide (non-optimal) solutions. ◮ Big gap between theoretical and actual performance! ◮ Some examples of worst case approximation ratios:

◮ Greedy for Minimum-weight Perfect Matching: O(nlog2(3/2)) ≈ O(n0.58) ◮ Nearest Neighbor (greedy) for TSP: O(log(n)) ◮ 2-Opt (local search) for TSP: O(√n) ◮ etc.

Random Shortest Path Metrics 2

slide-4
SLIDE 4

Optimization in practice

◮ Large scale optimization problems are hard to solve within reasonable time. ◮ Often heuristics are used to provide (non-optimal) solutions. ◮ Big gap between theoretical and actual performance! ◮ Probabilistic analysis and other ‘beyond worst-case analysis’ methods are nowadays used for analysis of the performance of these heuristics. ◮ Interested in E ALG

OPT

  • (instead of E[ALG]

E[OPT]). Random Shortest Path Metrics 2

slide-5
SLIDE 5

Random (Metric) Spaces

random in [0, 1]2

8 1 4 15 11 8 2 3 7 4 2

independent edge lengths

Random Shortest Path Metrics 3

slide-6
SLIDE 6

Framework for Random Metric Spaces

◮ We look at different models for random metric spaces. ◮ We study them and analyse the performance of heuristics

  • n them.

◮ Goal:

◮ help choosing the right heuristic for a given problem; ◮ facilitate the design of better heuristics.

Random Shortest Path Metrics 4

slide-7
SLIDE 7

Random Shortest Path Metrics

◮ Graph G = (V , E) (on n vertices) ◮ Random ‘edge weights’ w(e) for all edges e ∈ E ◮ Distances d(u, v) given by the shortest u, v-path w.r.t. weights, for all vertices u, v ∈ V

◮ d(v, v) = 0 for all v ∈ V ◮ Symmetry: d(u, v) = d(v, u) for all u, v ∈ V ◮ Triangle inequality: d(u, v) ≤ d(u, s) + d(s, v) for all u, s, v ∈ V

Random Shortest Path Metrics 5

slide-8
SLIDE 8

Random Shortest Path Metrics – Example

15 8 3 7 2

A B C D E Random Shortest Path Metrics 6

slide-9
SLIDE 9

Random Shortest Path Metrics – Example

15 8 3 7 2

A B C D E

d A B C D E A B C D E

Random Shortest Path Metrics 6

slide-10
SLIDE 10

Random Shortest Path Metrics – Example

15 8 3 7 2

A B C D E

d A B C D E A B C D E

Random Shortest Path Metrics 6

slide-11
SLIDE 11

Random Shortest Path Metrics – Example

15 8 3 7 2 20

A B C D E

d A B C D E A 20 B C D E

Random Shortest Path Metrics 6

slide-12
SLIDE 12

Random Shortest Path Metrics – Example

15 8 3 7 2 20

A B C D E

d A B C D E A 20 3 B C D E

Random Shortest Path Metrics 6

slide-13
SLIDE 13

Random Shortest Path Metrics – Example

15 8 3 7 2 20

A B C D E

d A B C D E A 20 3 B C D E

Random Shortest Path Metrics 6

slide-14
SLIDE 14

Random Shortest Path Metrics – Example

15 8 3 7 2 20 1 3

A B C D E

d A B C D E A 20 3 13 B C D E

Random Shortest Path Metrics 6

slide-15
SLIDE 15

Random Shortest Path Metrics – Example

15 8 3 7 2 20 1 3

A B C D E

d A B C D E A 20 3 13 B C D E

Random Shortest Path Metrics 6

slide-16
SLIDE 16

Random Shortest Path Metrics – Example

15 8 3 7 2 20 1 3 11 XX

A B C D E

d A B C D E A 20 3 13 11 B C D E

Random Shortest Path Metrics 6

slide-17
SLIDE 17

Random Shortest Path Metrics – Example

15 8 3 7 2 20 1 3 11 XX 1 7 10 9

A B C D E

d A B C D E A 20 3 13 11 B 20 17 7 9 C 3 17 10 8 D 13 7 10 2 E 11 9 8 2

Random Shortest Path Metrics 6

slide-18
SLIDE 18

Random Shortest Path Metrics – Example

15 8 3 7 2 20 1 3 11 XX 1 7 10 9

A B C D E

d A B C D E A 20 3 13 11 B 20 17 7 9 C 3 17 10 8 D 13 7 10 2 E 11 9 8 2 ◮ Edge weights from (standard)exponential distribution ⇒ ‘memorylessness property’: P(X > s + t | X > t) = P(X > s) for all s, t ≥ 0. ⇒ ‘minimum property’: X1, . . . , Xk ∼ Exp(1) ⇒ min(Xi) ∼ Exp(k).

Random Shortest Path Metrics 6

slide-19
SLIDE 19

Random Shortest Path Metrics (RSPM)

◮ Graph G = (V , E) (on n vertices) ◮ Random ‘edge weights’ w(e) for all edges e ∈ E ◮ Distances d(u, v) given by the shortest u, v-path w.r.t. weights, for all vertices u, v ∈ V ◮ Also known as First Passage Percolation (FPP) ◮ A widely studied model, but (until recently) not used for probabilistic analysis

Random Shortest Path Metrics 7

slide-20
SLIDE 20

Related results

◮ Probabilistic analysis using RSPM on complete graphs proposed by Karp & Steele (1985)

Theorem (Bringmann, Engels, Manthey, Rao 2013)

On RSPM generated from complete graphs, the following heuristics have expected approximation ratio O(1): ◮ Greedy for Minimum-Distance Perfect Matching; ◮ Nearest Neighbor Heuristic for TSP; ◮ Insertion Heuristics for TSP (for any insertion rule R). Also a ‘trivial’ O(log(n)) approximation ratio for 2-opt for TSP, open question whether this can be improved.

Random Shortest Path Metrics 8

slide-21
SLIDE 21

Related results

◮ Recent efforts to adapt the model to a more realistic one.

Theorem (K., Manthey, Visser 2019)

On RSPM generated from (dense) Erd˝

  • s–R´

enyi random graphs, the following heuristics have expected approximation ratio O(1): ◮ Greedy for Minimum-Distance Perfect Matching; ◮ Nearest Neighbor Heuristic for TSP; ◮ Insertion Heuristics for TSP (for any insertion rule R).

Random Shortest Path Metrics 9

slide-22
SLIDE 22

Related results

◮ Recent efforts to adapt the model to a more realistic one.

Theorem (K., Manthey, Visser 2019)

On RSPM generated from (dense) Erd˝

  • s–R´

enyi random graphs, the following heuristics have expected approximation ratio O(1): ◮ Greedy for Minimum-Distance Perfect Matching; ◮ Nearest Neighbor Heuristic for TSP; ◮ Insertion Heuristics for TSP (for any insertion rule R). Next step: RSPM generated from sparse graphs. ◮ Start from grid graphs, because most studied in FPP.

Random Shortest Path Metrics 9

slide-23
SLIDE 23

Main Result

Theorem (K., Manthey 2020)

On RSPM generated from square grid graphs, the following heuristics have expected approximation ratio O(1): ◮ Greedy for Minimum-Distance Perfect Matching;∗ ◮ Nearest Neighbor Heuristic for TSP;∗ ◮ Insertion Heuristics for TSP (for any insertion rule R);∗ ◮ 2-opt for TSP (for any choice of the improvements).†

∗ Also for RSPM generated from a certain wide class of

sparse graphs.

† Also for RSPM generated from arbitrary sparse graphs. Random Shortest Path Metrics 10

slide-24
SLIDE 24

Main Result

Theorem (K., Manthey 2020)

On RSPM generated from square grid graphs, the following heuristics have expected approximation ratio O(1): ◮ Greedy for Minimum-Distance Perfect Matching;∗ ◮ Nearest Neighbor Heuristic for TSP;∗ ◮ Insertion Heuristics for TSP (for any insertion rule R);∗ ◮ 2-opt for TSP (for any choice of the improvements).† ◮ Remainder of this presentation:

◮ Idea for the 2-opt result; ◮ Quick sketch of the ‘road’ to the greedy matching result.

Random Shortest Path Metrics 10

slide-25
SLIDE 25

Idea for the 2-opt result

Observation

Consider the shortest paths corresponding to an arbitrary 2-optimal solution. Then, every edge of G is used at most twice (once per direction).

2-exchange

= ⇒

Random Shortest Path Metrics 11

slide-26
SLIDE 26

Idea for the 2-opt result

Observation

Consider the shortest paths corresponding to an arbitrary 2-optimal solution. Then, every edge of G is used at most twice (once per direction). ◮ Any 2-optimal solution has length at most twice the sum

  • f all edge weights, so E[WLO] ≤ O(n).

◮ Any TSP solution uses at least n − 1 different edge weights, so E[TSP] ≥ Ω(n).

Random Shortest Path Metrics 11

slide-27
SLIDE 27

Idea for the 2-opt result

Observation

Consider the shortest paths corresponding to an arbitrary 2-optimal solution. Then, every edge of G is used at most twice (once per direction). ◮ Any 2-optimal solution has length at most twice the sum

  • f all edge weights, so E[WLO] ≤ O(n).

◮ Any TSP solution uses at least n − 1 different edge weights, so E[TSP] ≥ Ω(n). ◮ E WLO TSP

  • = O(1)

Random Shortest Path Metrics 11

slide-28
SLIDE 28

RSPM (general graphs) – Structural properties

Theorem (Davis, Prieditis 1993)

Let G be a complete graph and let τk(v) denote the distance to the k-th closest vertex from v. Then, for any k and v, τk(v) ∼

k−1

  • i=1

Exp(i · (n − i)).

Random Shortest Path Metrics 12

slide-29
SLIDE 29

RSPM (general graphs) – Structural properties

Theorem (Davis, Prieditis 1993)

Let G be a complete graph and let τk(v) denote the distance to the k-th closest vertex from v. Then, for any k and v, τk(v) ∼

k−1

  • i=1

Exp(i · (n − i)).

Generalization

Suppose that |δ(U)| ≥ f (|U|) for some function f (·) and all U ⊆ V . Then, for any k ∈ [n] and any v ∈ V , τk(v)

k−1

  • i=1

Exp(f (i)).

Random Shortest Path Metrics 12

slide-30
SLIDE 30

RSPM (grid graphs) – cut sizes

Theorem (Bollob´ as, Leader 1991)

Let G be a finite square grid graph on n = N2 vertices. Then, for any U ⊆ V : |δ(U)| ≥      2

  • |U|

if |U| ≤ n/4, √n if n/4 ≤ |U| ≤ 3n/4, 2

  • n − |U|

if |U| ≥ 3n/4.

Random Shortest Path Metrics 13

slide-31
SLIDE 31

RSPM (grid graphs) – cut sizes

Theorem (Bollob´ as, Leader 1991)

Let G be a finite square grid graph on n = N2 vertices. Then, for any U ⊆ V : |δ(U)| ≥      2

  • |U|

if |U| ≤ n/4, √n if n/4 ≤ |U| ≤ 3n/4, 2

  • n − |U|

if |U| ≥ 3n/4.

Remark

All results that follow can be generalized to any family of graphs that satisfies |δ(U)| ≥ Ω(|U|ε) for all U ⊆ V with |U| ≤ cn (where ε, c ∈ (0, 1) are constants).

Random Shortest Path Metrics 13

slide-32
SLIDE 32

RSPM (grid graphs) – cut sizes

Theorem (Bollob´ as, Leader 1991)

Let G be a finite square grid graph on n = N2 vertices. Then, for any U ⊆ V : |δ(U)| ≥      2

  • |U|

if |U| ≤ n/4, √n if n/4 ≤ |U| ≤ 3n/4, 2

  • n − |U|

if |U| ≥ 3n/4.

Corollary

Let τk(v) denote the distance to the k-th closest vertex from

  • v. Then, for any k ≤ n/4 and any v ∈ V ,

τk(v)

k−1

  • i=1

Exp(2 √ i).

Random Shortest Path Metrics 13

slide-33
SLIDE 33

RSPM (general graphs) – ‘toolbox’

Theorem (Clustering)

Let ∆ > 0. If we partition the instance into clusters of diameter at most 4∆, then the expected number of clusters needed is O(1 + n/∆2).

Lemma (Tail bound for ∆max)

Let ∆max := maxu,v d(u, v). Then for x ≥ 9√n we have P(∆max ≥ x) = ne−x.

Random Shortest Path Metrics 14

slide-34
SLIDE 34

Greedy Matching

Random Shortest Path Metrics 15

slide-35
SLIDE 35

Greedy Matching

Random Shortest Path Metrics 15

slide-36
SLIDE 36

Greedy Matching

Random Shortest Path Metrics 15

slide-37
SLIDE 37

Greedy Matching

Random Shortest Path Metrics 15

slide-38
SLIDE 38

Greedy Matching

Random Shortest Path Metrics 15

slide-39
SLIDE 39

Greedy Matching

Random Shortest Path Metrics 15

slide-40
SLIDE 40

Greedy Matching

Random Shortest Path Metrics 15

slide-41
SLIDE 41

Greedy Matching

Random Shortest Path Metrics 15

slide-42
SLIDE 42

Greedy Matching

Random Shortest Path Metrics 15

slide-43
SLIDE 43

Greedy Matching

Random Shortest Path Metrics 15

slide-44
SLIDE 44

Greedy Matching

Lemma

Greedy outputs a matching with expected costs at most O(n).

Theorem

Greedy has an expected approximation ratio of O(1) on RSPM generated from grid graphs.

Random Shortest Path Metrics 15

slide-45
SLIDE 45

Proof idea

· · · · · ·

d ∈ (0, 4] d ∈ (4, 8] d ∈ (8, 12] d ∈ (4(i − 1), 4i] X1 X2 X3 Xi Yi Y3 Y2 Y1

Random Shortest Path Metrics 16

slide-46
SLIDE 46

Proof idea

· · · · · ·

d ∈ (0, 4] d ∈ (4, 8] d ∈ (8, 12] d ∈ (4(i − 1), 4i] X1 X2 X3 Xi Yi Y3 Y2 Y1 ◮ E[GR] ≤ ∞

i=1 4i · E[Xi] = ∞ i=1 4 · E[Yi].

◮ Y1 = n/2.

Random Shortest Path Metrics 16

slide-47
SLIDE 47

Proof idea

◮ E[GR] ≤ ∞

i=1 4i · E[Xi] = ∞ i=1 4 · E[Yi].

A partitioning in clusters of diameter ≤ 4(i − 1) needs ≤ O(1 + n/(i − 1)2) clusters. When ‘Greedy’ reaches bin i, at most O(1 + n/(i − 1)2) unmatched vertices remain. So E[Yi] ≤ O(1 + n/(i − 1)2) for i > 1.

· · · · · ·

d ∈ (0, 4] d ∈ (4, 8] d ∈ (8, 12] d ∈ (4(i − 1), 4i] X1 X2 X3 Xi Yi Y3 Y2 Y1

Random Shortest Path Metrics 17

slide-48
SLIDE 48

Proof idea

◮ E[GR] ≤ ∞

i=1 4i · E[Xi] = ∞ i=1 4 · E[Yi].

◮ A partitioning in clusters of diameter ≤ 4(i − 1) needs ≤ O(1 + n/(i − 1)2) clusters. When ‘Greedy’ reaches bin i, at most O(1 + n/(i − 1)2) unmatched vertices remain. So E[Yi] ≤ O(1 + n/(i − 1)2) for i > 1.

· · · · · ·

d ∈ (0, 4] d ∈ (4, 8] d ∈ (8, 12] d ∈ (4(i − 1), 4i] X1 X2 X3 Xi Yi Y3 Y2 Y1

Random Shortest Path Metrics 17

slide-49
SLIDE 49

Proof idea

◮ E[GR] ≤ ∞

i=1 4i · E[Xi] = ∞ i=1 4 · E[Yi].

◮ A partitioning in clusters of diameter ≤ 4(i − 1) needs ≤ O(1 + n/(i − 1)2) clusters. ◮ When ‘Greedy’ reaches bin i, at most O(1 + n/(i − 1)2) unmatched vertices remain. So E[Yi] ≤ O(1 + n/(i − 1)2) for i > 1.

· · · · · ·

d ∈ (0, 4] d ∈ (4, 8] d ∈ (8, 12] d ∈ (4(i − 1), 4i] X1 X2 X3 Xi Yi Y3 Y2 Y1

Random Shortest Path Metrics 17

slide-50
SLIDE 50

Proof idea

◮ E[GR] ≤ ∞

i=1 4i · E[Xi] = ∞ i=1 4 · E[Yi].

◮ A partitioning in clusters of diameter ≤ 4(i − 1) needs ≤ O(1 + n/(i − 1)2) clusters. ◮ When ‘Greedy’ reaches bin i, at most O(1 + n/(i − 1)2) unmatched vertices remain. ◮ So E[Yi] ≤ O(1 + n/(i − 1)2) for i > 1.

· · · · · ·

d ∈ (0, 4] d ∈ (4, 8] d ∈ (8, 12] d ∈ (4(i − 1), 4i] X1 X2 X3 Xi Yi Y3 Y2 Y1

Random Shortest Path Metrics 17

slide-51
SLIDE 51

Proof idea

◮ E[GR] ≤ ∞

i=1 4i · E[Xi] = ∞ i=1 4 · E[Yi].

◮ A partitioning in clusters of diameter ≤ 4(i − 1) needs ≤ O(1 + n/(i − 1)2) clusters. ◮ When ‘Greedy’ reaches bin i, at most O(1 + n/(i − 1)2) unmatched vertices remain. ◮ So E[Yi] ≤ O(1 + n/(i − 1)2) for i > 1. ◮ For ‘large’ i we have E[Yi] ≤ n · P(∆max ≥ 4(i − 1)) ≤ n2e−4(i−1). Summing over all i yields E[GR] ≤ O(n).

Random Shortest Path Metrics 17

slide-52
SLIDE 52

Proof idea

◮ E[GR] ≤ ∞

i=1 4i · E[Xi] = ∞ i=1 4 · E[Yi].

◮ A partitioning in clusters of diameter ≤ 4(i − 1) needs ≤ O(1 + n/(i − 1)2) clusters. ◮ When ‘Greedy’ reaches bin i, at most O(1 + n/(i − 1)2) unmatched vertices remain. ◮ So E[Yi] ≤ O(1 + n/(i − 1)2) for i > 1. ◮ For ‘large’ i we have E[Yi] ≤ n · P(∆max ≥ 4(i − 1)) ≤ n2e−4(i−1). ◮ Summing over all i yields E[GR] ≤ O(n).

Random Shortest Path Metrics 17

slide-53
SLIDE 53

Open problems

◮ RSPM on arbitrary sparse graphs? ◮ Only using a subset of the vertices? ◮ ‘Hybrid heuristics’? ◮ ...

Random Shortest Path Metrics 18