Data Reduction for Maximum weight Matchings in Real World Graphs - - PowerPoint PPT Presentation

data reduction for maximum weight matchings in real world
SMART_READER_LITE
LIVE PREVIEW

Data Reduction for Maximum weight Matchings in Real World Graphs - - PowerPoint PPT Presentation

Data Reduction for Maximum weight Matchings in Real World Graphs Pitta Venkat Indian Institute of Technology Madras cs16b017@smail.iitm.ac.in November 1, 2019 Pitta Venkat (IITM) Data Reduction November 1, 2019 1 / 12 Definitions and


slide-1
SLIDE 1

Data Reduction for Maximum weight Matchings in Real World Graphs

Pitta Venkat

Indian Institute of Technology Madras cs16b017@smail.iitm.ac.in

November 1, 2019

Pitta Venkat (IITM) Data Reduction November 1, 2019 1 / 12

slide-2
SLIDE 2

Definitions and Notations

Feedback Edge Set

A set of Edges X in a Graph G is called feedback edge set if G − X is a forest.

Feedback Edge Number

Minimum Cardinality of all Feedback Edge Sets of Graph G is called Feedback Edge Number of G.

Notations

w(G):- Denotes weight of maximum weight Matching in Graph G. w(M):- Denotes weight of Matching M. Maximum weight Matching Problem :- Input:- weighted undirected Graph G and S. Output:- Yes iff ∃ Matching M in G s.t w(M) ≥ S.

Pitta Venkat (IITM) Data Reduction November 1, 2019 2 / 12

slide-3
SLIDE 3

Definitions and Notations

Parameterized problem

A Set of Instances (I, k) where I ∈ Σ∗ and K ∈ N is a parameter.

Equivalance of Instances

Instance (I, K) and Instance (I 1, K 1) of problem P are equivalent if (I, K) is an YES instance of P iff (I 1, K 1) is an YES instance of P.

Kernelization and Data Reduction Rules

Kernelization is an Algorithm that given an Instance (I, K) of parameterized problem P, computes in polynomial time equivalent Instance (I 1, K 1) s.t |I 1| + K 1 ≤ f (K) for some computable function f . If f (K) ∈ K O(1) then we say P admits a polynomial Kernel. Kernelization is generally acheived by applying poly-time executable Data reduction Rules. A Rule R is correct if on applying R on (I, K) results (I 1, K 1) equivalent to (I, K).

Pitta Venkat (IITM) Data Reduction November 1, 2019 3 / 12

slide-4
SLIDE 4

Kernelization for Maximum cardinality Matching in unweighted graphs

Reduction Rule 1

Let v ∈ V . If deg(v) = 0 then remove v from G. If deg(v) = 1 then remove v and it’s neighbour and decrease S by 1.

Reduction Rule 2

Let v ∈ V . If deg(v) = 2 and u, w are neighbour’s of v then remove v and merge u, w and decrease S by 1.

Theorm1

Maximum cardinality Matching admits linear-time computable linear-size Kernel with respective to parameter as feedback edge number.

Result from Theorm1

Total time taken to compute Maximum cardinality Matching is O(m + n + K 1.5).

Pitta Venkat (IITM) Data Reduction November 1, 2019 4 / 12

slide-5
SLIDE 5

Kernelization for Maximum weight Matching

Reduction Rule 1

Let v ∈ V . If deg(v) = 0 then remove v. Or Let e ∈ E If w(e) = 0 then remove e.

Reduction Rule 2

Let v ∈ V . If deg(v) = 1 and u is neighbour of v then ∀e ∈ E s.t u incident on e set weight of edge as max{0, w(e) − w(uv)}, remove v and decrease S by w(uv)

Lemma1

Rule 1 and Rule 2 are correct.

Lemma2

Rule 1 and Rule 2 can be applied exhaustively in linear time.

Pitta Venkat (IITM) Data Reduction November 1, 2019 5 / 12

slide-6
SLIDE 6

Proof of Lemma1

Correctness of Rule 1 :- is simple to see because by applying Rule 1 Matchings doesnot effected. Correctness of Rule 2 :- Let us consider v be with deg(v) = 1 and uv ∈ E, G 1 is the graph obtained by applying Rule 2. If ∃ Matching M s.t w(M) ≥ S. Let X be set of all edges incident on u, v and M1 = M − X. M1 exists in G 1 with same weight. if uv ∈ M then w(M1) ≥ S − w(uv). if e ∈ X − uv ∈ M then wG 1(M1 + e) ≥ S − w(e) + Max ≥ S − w(uv). Hence ∃ Matching M11 s.t wG 1(M11) ≥ S − w(uv). We can also proove other way also by following similar procedure. Hence Rule 2 is correct.

Pitta Venkat (IITM) Data Reduction November 1, 2019 6 / 12

slide-7
SLIDE 7

Proof of Lemma2

Rule 1 can be exauhstively applied in linear time by collecting all vertices of 0 degree and all edges of 0 weight in one reading. We will give an Algorithm to show Rule 2 can be exauhstively applied in linear time as follows. There are counters for each vertex initialised to 0. weight of edge xy at any iteration is w(xy) − c(x) − c(y). Each time when the Rule 2 is applied Let deg(x) = 1 and y is neighbour of x we will decrease S by max{0, w(xy) − c(x) − c(y)} and c(y) = c(y) + max{0, w(xy) − c(x) − c(y)}. So we will first collect all vertices of degree 1 in O(n + m) time. We will go through each vertex we collected and apply Rule 2. After that change all weights of edges as w(xy) = w(xy) + c(x) + c(y). correctness of Algorithm can be easily varified.

Pitta Venkat (IITM) Data Reduction November 1, 2019 7 / 12

slide-8
SLIDE 8

Definitions and Reduction Rules of degree 2 vertices

Maximal path

Let G be a Graph. A path P = v0v1..vl is a Maximal path in G if l ≥ 3. and deg(V1) = deg(v2) = .. = deg(vl−1) = 2, deg(v0) = 2, deg(vl) = 2.

Pending cycle

Let G be the Graph. A Cycle C = c0c1..cl is a pending cycle if atmost one vertex in cycle has degree greater than 2.

Reduction Rule 3

Let G be a graph and C be a pending cycle in G where u ∈ C with deg(u) ≥ 3.Then replace the Cycle by an edge uz with w(uz) = w(C) − w(C − u) and decrease S by w(C − u).

Pitta Venkat (IITM) Data Reduction November 1, 2019 8 / 12

slide-9
SLIDE 9

Reduction Rules and Lemma’s

Reduction Rule 4

Let G be a graph. P be a maximal path with end points u, v then remove all vertices in P execpt u, v and add vertex z. s.t w(uz) = w(P − v) − w(P − u − v), w(vz) = w(P − u) − w(p − u − v) and w(uv) = max{w(uv), w(P) − w(P − u − v)}. and decrease S by w(P − u − v)

Lemma 3

Maximal paths in a Graph is atmost feedback edge number.

Lemma 4

Reduction Rules 3 doesnot increase feedback edge number and Reduction Rule 4 can increase atmost double.

Lemma 5

Rules 3,4 are correct and can be applied exhaustively in linear time.

Pitta Venkat (IITM) Data Reduction November 1, 2019 9 / 12

slide-10
SLIDE 10

Kernel Size for Maximum weight Matching

Kernelization for Maximum weight Matching is exhaustively applying Rule 1 , Rule 3 , Rule 4 and Rule 2 in Order. Let G 1 be the final graph obtained K 1 ≤ 2 ∗ K because feedback edge number doesnot increase in any rule. Let X is minimum size feedback edge set in G 1. Consider the graph G 1 − X. Devide the vertices into V1, V2, V3 represents vertices of degree 1 , degree 2 and degree ≥ 3 respectively in G 1 − X. V3 < V1 (2V1 + 2V2 + 2V3 − 2 ≤ V1 + 2V2 + 3V3). in G 1 there are no degree 1 vertices so in G 1 − X degree 1 vertices are obtained by removing X so V1 < 4 ∗ K. degree 2 vertices in G are incident to X or V3 + 4 ∗ K so V2 < 12 ∗ K. Total vertices < 4 ∗ K + 4 ∗ K + 12 ∗ K = 20 ∗ K and Total Edges < 20 ∗ K + 2 ∗ K in G 1. So the Kernel as atmost 20 ∗ K vertices and atmost 22 ∗ K edges.

Pitta Venkat (IITM) Data Reduction November 1, 2019 10 / 12

slide-11
SLIDE 11

Correctness of Rules 3 and 4

First we consider the Instances before and After applying Reduction Rules From Matchings in G we will remove edges which are different from both instances and add some other edges We will Obtain relation ship like w(G 1) ≥ w(G) − w(C − u) and w(G 1) ≥ w(G) − w(P − u − v). From Matchings in G 1 we will remove edges which are different from both instances and add some other edges We will Obtain relation ship like w(G) ≥ w(G 1) + w(C − u) and w(G) ≥ w(G 1) + w(P − u − v). Hence we can conclude equivalance of 2 Instances.

Pitta Venkat (IITM) Data Reduction November 1, 2019 11 / 12

slide-12
SLIDE 12

Conclusion

Comparing Kolmogorov’s Algorithm with and without data Reductions the average speedup is 3800% in case of unweighted graphs. Comparing Kolmogorov’s Algorithm with and without data Reductions the average speedup is 30% in case of weighted graphs. because weighted graphs can reduce only 50% size on average where unweighted graphs can reduce 80% size. Hence weighted graph Kernelization is not as good as unweighted Graph Kernelization

Pitta Venkat (IITM) Data Reduction November 1, 2019 12 / 12