cs 473 algorithms
play

CS 473: Algorithms Ruta Mehta University of Illinois, - PowerPoint PPT Presentation

CS 473: Algorithms Ruta Mehta University of Illinois, Urbana-Champaign Spring 2018 Ruta (UIUC) CS473 1 Spring 2018 1 / 53 CS 473: Algorithms, Spring 2018 Introduction to Randomized Algorithms: QuickSort Lecture 7 Feb 6, 2018 Most


  1. Independent Events Definition Given a probability space (Ω , Pr) and two events A , B are independent if and only if Pr[ A ∩ B ] = Pr[ A ] Pr[ B ] . Otherwise they are dependent . In other words A , B independent implies one does not affect the other. Example Two coins. Ω = { HH , TT , HT , TH } and Pr[ HH ] = Pr[ TT ] = Pr[ HT ] = Pr[ TH ] = 1 / 4 . A : the first coin is heads. B : second coin is tails. 1 Pr[ A ] =1 / 2 , Pr[ B ] = 1 / 2 , Pr[ A ∩ B ] = Ruta (UIUC) CS473 13 Spring 2018 13 / 53

  2. Independent Events Definition Given a probability space (Ω , Pr) and two events A , B are independent if and only if Pr[ A ∩ B ] = Pr[ A ] Pr[ B ] . Otherwise they are dependent . In other words A , B independent implies one does not affect the other. Example Two coins. Ω = { HH , TT , HT , TH } and Pr[ HH ] = Pr[ TT ] = Pr[ HT ] = Pr[ TH ] = 1 / 4 . A : the first coin is heads. B : second coin is tails. 1 Pr[ A ] =1 / 2 , Pr[ B ] = 1 / 2 , Pr[ A ∩ B ] =1 / 4 . independent. Ruta (UIUC) CS473 13 Spring 2018 13 / 53

  3. Independent Events Definition Given a probability space (Ω , Pr) and two events A , B are independent if and only if Pr[ A ∩ B ] = Pr[ A ] Pr[ B ] . Otherwise they are dependent . In other words A , B independent implies one does not affect the other. Example Two coins. Ω = { HH , TT , HT , TH } and Pr[ HH ] = Pr[ TT ] = Pr[ HT ] = Pr[ TH ] = 1 / 4 . A : the first coin is heads. B : second coin is tails. 1 Pr[ A ] =1 / 2 , Pr[ B ] = 1 / 2 , Pr[ A ∩ B ] =1 / 4 . independent. A : both are not tails. B : second coin is heads. 2 Pr[ A ] = Ruta (UIUC) CS473 13 Spring 2018 13 / 53

  4. Independent Events Definition Given a probability space (Ω , Pr) and two events A , B are independent if and only if Pr[ A ∩ B ] = Pr[ A ] Pr[ B ] . Otherwise they are dependent . In other words A , B independent implies one does not affect the other. Example Two coins. Ω = { HH , TT , HT , TH } and Pr[ HH ] = Pr[ TT ] = Pr[ HT ] = Pr[ TH ] = 1 / 4 . A : the first coin is heads. B : second coin is tails. 1 Pr[ A ] =1 / 2 , Pr[ B ] = 1 / 2 , Pr[ A ∩ B ] =1 / 4 . independent. A : both are not tails. B : second coin is heads. 2 Pr[ A ] =3 / 4 , Pr[ B ] = 1 / 2 , Pr[ A ∩ B ] = Ruta (UIUC) CS473 13 Spring 2018 13 / 53

  5. Independent Events Definition Given a probability space (Ω , Pr) and two events A , B are independent if and only if Pr[ A ∩ B ] = Pr[ A ] Pr[ B ] . Otherwise they are dependent . In other words A , B independent implies one does not affect the other. Example Two coins. Ω = { HH , TT , HT , TH } and Pr[ HH ] = Pr[ TT ] = Pr[ HT ] = Pr[ TH ] = 1 / 4 . A : the first coin is heads. B : second coin is tails. 1 Pr[ A ] =1 / 2 , Pr[ B ] = 1 / 2 , Pr[ A ∩ B ] =1 / 4 . independent. A : both are not tails. B : second coin is heads. 2 Pr[ A ] =3 / 4 , Pr[ B ] = 1 / 2 , Pr[ A ∩ B ] =1 / 2 . dependent . Ruta (UIUC) CS473 13 Spring 2018 13 / 53

  6. Union bound The probability of the union of two events, is no bigger than the sum of their probabilities. Lemma For any two events E and F , we have that � � � � � � Pr E ∪ F ≤ Pr E + Pr F . Proof. Consider E and F to be a collection of elmentery events (which they are). We have � � � Pr E ∪ F = Pr[ x ] x ∈ E ∪ F � � � � � � ≤ Pr[ x ] + Pr[ x ] = Pr E + Pr F . x ∈ E x ∈ F Ruta (UIUC) CS473 14 Spring 2018 14 / 53

  7. Random Variables Definition (Random Variable) Given a probability space (Ω , Pr) a (real-valued) random variable X over Ω is a function that maps each elementary event to a real number. In other words X : Ω → R . Ruta (UIUC) CS473 15 Spring 2018 15 / 53

  8. Random Variables Definition (Random Variable) Given a probability space (Ω , Pr) a (real-valued) random variable X over Ω is a function that maps each elementary event to a real number. In other words X : Ω → R . Definition (Expectation) For a random variable X over a probability space (Ω , Pr) the expectation of X is defined as � ω ∈ Ω Pr[ ω ] X ( ω ) . In other words, the expectation is the average value of X according to the probabilities given by Pr[ · ] . Ruta (UIUC) CS473 15 Spring 2018 15 / 53

  9. Expectation Example A 6-sided unbiased die. Ω = { 1 , 2 , 3 , 4 , 5 , 6 } and Pr[ i ] = 1 / 6 for each i ∈ Ω . X : Ω → R where X ( i ) = i mod 2 ∈ { 0 , 1 } . 1 Ruta (UIUC) CS473 16 Spring 2018 16 / 53

  10. Expectation Example A 6-sided unbiased die. Ω = { 1 , 2 , 3 , 4 , 5 , 6 } and Pr[ i ] = 1 / 6 for each i ∈ Ω . X : Ω → R where X ( i ) = i mod 2 ∈ { 0 , 1 } . Then 1 E[ X ] = � 6 � 6 i =1 Pr[ i ] · X ( i ) = 1 i =1 X ( i ) = 1 / 2 . 6 Ruta (UIUC) CS473 16 Spring 2018 16 / 53

  11. Expectation Example A 6-sided unbiased die. Ω = { 1 , 2 , 3 , 4 , 5 , 6 } and Pr[ i ] = 1 / 6 for each i ∈ Ω . X : Ω → R where X ( i ) = i mod 2 ∈ { 0 , 1 } . Then 1 E[ X ] = � 6 � 6 i =1 Pr[ i ] · X ( i ) = 1 i =1 X ( i ) = 1 / 2 . 6 Y : Ω → R where Y ( i ) = i 2 . 2 Ruta (UIUC) CS473 16 Spring 2018 16 / 53

  12. Expectation Example A 6-sided unbiased die. Ω = { 1 , 2 , 3 , 4 , 5 , 6 } and Pr[ i ] = 1 / 6 for each i ∈ Ω . X : Ω → R where X ( i ) = i mod 2 ∈ { 0 , 1 } . Then 1 E[ X ] = � 6 � 6 i =1 Pr[ i ] · X ( i ) = 1 i =1 X ( i ) = 1 / 2 . 6 Y : Ω → R where Y ( i ) = i 2 . Then 2 6 · i 2 = 91 / 6 . E[ Y ] = � 6 1 i =1 Ruta (UIUC) CS473 16 Spring 2018 16 / 53

  13. Expected number of vertices? Let G = ( V , E ) be a graph with n vertices and m edges. Let H be the graph resulting from independently deleting every vertex of G with probability 1 / 2 . Compute the expected number of vertices in H. (A) n/2. (B) n/4. (C) m/2. (D) m/4. (E) none of the above. Ruta (UIUC) CS473 17 Spring 2018 17 / 53

  14. Expected number of vertices is: Probability Space Ω = { 0 , 1 } n . For ω ∈ { 0 , 1 } n , ω v = 1 if vertex v is present in H, else is zero. 1 For each ω ∈ Ω , Pr[ ω ] = 2 n . Ruta (UIUC) CS473 18 Spring 2018 18 / 53

  15. Expected number of vertices is: Probability Space Ω = { 0 , 1 } n . For ω ∈ { 0 , 1 } n , ω v = 1 if vertex v is present in H, else is zero. 1 For each ω ∈ Ω , Pr[ ω ] = 2 n . X ( ω ) = # vertices in H as per ω = # 1s in ω . Ruta (UIUC) CS473 18 Spring 2018 18 / 53

  16. Expected number of vertices is: Probability Space Ω = { 0 , 1 } n . For ω ∈ { 0 , 1 } n , ω v = 1 if vertex v is present in H, else is zero. 1 For each ω ∈ Ω , Pr[ ω ] = 2 n . X ( ω ) = # vertices in H as per ω = # 1s in ω . � E[ X ] = ω ∈ Ω Pr[ ω ] X ( ω ) = � ω ∈ Ω 1 / 2 n X ( ω ) 1 / 2 n � n � n � = k k =0 k 1 / 2 n (2 n n = 2 ) = n / 2 Ruta (UIUC) CS473 18 Spring 2018 18 / 53

  17. Expected number of edges? Let G = ( V , E ) be a graph with n vertices and m edges. Let H be the graph resulting from independently deleting every vertex of G with probability 1 / 2 . The expected number of edges in H is (A) n/2. (B) n/4. (C) m/2. (D) m/4. (E) none of the above. Ruta (UIUC) CS473 19 Spring 2018 19 / 53

  18. Expected number of edges is: Probability Space Ω = { 0 , 1 } n . For ω ∈ { 0 , 1 } n , ω v = 1 if vertex v is present in H, else is zero. 1 For each ω ∈ Ω , Pr[ ω ] = 2 n . Ruta (UIUC) CS473 20 Spring 2018 20 / 53

  19. Expected number of edges is: Probability Space Ω = { 0 , 1 } n . For ω ∈ { 0 , 1 } n , ω v = 1 if vertex v is present in H, else is zero. 1 For each ω ∈ Ω , Pr[ ω ] = 2 n . X ( ω ) = # edges present in H as per ω = ?? Ruta (UIUC) CS473 20 Spring 2018 20 / 53

  20. Expected number of edges is: Probability Space Ω = { 0 , 1 } n . For ω ∈ { 0 , 1 } n , ω v = 1 if vertex v is present in H, else is zero. 1 For each ω ∈ Ω , Pr[ ω ] = 2 n . X ( ω ) = # edges present in H as per ω = ?? How to compute E[ X ] ? Ruta (UIUC) CS473 20 Spring 2018 20 / 53

  21. Indicator Random Variables Definition A binary random variable is one that takes on values in { 0 , 1 } . Ruta (UIUC) CS473 21 Spring 2018 21 / 53

  22. Indicator Random Variables Definition A binary random variable is one that takes on values in { 0 , 1 } . Special type of random variables that are quite useful. Definition Given a probability space (Ω , Pr) and an event A ⊆ Ω the indicator random variable X A is a binary random variable where X A ( ω ) = 1 if ω ∈ A and X A ( ω ) = 0 if ω �∈ A . Ruta (UIUC) CS473 21 Spring 2018 21 / 53

  23. Indicator Random Variables Definition A binary random variable is one that takes on values in { 0 , 1 } . Special type of random variables that are quite useful. Definition Given a probability space (Ω , Pr) and an event A ⊆ Ω the indicator random variable X A is a binary random variable where X A ( ω ) = 1 if ω ∈ A and X A ( ω ) = 0 if ω �∈ A . Example A 6-sided unbiased die. Ω = { 1 , 2 , 3 , 4 , 5 , 6 } and Pr[ i ] = 1 / 6 for each i ∈ Ω . Let A be the even that i is divisible by 3 , i.e., A = { 3 , 6 } . Then X A ( i ) = 1 if i ∈ { 3 , 6 } and 0 otherwise. Ruta (UIUC) CS473 21 Spring 2018 21 / 53

  24. Expectation Proposition For an indicator variable X A , E[ X A ] = Pr[ A ] . Proof. � E[ X A ] = X A ( ω ) Pr[ ω ] ω ∈ Ω � � = 1 · Pr[ ω ] + 0 · Pr[ ω ] ω ∈ A ω ∈ Ω \ A � = Pr[ ω ] ω ∈ A = Pr[ A ] . Ruta (UIUC) CS473 22 Spring 2018 22 / 53

  25. Linearity of Expectation Lemma Let X , Y be two random variables (not necessarily independent) over a probability space (Ω , Pr) . Then E[ X + Y ] = E[ X ] + E[ Y ] . Proof. � E[ X + Y ] = Pr[ ω ] ( X ( ω ) + Y ( ω )) ω ∈ Ω � � = Pr[ ω ] X ( ω ) + Pr[ ω ] Y ( ω ) = E[ X ] + E[ Y ] . ω ∈ Ω ω ∈ Ω Ruta (UIUC) CS473 23 Spring 2018 23 / 53

  26. Linearity of Expectation Lemma Let X , Y be two random variables (not necessarily independent) over a probability space (Ω , Pr) . Then E[ X + Y ] = E[ X ] + E[ Y ] . Proof. � E[ X + Y ] = Pr[ ω ] ( X ( ω ) + Y ( ω )) ω ∈ Ω � � = Pr[ ω ] X ( ω ) + Pr[ ω ] Y ( ω ) = E[ X ] + E[ Y ] . ω ∈ Ω ω ∈ Ω Corollary E[ a 1 X 1 + a 2 X 2 + . . . + a n X n ] = � n i =1 a i E[ X i ] . Ruta (UIUC) CS473 23 Spring 2018 23 / 53

  27. Expected number of edges? Let G = ( V , E ) be a graph with n vertices and m edges. Let H be the graph resulting from independently deleting every vertex of G with probability 1 / 2 . The expected number of edges in H is Ruta (UIUC) CS473 24 Spring 2018 24 / 53

  28. Expected number of edges? Let G = ( V , E ) be a graph with n vertices and m edges. Let H be the graph resulting from independently deleting every vertex of G with probability 1 / 2 . The expected number of edges in H is Event A e = edge e ∈ E is present in H. � � Pr = Pr[ u and v both are present ] = A e =( u , v ) Pr[ u is present ] · Pr[ v is present ] = 1 2 · 1 2 = 1 4 . Ruta (UIUC) CS473 24 Spring 2018 24 / 53

  29. Expected number of edges? Let G = ( V , E ) be a graph with n vertices and m edges. Let H be the graph resulting from independently deleting every vertex of G with probability 1 / 2 . The expected number of edges in H is Event A e = edge e ∈ E is present in H. � � Pr = Pr[ u and v both are present ] = A e =( u , v ) Pr[ u is present ] · Pr[ v is present ] = 1 2 · 1 2 = 1 4 . X A e indicator random variables, then E[ X A e ] = Pr[ A e ] . Ruta (UIUC) CS473 24 Spring 2018 24 / 53

  30. Expected number of edges? Let G = ( V , E ) be a graph with n vertices and m edges. Let H be the graph resulting from independently deleting every vertex of G with probability 1 / 2 . The expected number of edges in H is Event A e = edge e ∈ E is present in H. � � Pr = Pr[ u and v both are present ] = A e =( u , v ) Pr[ u is present ] · Pr[ v is present ] = 1 2 · 1 2 = 1 4 . X A e indicator random variables, then E[ X A e ] = Pr[ A e ] . Let X = � e ∈ E X A e (Number of edges in H) Ruta (UIUC) CS473 24 Spring 2018 24 / 53

  31. Expected number of edges? Let G = ( V , E ) be a graph with n vertices and m edges. Let H be the graph resulting from independently deleting every vertex of G with probability 1 / 2 . The expected number of edges in H is Event A e = edge e ∈ E is present in H. � � Pr = Pr[ u and v both are present ] = A e =( u , v ) Pr[ u is present ] · Pr[ v is present ] = 1 2 · 1 2 = 1 4 . X A e indicator random variables, then E[ X A e ] = Pr[ A e ] . Let X = � e ∈ E X A e (Number of edges in H) �� � Pr[ A e ] = m � � E[ X ] = E = E[ X A e ] = X A e 4 e ∈ E e ∈ E e ∈ E Ruta (UIUC) CS473 24 Spring 2018 24 / 53

  32. Expected number of edges? Let G = ( V , E ) be a graph with n vertices and m edges. Let H be the graph resulting from independently deleting every vertex of G with probability 1 / 2 . The expected number of edges in H is Event A e = edge e ∈ E is present in H. � � Pr = Pr[ u and v both are present ] = A e =( u , v ) Pr[ u is present ] · Pr[ v is present ] = 1 2 · 1 2 = 1 4 . X A e indicator random variables, then E[ X A e ] = Pr[ A e ] . Let X = � e ∈ E X A e (Number of edges in H) �� � Pr[ A e ] = m � � E[ X ] = E = E[ X A e ] = X A e 4 e ∈ E e ∈ E e ∈ E It is important to setup random variables carefully. Ruta (UIUC) CS473 24 Spring 2018 24 / 53

  33. Expected number of triangles? Let G = ( V , E ) be a graph with n vertices and m edges. Assume G has t triangles (i.e., a triangle is a simple cycle with three vertices). Let H be the graph resulting from deleting independently each vertex of G with probability 1 / 2 . The expected number of triangles in H is (A) t/2. (B) t/4. (C) t/8. (D) t/16. (E) none of the above. Ruta (UIUC) CS473 25 Spring 2018 25 / 53

  34. Independent Random Variables Definition Random variables X , Y are said to be independent if ∀ x , y ∈ R , Pr[ X = x ∧ Y = y ] = Pr[ X = x ] Pr[ Y = y ] . Ruta (UIUC) CS473 26 Spring 2018 26 / 53

  35. Independent Random Variables Definition Random variables X , Y are said to be independent if ∀ x , y ∈ R , Pr[ X = x ∧ Y = y ] = Pr[ X = x ] Pr[ Y = y ] . Examples Two independent un-biased coin flips: Ω = { HH , HT , TH , TT } . X = 1 if first coin is H else 0 . Y = 1 if second coin is H else 0 . Ruta (UIUC) CS473 26 Spring 2018 26 / 53

  36. Independent Random Variables Definition Random variables X , Y are said to be independent if ∀ x , y ∈ R , Pr[ X = x ∧ Y = y ] = Pr[ X = x ] Pr[ Y = y ] . Examples Two independent un-biased coin flips: Ω = { HH , HT , TH , TT } . X = 1 if first coin is H else 0 . Y = 1 if second coin is H else 0 . Independent. Ruta (UIUC) CS473 26 Spring 2018 26 / 53

  37. Independent Random Variables Definition Random variables X , Y are said to be independent if ∀ x , y ∈ R , Pr[ X = x ∧ Y = y ] = Pr[ X = x ] Pr[ Y = y ] . Examples Two independent un-biased coin flips: Ω = { HH , HT , TH , TT } . X = 1 if first coin is H else 0 . Y = 1 if second coin is H else 0 . Independent. X = # H , Y = # T . Ruta (UIUC) CS473 26 Spring 2018 26 / 53

  38. Independent Random Variables Definition Random variables X , Y are said to be independent if ∀ x , y ∈ R , Pr[ X = x ∧ Y = y ] = Pr[ X = x ] Pr[ Y = y ] . Examples Two independent un-biased coin flips: Ω = { HH , HT , TH , TT } . X = 1 if first coin is H else 0 . Y = 1 if second coin is H else 0 . Independent. X = # H , Y = # T . Dependent. Why? Ruta (UIUC) CS473 26 Spring 2018 26 / 53

  39. Independent Random Variables Lemma If X and Y are independent then E[ X · Y ] = E[ X ] · E[ Y ] Proof. � E[ X · Y ] = Pr[ ω ] ( X ( ω ) · Y ( ω )) ω ∈ Ω � = Pr[ X = x ∧ Y = y ] ( x · y ) x , y ∈ R � = Pr[ X = x ] · Pr[ Y = y ] · x · y x , y ∈ R � � = ( Pr[ X = x ] x )( Pr[ Y = y ] y ) = E[ X ] E[ Y ] x ∈ R y ∈ R Ruta (UIUC) CS473 27 Spring 2018 27 / 53

  40. Types of Randomized Algorithms Typically one encounters the following types: Las Vegas randomized algorithms: for a given input x 1 output of algorithm is always correct but the running time is a random variable . In this case we are interested in analyzing the expected running time. Ruta (UIUC) CS473 28 Spring 2018 28 / 53

  41. Types of Randomized Algorithms Typically one encounters the following types: Las Vegas randomized algorithms: for a given input x 1 output of algorithm is always correct but the running time is a random variable . In this case we are interested in analyzing the expected running time. Monte Carlo randomized algorithms: for a given input x the 2 running time is deterministic but the output is random; correct with some probability . In this case we are interested in analyzing the probability of the correct output (and also the running time). Algorithms whose running time and output may both be random. 3 Ruta (UIUC) CS473 28 Spring 2018 28 / 53

  42. Analyzing Las Vegas Algorithms Deterministic algorithm Q for a problem Π : Let Q ( x ) be the time for Q to run on input x of length | x | . 1 Worst-case analysis: run time on worst input for a given size n . 2 T wc ( n ) = max x : | x | = n Q ( x ) . Ruta (UIUC) CS473 29 Spring 2018 29 / 53

  43. Analyzing Las Vegas Algorithms Deterministic algorithm Q for a problem Π : Let Q ( x ) be the time for Q to run on input x of length | x | . 1 Worst-case analysis: run time on worst input for a given size n . 2 T wc ( n ) = max x : | x | = n Q ( x ) . Randomized algorithm R for a problem Π : Let R ( x ) be the time for Q to run on input x of length | x | . 1 R ( x ) is a random variable: depends on random bits used by R . 2 E[ R ( x )] is the expected running time for R on x 3 Ruta (UIUC) CS473 29 Spring 2018 29 / 53

  44. Analyzing Las Vegas Algorithms Deterministic algorithm Q for a problem Π : Let Q ( x ) be the time for Q to run on input x of length | x | . 1 Worst-case analysis: run time on worst input for a given size n . 2 T wc ( n ) = max x : | x | = n Q ( x ) . Randomized algorithm R for a problem Π : Let R ( x ) be the time for Q to run on input x of length | x | . 1 R ( x ) is a random variable: depends on random bits used by R . 2 E[ R ( x )] is the expected running time for R on x 3 Worst-case analysis: expected time on worst input of size n 4 T rand − wc ( n ) = max x : | x | = n E[ R ( x )] . Ruta (UIUC) CS473 29 Spring 2018 29 / 53

  45. Analyzing Monte Carlo Algorithms Randomized algorithm M for a problem Π : Let M ( x ) be the time for M to run on input x of length | x | . 1 For Monte Carlo, assumption is that run time is deterministic. Let Pr[ x ] be the probability that M is correct on x . 2 Pr[ x ] is a random variable: depends on random bits used by M . 3 Ruta (UIUC) CS473 30 Spring 2018 30 / 53

  46. Analyzing Monte Carlo Algorithms Randomized algorithm M for a problem Π : Let M ( x ) be the time for M to run on input x of length | x | . 1 For Monte Carlo, assumption is that run time is deterministic. Let Pr[ x ] be the probability that M is correct on x . 2 Pr[ x ] is a random variable: depends on random bits used by M . 3 Worst-case analysis: success probability on worst input 4 P rand − wc ( n ) = min x : | x | = n Pr[ x ] . Ruta (UIUC) CS473 30 Spring 2018 30 / 53

  47. Part III Why does randomization help? Ruta (UIUC) CS473 31 Spring 2018 31 / 53

  48. Ping and find. Consider a deterministic algorithm A that is trying to find an element in an array X of size n . At every step it is allowed to ask the value of one cell in the array, and the adversary is allowed after each such ping, to shuffle elements around in the array in any way it seems fit. For the best possible deterministic algorithm the number of rounds it has to play this game till it finds the required element is (A) O (1) (B) O ( n ) (C) O ( n log n ) (D) O ( n 2 ) (E) ∞ . Ruta (UIUC) CS473 32 Spring 2018 32 / 53

  49. Ping and find randomized. Consider an algorithm randFind that is trying to find an element in an array X of size n . At every step it asks the value of one random cell in the array, and the adversary is allowed after each such ping, to shuffle elements around in the array in any way it seems fit. This algorithm would stop in expectation after (A) O (1) (B) O (log n ) (C) O ( n ) (D) O ( n 2 ) (E) ∞ . steps. Ruta (UIUC) CS473 33 Spring 2018 33 / 53

  50. Abundance of witnesses Consider the problem of finding an “approximate median” of an unsorted array A [1 .. n ] : an element of A with rank between n / 4 and 3 n / 4 . Finding an approximate median is not any easier than a proper median. Ruta (UIUC) CS473 34 Spring 2018 34 / 53

  51. Abundance of witnesses Consider the problem of finding an “approximate median” of an unsorted array A [1 .. n ] : an element of A with rank between n / 4 and 3 n / 4 . Finding an approximate median is not any easier than a proper median. n / 2 elements of A qualify as approximate medians and hence a random element is good with probability 1 / 2 ! Ruta (UIUC) CS473 34 Spring 2018 34 / 53

  52. Part IV Randomized Quick Sort Ruta (UIUC) CS473 35 Spring 2018 35 / 53

  53. QuickSort Deterministic QuickSort Pick a pivot element from array 1 Split array into 3 subarrays: those smaller than pivot, those 2 larger than pivot, and the pivot itself. Recursively sort the subarrays, and concatenate them. 3 Ruta (UIUC) CS473 36 Spring 2018 36 / 53

  54. QuickSort Deterministic QuickSort Pick a pivot element from array 1 Split array into 3 subarrays: those smaller than pivot, those 2 larger than pivot, and the pivot itself. Recursively sort the subarrays, and concatenate them. 3 Randomized QuickSort Pick a pivot element uniformly at random from the array 1 Split array into 3 subarrays: those smaller than pivot, those 2 larger than pivot, and the pivot itself. Recursively sort the subarrays, and concatenate them. 3 Ruta (UIUC) CS473 36 Spring 2018 36 / 53

  55. Randomized Quicksort Recall: Deterministic QuickSort can take Ω( n 2 ) time to sort array of size n . Ruta (UIUC) CS473 37 Spring 2018 37 / 53

  56. Randomized Quicksort Recall: Deterministic QuickSort can take Ω( n 2 ) time to sort array of size n . Theorem Randomized QuickSort sorts a given array of length n in O ( n log n ) expected time. Ruta (UIUC) CS473 37 Spring 2018 37 / 53

  57. Randomized Quicksort Recall: Deterministic QuickSort can take Ω( n 2 ) time to sort array of size n . Theorem Randomized QuickSort sorts a given array of length n in O ( n log n ) expected time. Note: On every input randomized QuickSort takes O ( n log n ) time in expectation. On every input it may take Ω( n 2 ) time with some small probability. Ruta (UIUC) CS473 37 Spring 2018 37 / 53

  58. Randomized QuickSort Randomized QuickSort Pick a pivot element uniformly at random from the array. 1 Split array into 3 subarrays: those smaller than pivot, those 2 larger than pivot, and the pivot itself. Recursively sort the subarrays, and concatenate them. 3 Ruta (UIUC) CS473 38 Spring 2018 38 / 53

  59. Analysis What events to count? Number of Comparisions. Ruta (UIUC) CS473 39 Spring 2018 39 / 53

  60. Analysis What events to count? Number of Comparisions. What is the probability space? All the coin tosses at all levels and parts of recursion. Ruta (UIUC) CS473 39 Spring 2018 39 / 53

  61. Analysis What events to count? Number of Comparisions. What is the probability space? All the coin tosses at all levels and parts of recursion. Too Big!! Ruta (UIUC) CS473 39 Spring 2018 39 / 53

  62. Analysis What events to count? Number of Comparisions. What is the probability space? All the coin tosses at all levels and parts of recursion. Too Big!! What random variables to define? What are the events of the algorithm? Ruta (UIUC) CS473 39 Spring 2018 39 / 53

  63. Analysis via Recurrence Given array A of n distinct numbers. 1 Q ( A ) : number of comparisons of randomized QuickSort on A . 2 Note that Q ( A ) is a random variable. Ruta (UIUC) CS473 40 Spring 2018 40 / 53

  64. Analysis via Recurrence Given array A of n distinct numbers. 1 Q ( A ) : number of comparisons of randomized QuickSort on A . 2 Note that Q ( A ) is a random variable. X i : Indicator random variable, which is set to 1 if pivot is of 3 rank i in A , else zero. Let A i left and A i right be the corresponding left and right subarrays. Ruta (UIUC) CS473 40 Spring 2018 40 / 53

  65. Analysis via Recurrence Given array A of n distinct numbers. 1 Q ( A ) : number of comparisons of randomized QuickSort on A . 2 Note that Q ( A ) is a random variable. X i : Indicator random variable, which is set to 1 if pivot is of 3 rank i in A , else zero. Let A i left and A i right be the corresponding left and right subarrays. n � � � Q ( A i left ) + Q ( A i Q ( A ) = n + X i · right ) . i =1 Ruta (UIUC) CS473 40 Spring 2018 40 / 53

  66. Analysis via Recurrence Given array A of n distinct numbers. 1 Q ( A ) : number of comparisons of randomized QuickSort on A . 2 Note that Q ( A ) is a random variable. X i : Indicator random variable, which is set to 1 if pivot is of 3 rank i in A , else zero. Let A i left and A i right be the corresponding left and right subarrays. n � � � Q ( A i left ) + Q ( A i Q ( A ) = n + X i · right ) . i =1 Since each element of A has probability exactly of 1 / n of being chosen: E[ X i ] = Pr[ pivot has rank i ] = 1 / n . Ruta (UIUC) CS473 40 Spring 2018 40 / 53

  67. Independence of Random Variables Lemma Random variables X i is independent of random variables Q ( A i left ) as well as Q ( A i right ) , i.e. � X i · Q ( A i � � Q ( A i � left ) = E[ X i ] E left ) E � � � � X i · Q ( A i Q ( A i right ) = E[ X i ] E right ) E Proof. This is because the algorithm, while recursing on Q ( A i left ) and Q ( A i right ) uses new random coin tosses that are independent of the coin tosses used to decide the first pivot. Only the latter decides value of X i . Ruta (UIUC) CS473 41 Spring 2018 41 / 53

  68. Analysis via Recurrence Let T ( n ) = max A : | A | = n E[ Q ( A )] be the worst-case expected running time of randomized QuickSort on arrays of size n . Ruta (UIUC) CS473 42 Spring 2018 42 / 53

  69. Analysis via Recurrence Let T ( n ) = max A : | A | = n E[ Q ( A )] be the worst-case expected running time of randomized QuickSort on arrays of size n . We have, for any A : n � � � Q ( A i left ) + Q ( A i Q ( A ) = n + X i right ) i =1 Ruta (UIUC) CS473 42 Spring 2018 42 / 53

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend