Randomized Algorithms Lecture 5: Two-point Sampling Sotiris - - PowerPoint PPT Presentation

randomized algorithms lecture 5 two point sampling
SMART_READER_LITE
LIVE PREVIEW

Randomized Algorithms Lecture 5: Two-point Sampling Sotiris - - PowerPoint PPT Presentation

Randomized Algorithms Lecture 5: Two-point Sampling Sotiris Nikoletseas Professor CEID - ETY Course 2017 - 2018 Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 1 / 26 Overview A. Pairwise independence of random


slide-1
SLIDE 1

Randomized Algorithms Lecture 5: “Two-point Sampling”

Sotiris Nikoletseas Professor

CEID - ETY Course 2017 - 2018

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 1 / 26

slide-2
SLIDE 2

Overview

  • A. Pairwise independence of random variables
  • B. The pairwise independent sampling theorem
  • C. Probability amplification via reduced randomness

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 2 / 26

slide-3
SLIDE 3
  • A. On the Additivity of Variance

In general the variance of a sum of random variables is not equal to the sum of their variances However, variances do add for independent variables (i.e. mutually independent variables) In fact, mutual independence is not necessary and pairwise independence suffices This is very useful, since in many situations the random variables involved are pairwise independent but not mutually independent.

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 3 / 26

slide-4
SLIDE 4

Conditional distributions

Let X, Y be discrete random variables. Their joint probability density function is f(x, y) = Pr{(X = x) ∩ (Y = y)} Clearly f1(x) = Pr{X = x} =

  • y

f(x, y) and f2(y) = Pr{Y = y} =

  • x

f(x, y) Also, the conditional probability density function is: f(x|y) = Pr{X = x|Y = y} = Pr{(X = x) ∩ (Y = y)} Pr{Y = y} = = f(x, y) f2(y) = f(x, y)

  • x f(x, y)

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 4 / 26

slide-5
SLIDE 5

Pairwise independence

Let random variables X1, X2, . . . , Xn. These are called pairwise independent iff for all i = j it is Pr{(Xi = x)|(Xj = y)} = Pr{Xi = x}, ∀x, y Equivalently, Pr{(Xi = x) ∩ (Xj = y)} = = Pr{Xi = x} · Pr{Xj = y}, ∀x, y Generalizing, the collection is k-wise independent iff, for every subset I ⊆ {1, 2, . . . , n} with |I| < k for every set of values {ai}, b and j / ∈ I, it is Pr

  • Xj = b|
  • i∈I

Xi = ai

  • = Pr{Xj = b}

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 5 / 26

slide-6
SLIDE 6

Mutual (or“full”) independence

The random variables X1, X2, . . . , Xn are mutually independent iff for any subset Xi1, Xi2, . . . , Xik, (2 ≤ k ≤ n) of them, it is Pr{(Xi1 = x1) ∩ (Xi2 = x2) ∩ · · · ∩ (Xik = xk)} = = Pr{Xi1 = x1} · Pr{Xi2 = x2} · · · Pr{Xik = xk} Example (for n = 3). Let A1, A2, A3 3 events. They are mutually independent iff all four equalities hold: Pr{A1A2} = Pr{A1} Pr{A2} (1) Pr{A2A3} = Pr{A2} Pr{A3} (2) Pr{A1A3} = Pr{A1} Pr{A3} (3) Pr{A1A2A3} = Pr{A1} Pr{A2} Pr{A3} (4) They are called pairwise independent if (1), (2), (3) hold.

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 6 / 26

slide-7
SLIDE 7

Mutual vs pairwise independence

Important notice: Pairwise independence does not imply mutual independence in general.

  • Example. Let a probability space including all

permutations of a, b, c as well as aaa, bbb, ccc (all 9 points considered equiprobable). Let Ak =“at place k there is an a” (for k = 1, 2, 3). It is Pr{A1} = Pr{A2} = Pr{A3} = 2+1

9

= 1

3

Also Pr{A1A2} = Pr{A2A3} = Pr{A1A3} = 1

9 = 1 3 · 1 3

thus A1, A2, A3 are pairwise independent. But Pr{A1A2A3} = 1

9 = Pr{A1} Pr{A2} Pr{A3} = 1 27

thus the events are not mutually independent

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 7 / 26

slide-8
SLIDE 8

Variance: key features

Definition: V ar(X) = E[(X − µ)2] =

  • x

(x − µ)2 Pr{X = x} where µ = E[X] =

  • x

x Pr{X = x} We call standard deviation of X the σ =

  • V ar(X)

Basic Properties:

(i) V ar(X) = E[X2] − E2[X] (ii) V ar(cX) = c2V ar(X), where c constant. (iii) V ar(X + c) = V ar(X), where c constant.

proof of (i): V ar(X) = E[(X − µ)2] = E[X2 − 2µX + µ2] = E[X2] + E[−2µX] + E[µ2] = E[X2] − 2µE[X] + µ2 = E[X2] − µ2

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 8 / 26

slide-9
SLIDE 9

The additivity of variance

Theorem: if X1, X2, . . . , Xn are pairwise independent random variables, then: V ar n

  • i=1

Xi

  • =

n

  • i=1

V ar(Xi) Proof: V ar(X1 +· · ·+Xn) = E[(X1 +· · ·+Xn)2]−E2[X1 +· · ·+Xn] = = E  

n

  • i=1

X2

i + n

  • 1≤i=j≤n

XiXj   −  

n

  • i=1

µ2

i + n

  • 1≤i=j≤n

µiµj   = =

n

  • i=1

(E[X2

i ] − µ2 i ) + n

  • 1≤i=j≤n

(E[XiXj] − µiµj) =

n

  • i=1

V ar(Xi) (since Xi pairwise independent, so ∀1 ≤ i = j ≤ n E(XiXj) = E(Xi)E(Xj = µiµj)

  • Note: As we see in the proof, the pairwise independence

suffices, and mutual (full) independence is not needed.

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 9 / 26

slide-10
SLIDE 10
  • B. The pairwise independent sampling theorem

Another Example. Birthday matching: Let us try to estimate the number of pairs of people in a room having birthday on the same day. Note 1: Matching birthdays for different pairs of students are pairwise independent, since knowing that (George, Takis) have a match tell us nothing about whether (George, Petros) match. Note 2: However, the events are not mutually independent. Indeed they are not even 3-wise independent since if (George,Takis) match and (Takis, Petros) match then (George, Petros) match!

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 10 / 26

slide-11
SLIDE 11

Birthday matching

Let us calculate the probability of having a certain number

  • f birthday matches.

Let B1, B2, . . . , Bn the birthdays of n independently chosen people and let Ei,j be the indicator variable for the event of a (i, j) match (i.e. Bi = Bj). As said, the events Ei,j are pairwise independent but not mutually independent. Clearly, Pr{Ei,j} = Pr{Bi = Bj} = 365 1

365 1 365 = 1 365 (for

i = j). Let D the number of matching pairs. Then D =

  • 1≤i<j≤n

Ei,j By linearity of expectation we have E[D] = E  

  • 1≤i<j≤n

Ei,j   =

  • 1≤i<j≤n

E[Ei,j] = n 2 1 365

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 11 / 26

slide-12
SLIDE 12

Birthday matching

Since the variances of pairwise independent variables Ei,j add up, it is: V ar[D] = V ar  

  • 1≤i<j≤n

Ei,j   =

  • 1≤i<j≤n

V ar[Ei,j] = = n

2

1

365(1 − 1 365)

As an example, for a class of n = 100 students, it is E[D] ≃ 14 and V ar[D] < 14(1 −

1 365) < 14. So by

Chebyshev’s inequality we have Pr{|D − 14| ≥ x} ≤ 14

x2

Letting x = 6, we conclude that with more than 50% chance the number of matching birthdays will be between 8 and 20.

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 12 / 26

slide-13
SLIDE 13

The Pairwise Independent Sampling Theorem (I)

We can actually generalize and not restrict to sums of zero-one (indicator) valued variables neither to variables with the same distribution. We below state the theorem for possibly different distributions with same mean and variance (but this is done for simplicity, and the result holds for distributions with different means and/or variances as well).

  • Theorem. Let X1, . . . , Xn pairwise independent variables

with the same mean µ and variance σ2. Let Sn = n

i=1 Xi

Then Pr

  • Sn

n − µ

  • ≥ x
  • ≤ 1

n

σ

x

2

  • Proof. Note that E[ Sn

n ] = nµ n = µ,

V ar[ Sn

n ] =

1

n

2 nσ2 = σ2

n and apply Chebyshev’s

inequality.

  • Sotiris Nikoletseas, Professor

Randomized Algorithms - Lecture 5 13 / 26

slide-14
SLIDE 14

The Pairwise Independent Sampling Theorem (II)

Note: This Theorem actually provides a precise general evaluation of how the average of pairwise independent random samples approaches their mean. If the number n of samples becomes large enough we can arbitrarily close approach the mean with confidence arbitrarily close to 100% (n > σ2

x2 ) i.e. a

large number of samples is needed for distributions of large variance and when we want to assure high concentration around the mean).

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 14 / 26

slide-15
SLIDE 15
  • C. Reduced randomness at probability amplification

Motivation: Randomized Algorithms, for a given input x, actually choose n random numbers (“witnesses”) and run a deterministic algorithm on the input, using each of these random numbers. intuitively, if the deterministic algorithm has a probability

  • f error ǫ (e.g. 1

2), t independent runs reduce the error

probability to ǫt(e.g.

1 2t ) and amplify the correctness

probability from 1

2 to 1 − 1 2t .

however, true randomness is quite expensive! What happens if we are constrained to use no more than a constant c random numbers? The simplest case is when c = 2 e.g. we choose just 2 random numbers (thus the name two-point sampling)

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 15 / 26

slide-16
SLIDE 16
  • C. Reduced randomness at probability amplification

Problem definition: If our randomized algorithm reduced the error probability to ǫt with t random numbers, what can we expect about the error probability with 2 random numbers only? an obvious bound is ǫ2(e.g. when ǫ = 1

2 reducing the error

probability from 1

2 to 1 4)

can we do any better? it turns out that we can indeed do much better and reduce the error probability to 1

t (which is much smaller than the

constant ǫ2)

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 16 / 26

slide-17
SLIDE 17
  • C. Reduced randomness at probability amplification

High level idea: generate t (“pseudo-random”) numbers based on the chosen 2 truly random numbers and use them in lieu of t truly independent random numbers. these generated numbers are dependent on the 2 chosen

  • numbers. Hence they are not independent, but are pairwise

independent. This loss of full independence reduces the accuracy of the algorithmic process but is still quite usable in reducing the error probability.

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 17 / 26

slide-18
SLIDE 18
  • C. Reduced randomness at probability amplification

The process (high level description):

1 Choose a large prime number p

(e.g. Mersenne prime, 231 − 1).

2 Define Zp as the ring of integers modulo p

(e.g 0, 1, 2, . . . , 231 − 2)

3 Choose 2 truly random numbers, a and b from Zp

(e.g. a = 220 + 781 and b = 227 − 44).

4 Generate t “pseudo-random” numbers yi = (ai + b) mod p

e.g. y0 = 227 − 44 y1 = (220 + 781) · 1 + 227 − 44 y2 = (220 + 781) · 2 + 227 − 44 and so on.

5 Use each of the yi’s as t witnesses (in lieu of t purely

independent random witnesses)

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 18 / 26

slide-19
SLIDE 19
  • C. Reduced randomness at probability amplification

Performance (high level discussion) As we will see: for any given error bound ǫ, the error probability is reduced from ǫ2 to 1

t .

however, instead of requiring O(log 1

ǫ) runs on the

deterministic algorithm (in the case of t independent random witnesses) we will require O( 1

ǫ) runs in the case of

2 independent random witnesses. thus, we gain in the probability amplification but loose on some efficiency. we need significantly less true randomness.

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 19 / 26

slide-20
SLIDE 20

The class RP (I)

  • Definition. The class RP (Random Polynomial time) consists of

all languages L admitting a randomized algorithm A running in worst case polynomial time such that for any input x: x ∈ L ⇒ Pr{A(x) accepts} ≥ 1

2

x / ∈ L ⇒ Pr{A(x) accepts} = 0 Notes: language recognition ← → computational decision problems the 1

2 value is arbitrary. The success probability needs just

to be lower-bounded by an inverse polynomial function of the input size (a polynomial number of algorithm repetitions would boost it to constant, in polynomial time). RP actually includes Monte Carlo algorithms than can err

  • nly when x ∈ L (one-sided error)

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 20 / 26

slide-21
SLIDE 21

The RP algorithm

An RP algorithm, for given input x, actually picks a random number r from Zp and computes A(x, r) with the following properties:

  • x ∈ L ⇒ A(x, r) = 1, for half of the possible values of r
  • x /

∈ L ⇒ A(x, r) = 0, for all possible choices of r

As said,

  • if we run algorithm A t times on the same input x, the error

probability is ≤

1 2t but this requires t truly random

numbers.

  • if we restrict ourselves to t = 2 true random numbers a, b

from Zp and run A(x, a) and A(x, b), the error probability can be as high as 1

4

  • but we can do better with t pseudo-random numbers: Let

ri = a · i + b (mod p), where a, b are truly random, as above, for i = 1, 2, . . . , t.

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 21 / 26

slide-22
SLIDE 22

Modulo rings pairwise independence (I)

Let p a prime number and Zp = {0, 1, 2, . . . , p − 1} denote the ring of the integers modulo p. Lemma 1. Given y, i ∈ Zp and choosing a, b randomly uniformly from Zp, the probability of y ≡ a · i + b (mod p) is 1

p

  • Proof. Imagine that we first choose a. Then when choosing b, it

must be y − a · i ≡ b (mod p). Since we choose b uniformly and modulo p can take p values, the probability is clearly 1

p indeed.

  • Sotiris Nikoletseas, Professor

Randomized Algorithms - Lecture 5 22 / 26

slide-23
SLIDE 23

Modulo rings pairwise independence (II)

Lemma 2. Given y, z, x, w ∈ Zp such that x = w, and choosing a, b randomly uniformly from Zp, the probability of y ≡ a · x + b (mod p) and z ≡ a · w + b (mod p) is

1 p2

  • Proof. It is y − z ≡ a · (x − w) (mod p). Since x − w = 0, the

equation holds for a unique value of a. This in turn implies a specific value for b. The probability that a, b get those two specific values is clearly 1

p · 1 p = 1 p2 .

  • Sotiris Nikoletseas, Professor

Randomized Algorithms - Lecture 5 23 / 26

slide-24
SLIDE 24

Modulo rings pairwise independence (III)

Lemma 3. Let i, j two distinct elements of Zp,and choose a, b randomly uniformly from Zp. Then the two variables Yi = a · i + b (mod p) and Yj = a · j + b (mod p) are uniformly distributed on Zp and are pairwise independent.

  • Proof. From lemma 1, it is clearly

Pr{Yi = α} = 1

p, for any α ∈ Zp, so Yi, Yj indeed have uniform

  • distribution. As for pairwise independence, note that:

Pr{Yi = α|Yj = β} = Pr{Yi=α∩Yj=β}

Pr{Yj=β}

=

1 p2 1 p = Pr{Yi = α}

Thus, Yi, Yj are pairwise independent.

  • Sotiris Nikoletseas, Professor

Randomized Algorithms - Lecture 5 24 / 26

slide-25
SLIDE 25

Modulo rings pairwise independence (IV)

Note: This is only pairwise independence. Indeed, consider the variables Y1, Y2 Y3, Y4 as defined above. Every pair of them are pairwise independent. But, if you give the value of Y1, Y2 then we know the values of Y3, Y4 immediately, since the values of Y1 and Y2 uniquely determine a and b, and thus we can compute all Yi variables.

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 25 / 26

slide-26
SLIDE 26

Modulo rings pairwise independence (V)

Thus Y = t

i=1 A(x, ri) is a sum of random variables

which are pairwise independent, since the ri values are pairwise independent. Assume that x ∈ L, then E[Y ] = t

2

and V ar[Y ] = t

i=1 V ar[A(x, ri)] = t 1 2 1 2 = t 4

The probability that all those t executions failed corresponds to the event Y = 0, and Pr{Y = 0} ≤ Pr{|Y − E[Y ]| ≥ E[Y ]} = Pr{|Y − t

2| ≥ t 2} ≤

≤ V ar[Y ] ( t

2) 2

=

t 4

( t

2) 2 = 1

t

from Chebyshev’s inequality. Thus the use of pseudo-randomness (t pseudo-random numbers) allows to “exploit” our 2 random bits to reduce the error probability from 1

4 to 1 t .

Sotiris Nikoletseas, Professor Randomized Algorithms - Lecture 5 26 / 26