a zoo of discrete
play

a zoo of (discrete) Probability: Mean, Variance: random variables - PowerPoint PPT Presentation

discrete uniform random variables A discrete random variable X equally likely to take any (integer) value between integers a and b , inclusive, is uniform. Notation: X ~ Unif (a,b) a zoo of (discrete) Probability: Mean, Variance:


  1. discrete uniform random variables A discrete random variable X equally likely to take any (integer) value between integers a and b , inclusive, is uniform. Notation: X ~ Unif (a,b) a zoo of (discrete) Probability: Mean, Variance: random variables 1 2 discrete uniform random variables Bernoulli random variables A discrete random variable X equally likely to take any An experiment results in “Success” or “Failure” (integer) value between integers a and b , inclusive, is uniform. X is an indicator random variable (1 = success, 0 = failure) P(X=1) = p and P(X=0) = 1-p Notation: X ~ Unif (a,b) X is called a Bernoulli random variable: X ~ Ber(p) E[X] = E[X 2 ] = p Probability: Var(X) = E[X 2 ] – (E[X]) 2 = p – p 2 = p(1-p) Mean, Variance: 0.22 Example: value shown on one 
 P(X=i) roll of a fair die is Unif(1,6): 0.16 P( X=i ) = 1/6 
 0.10 E[ X ] = 7/2 
 Var[ X ] = 35/12 0 1 2 3 4 5 6 7 3 4 i

  2. Bernoulli random variables binomial random variables Consider n independent random variables Y i ~ Ber(p) An experiment results in “Success” or “Failure” X = Σ i Y i is the number of successes in n trials X is an indicator random variable (1 = success, 0 = failure) X is a Binomial random variable: X ~ Bin(n,p) P(X=1) = p and P(X=0) = 1-p X is called a Bernoulli random variable: X ~ Ber(p) Examples: coin flip Examples random binary digit # of heads in n coin flips whether a disk drive crashed # of 1’s in a randomly generated length n bit string # of disk drive crashes in a 1000 computer cluster 5 6 binomial pmfs binomial pmfs PMF for X ~ Bin(10,0.5) PMF for X ~ Bin(10,0.25) PMF for X ~ Bin(30,0.5) PMF for X ~ Bin(30,0.1) 0.30 0.30 0.25 0.25 0.25 0.25 0.20 0.20 0.20 0.20 0.15 0.15 µ ± σ P(X=k) P(X=k) P(X=k) P(X=k) µ ± σ 0.15 0.15 µ ± σ 0.10 0.10 µ ± σ 0.10 0.10 0.05 0.05 0.05 0.05 0.00 0.00 0.00 0.00 0 2 4 6 8 10 0 2 4 6 8 10 0 5 10 15 20 25 30 0 5 10 15 20 25 30 k k k k 7 8

  3. mean, variance of the binomial (II) geometric distribution In a series X 1 , X 2 , ... of Bernoulli trials with success probability p, let Y be the index of the first success, i.e., X 1 = X 2 = ... = X Y-1 = 0 & X Y = 1 Then Y is a geometric random variable with parameter p. Examples: Number of coin flips until first head Number of blind guesses on SAT until I get one right Number of darts thrown until you hit a bullseye Number of random probes into hash table until empty slot Number of wild guesses at a password until you hit it 9 10 geometric distribution geometric distribution In a series X 1 , X 2 , ... of Bernoulli trials with success probability Flip a (biased) coin repeatedly until 1 st head observed p, let Y be the index of the first success, i.e., How many flips? Let X be that number. X 1 = X 2 = ... = X Y-1 = 0 & X Y = 1 P(X=1) = P(H) = p Then Y is a geometric random variable with parameter p. P(X=2) = P(TH) = (1-p)p P(X=3) = P(TTH) = (1-p) 2 p P(Y=k) = (1-p) k-1 p; Mean 1/p; Variance (1-p)/p 2 ... memorize me! Check that it is a valid probability distribution: 1) 2) 11 12

  4. geometric random variable Geo(p) models & reality Let X be the number of flips up to & including 1 st head Sending a bit string over the network observed in repeated flips of a biased coin. n = 4 bits sent, each corrupted with probability 0.1 X = # of corrupted bits, X ~ Bin(4, 0.1) In real networks, large bit strings (length n ≈ 10 4 ) P ( H ) = p ; P ( T ) = 1 − p = q Corruption probability is very small: p ≈ 10 -6 pq i − 1 p ( i ) = ← PMF i ≥ 1 ipq i − 1 = p P X ~ Bin(10 4 , 10 -6 ) is unwieldy to compute i ≥ 1 iq i − 1 E [ X ] = P i ≥ 1 ip ( i ) = P ( ∗ ) Extreme n and p values arise in many cases A calculus trick: # bit errors in file written to disk 
 # of typos in a book # of elements in particular bucket of large hash table 
 dy 0 /dy = 0 So ( * ) becomes: (1 − q ) 2 = p p p 2 = 1 # of server crashes per day in giant data center iq i − 1 = X E [ X ] = p p # facebook login requests sent to a particular server i ≥ 1 E.g.: p =1/2; on average head every 2 nd flip p =1/10; on average, head every 10 th flip. 13 14 Poisson random variables poisson random variables Suppose “events” happen, independently, at an average rate of λ per unit time. Let X be 
 the actual number of events happening in a given time unit. Then X is a Poisson r.v. with 0.6 λ = 0.5 parameter λ (denoted X ~ Poi( λ )) and has λ = 3 distribution (PMF): 0.5 Siméon Poisson, 1781-1840 0.4 P(X=i) Examples: 0.3 # of alpha particles emitted by a lump of radium in 1 sec. 0.2 # of traffic accidents in Seattle in one year 0.1 # of babies born in a day at UW Med center # of visitors to my web page today 0.0 See B&T Section 6.2 for more on theoretical basis for Poisson. 0 1 2 3 4 5 6 15 16 i

  5. poisson random variables poisson random variables X is a Poisson r.v. with parameter λ if it has PMF: X is a Poisson r.v. with parameter λ if it has PMF: Is it a valid distribution? Recall Taylor series: Is it a valid distribution? Recall Taylor series: So So 17 18 expected value of poisson r.v.s binomial —> Poisson in the limit X ~ Binomial(n,p) Poisson approximates binomial when n is large, p is small, i = 0 term is zero and λ = np is “moderate” j = i-1 As expected, given definition in terms of “average rate λ ” (Var[X] = λ , too; proof similar, see B&T example 6.20) 19 20

  6. binomial → poisson in the limit binomial random variable is poisson in the limit X ~ Binomial(n,p) Poisson approximates binomial when n is large, p is small, and λ = np is “moderate” Different interpretations of “moderate,” e.g. n > 20 and p < 0.05 n > 100 and p < 0.1 Formally, Binomial is Poisson in the limit as 
 n → ∞ (equivalently, p → 0) while holding np = λ I.e., Binomial ≈ Poisson for large n, small p, moderate i, λ . Handy: Poisson has only 1 parameter–the expected # of successes 21 22 sending data on a network binomial vs poisson Consider sending bit string over a network Binomial(10, 0.3) Send bit string of length n = 10 4 Binomial(100, 0.03) 0.20 Probability of (independent) bit corruption is p = 10 -6 Poisson(3) X ~ Poi( λ = 10 4 •10 -6 = 0.01) P(X=k) What is probability that message arrives uncorrupted? 0.10 Using Y ~ Bin(10 4 , 10 -6 ): P(Y=0) ≈ 0.990049829 0.00 I.e., Poisson approximation (here) is accurate to ~5 parts per billion 0 2 4 6 8 10 k 23 24

  7. expectation and variance of a poisson expectation and variance of a poisson Recall: if Y ~ Bin(n,p), then: Recall: if Y ~ Bin(n,p), then: E[Y] = pn E[Y] = pn Var[Y] = np(1-p) Var[Y] = np(1-p) And if X ~ Poi( λ ) where λ = np (n →∞ , p → 0) then And if X ~ Poi( λ ) where λ = np (n →∞ , p → 0) then E[X] = λ = np = E[Y] E[X] = λ = np = E[Y] Var[X] = λ ≈ λ (1- λ /n) = np(1-p) = Var[X] = λ ≈ λ (1- λ /n) = np(1-p) = Var[Y] Var[Y] Expectation and variance of Poisson are the same ( λ ) Expectation is the same as corresponding binomial Variance almost the same as corresponding binomial Note: when two different distributions share the same 
 mean & variance, it suggests (but doesn’t prove) that 
 one may be a good approximation for the other. 25 26 some important (discrete) distributions balls in urns – the hypergeometric distribution Draw n balls (without replacement) from an urn containing E(X) E(X 2 ) E [ k 2 ] σ 2 PMF E [ k ] Name N , of which m are white, the rest black. ( b − a +1) 2 − 1 n 1 a + b Uniform( a, b ) f ( k ) = ( b − a +1) , k = a, a + 1 , . . . , b 2 12 Let X = number of white balls drawn ( 1 − p if k = 0 N Bernoulli( p ) f ( k ) = p p p (1 − p ) if k = 1 � m �� N − m p � i n − i P ( X = i ) = � N � n � p k (1 − p ) n − k , k = 0 , 1 , . . . , n � Binomial( p, n ) f ( k ) = np np (1 − p ) k n f ( k ) = e − λ λ k Poisson( λ ) k ! , k = 0 , 1 , . . . λ λ ( λ + 1) λ f ( k ) = p (1 − p ) k − 1 , k = 1 , 2 , . . . 1 2 − p 1 − p Geometric( p ) E[X] = np, where p = m/N (the fraction of white balls) p p 2 p 2 proof: Let X j be 0/1 indicator for j-th ball is white, X = Σ X j ⇣ ( n − 1)( m − 1) f ( k ) = ( m k )( N − m Hypergeomet- n − k ) ⌘ nm nm + 1 − nm , k = 0 , 1 , . . . , N ric( n, N, m ) ( N n ) N N N − 1 N The X j are dependent , but E[X] = E[ Σ X j ] = Σ E[X j ] = np Var[X] = np(1-p)(1-(n-1)/(N-1)) like binomial (almost) 27 28

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend