cs70 jean walrand lecture 37
play

CS70: Jean Walrand: Lecture 37. Gaussian RVs and CLT 1. Review: - PowerPoint PPT Presentation

CS70: Jean Walrand: Lecture 37. Gaussian RVs and CLT 1. Review: Continuous Probability 2. Normal Distribution 3. Central Limit Theorem 4. Confidence Intervals 5. Bayes Rule with Continuous RVs Continuous Probability 1. pdf: Pr [ X ( x


  1. CS70: Jean Walrand: Lecture 37. Gaussian RVs and CLT 1. Review: Continuous Probability 2. Normal Distribution 3. Central Limit Theorem 4. Confidence Intervals 5. Bayes’ Rule with Continuous RVs

  2. Continuous Probability 1. pdf: Pr [ X ∈ ( x , x + δ ]] = f X ( x ) δ . � x 2. CDF: Pr [ X ≤ x ] = F X ( x ) = − ∞ f X ( y ) dy . 3. U [ a , b ] , Expo ( λ ) , target. � ∞ 4. Expectation: E [ X ] = − ∞ xf X ( x ) dx . � ∞ 5. Expectation of function: E [ h ( X )] = − ∞ h ( x ) f X ( x ) dx . 6. Variance: var [ X ] = E [( X − E [ X ]) 2 ] = E [ X 2 ] − E [ X ] 2 . 7. Variance of Sum of Independent RVs: If X n are pairwise independent, var [ X 1 + ··· + X n ] = var [ X 1 ]+ ··· + var [ X n ]

  3. Normal (Gaussian) Distribution. For any µ and σ , a normal (aka Gaussian ) random variable Y , which we write as Y = N ( µ , σ 2 ) , has pdf 1 2 πσ 2 e − ( y − µ ) 2 / 2 σ 2 . f Y ( y ) = √ Standard normal has µ = 0 and σ = 1 . Note: Pr [ | Y − µ | > 1 . 65 σ ] = 10 %; Pr [ | Y − µ | > 2 σ ] = 5 % .

  4. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . Now, √ σ f X ( y − µ 1 f Y ( y ) = ) σ 2 πσ 2 exp {− ( y − µ ) 2 1 = √ } . 2 σ 2

  5. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,.... 2 π exp {− x 2 1 Thus, f X ( x ) = 2 } . √ First note that E [ X ] = 0 , by symmetry. exp {− x 2 1 � E [ X 2 ] = x 2 √ var [ X ] = 2 } dx 2 π xd exp {− x 2 exp {− x 2 1 1 � � 2 } dx by IBP 1 = − √ 2 } = √ 2 π 2 π � = f X ( x ) dx = 1 . � b � b 1 Integration by Parts: a fdg = [ fg ] b a − a gdf .

  6. Review: Law of Large Numbers. Theorem: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expectation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 / n . Thus, = σ 2 Pr [ | A n − µ | > ε ] ≤ var [ A n ] n ε → 0 . ε 2

  7. Central Limit Theorem Central Limit Theorem Let X 1 , X 2 ,... be i.i.d. with E [ X 1 ] = µ and var ( X 1 ) = σ 2 . Define S n := A n − µ σ / √ n = X 1 + ··· + X n − n µ σ √ n . Then, S n → N ( 0 , 1 ) , as n → ∞ . That is, � α 1 − ∞ e − x 2 / 2 dx . √ Pr [ S n ≤ α ] → 2 π Proof: See EE126. Note: 1 σ / √ n ( E ( A n ) − µ ) = 0 E ( S n ) = 1 Var ( S n ) = σ 2 / nVar ( A n ) = 1 .

  8. CI for Mean Let X 1 , X 2 ,... be i.i.d. with mean µ and variance σ 2 . Let A n = X 1 + ··· + X n . n The CLT states that A n − µ σ / √ n = X 1 + ··· + X n − n µ σ √ n → N ( 0 , 1 ) as n → ∞ . Thus, for n ≫ 1, one has Pr [ − 2 ≤ | A n − µ σ / √ n | ≤ 2 ] ≈ 95 % . Equivalently, Pr [ µ ∈ [ A n − 2 σ √ n , A n + 2 σ √ n ]] ≈ 95 % . That is, [ A n − 2 σ √ n , A n + 2 σ √ n ] is a 95 % − CI for µ .

  9. CI for Mean Let X 1 , X 2 ,... be i.i.d. with mean µ and variance σ 2 . Let A n = X 1 + ··· + X n . n The CLT states that X 1 + ··· + X n − n µ σ √ n → N ( 0 , 1 ) as n → ∞ . Also, [ A n − 2 σ √ n , A n + 2 σ √ n ] is a 95 % − CI for µ . Recall: Using Chebyshev, we found that [ A n − 4 . 5 σ √ n , A n + 4 . 5 σ √ n ] is a 95 % − CI for µ . Thus, the CLT provides a smaller confidence interval.

  10. Coins and normal. Let X 1 , X 2 ,... be i.i.d. B ( p ) . Thus, X 1 + ··· + X n = B ( n , p ) . � Here, µ = p and σ = p ( 1 − p ) . CLT states that X 1 + ··· + X n − np → N ( 0 , 1 ) . � p ( 1 − p ) n

  11. Coins and normal. Let X 1 , X 2 ,... be i.i.d. B ( p ) . Thus, X 1 + ··· + X n = B ( n , p ) . � Here, µ = p and σ = p ( 1 − p ) . CLT states that X 1 + ··· + X n − np → N ( 0 , 1 ) � p ( 1 − p ) n and [ A n − 2 σ √ n , A n + 2 σ √ n ] is a 95 % − CI for µ with A n = ( X 1 + ··· + X n ) / n . Hence, [ A n − 2 σ √ n , A n + 2 σ √ n ] is a 95 % − CI for p . Since σ ≤ 0 . 5 , [ A n − 20 . 5 √ n , A n + 20 . 5 √ n ] is a 95 % − CI for p . Thus, [ A n − 1 √ n , A n + 1 √ n ] is a 95 % − CI for p .

  12. Application: Polling. How many people should one poll to estimate the fraction of votes that will go for Trump? Say we want to estimate that fraction within 3 % (margin of error), with 95 % confidence. This means that if the fraction is p , we want an estimate ˆ p such that Pr [ˆ p − 0 . 03 < p < ˆ p + 0 . 03 ] ≥ 95 % . p = X 1 + ··· + X n We choose ˆ where X m = 1 if person m says she will vote n for Trump, 0 otherwise. We assume X m are i.i.d. B ( p ) . p ± 1 Thus, ˆ √ n is a 95 % -confidence interval for p . We need 1 √ n = 0 . 03 , i.e., n = 1112 .

  13. Application: Testing Lightbulbs. Assume that lightbulbs have i.i.d. Expo ( λ ) lifetimes. We want to make sure that λ − 1 > 1. Say that we measure the average lifetime A n of n = 100 bulbs and we find that it is equal to 1 . 2. What is the confidence that we have that λ − 1 > 1? We have, A n − λ − 1 √ λ − 1 / √ n = n ( λ A n − 1 ) ≈ N ( 0 , 1 ) . Thus, √ √ √ Pr [ n ( λ A n − 1 ) > n ( λ 1 . 2 − 1 )] ≈ Pr [ N ( 0 , 1 ) > n ( λ 1 . 2 − 1 )] . If λ − 1 < 1, this probability is at most Pr [ N ( 0 , 1 ) > √ n ( 1 . 2 − 1 )] = Pr [ N ( 0 , 1 ) > 2 ] = 2 . 5 % . Thus, we conclude that Pr [ λ − 1 > 1 ] ≥ 97 . 5 % .

  14. Continuous RV and Bayes’ Rule Example 1: W.p. 1 / 2, X , Y are i.i.d. Expo ( 1 ) and w.p. 1 / 2, they are i.i.d. Expo ( 3 ) . Calculate E [ Y | X = x ] . Let B be the event that X ∈ [ x , x + δ ] where 0 < δ ≪ 1. Let A be the event that X , Y are Expo ( 1 ) . Then, ( 1 / 2 ) Pr [ B | A ] exp {− x } δ Pr [ A | B ] = A ] = ( 1 / 2 ) Pr [ B | A ]+( 1 / 2 ) Pr [ B | ¯ exp {− x } δ + 3exp {− 3 x } δ e 2 x exp {− x } = exp {− x } + 3exp {− 3 x } = 3 + e 2 x . Now, E [ Y | A ] Pr [ A | X = x ]+ E [ Y | ¯ A ] Pr [¯ E [ Y | X = x ] = A | X = x ] A | X = x ] ... = 1 + e 2 x 1 × Pr [ A | X = x ]+( 1 / 3 ) Pr [¯ = 3 + e 2 x . We used Pr [ Z ∈ [ x , x + δ ]] ≈ f Z ( x ) δ and given A one has f X ( x ) = exp {− x } whereas given ¯ A one has f X ( x ) = 3exp {− 3 x } .

  15. Continuous RV and Bayes’ Rule Example 2: W.p. 1 / 2, Bob is a good dart player and shoots uniformly in a circle with radius 1. Otherwise, Bob is a very good dart player and shoots uniformly in a circle with radius 1 / 2. The first dart of Bob is at distance 0 . 3 from the center of the target. (a) What is the probability that he is a very good dart player? (b) What is the expected distance of his second dart to the center of the target? Note: If uniform in radius r , then Pr [ X ≤ x ] = ( π x 2 ) / ( π r 2 ) , so that f X ( x ) = 2 x / ( r 2 ) . (a) We use Bayes’ Rule: Pr [ VG ] Pr [ ≈ 0 . 3 | VG ] Pr [ VG | 0 . 3 ] = Pr [ VG ] Pr [ ≈ 0 . 3 | VG ]+ Pr [ G ] Pr [ ≈ 0 . 3 | G ] 0 . 5 × 2 ( 0 . 3 2 ) ε / ( 0 . 5 2 ) = 0 . 5 × 2 ( 0 . 3 2 ) ε / ( 0 . 5 2 )+ 0 . 5 × 2 ε ( 0 . 3 2 ) = 0 . 8 . (b) E [ X ] = 0 . 8 × 0 . 5 × 2 3 + 0 . 2 × 2 3 = 0 . 4 .

  16. Summary Gaussian and CLT 1. Gaussian: N ( µ , σ 2 ) : f X ( x ) = ... “bell curve” ⇒ A n − µ 2. CLT: X n i.i.d. = σ / √ n → N ( 0 , 1 ) 3. CI: [ A n − 2 σ √ n , A n + 2 σ √ n ] = 95 % -CI for µ . 4. Bayes’ Rule: Replace { X = x } by { X ∈ ( x , x + ε ) } .

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend