cs70 jean walrand lecture 36
play

CS70: Jean Walrand: Lecture 36. Gaussian and CLT CS70: Jean - PowerPoint PPT Presentation

CS70: Jean Walrand: Lecture 36. Gaussian and CLT CS70: Jean Walrand: Lecture 36. Gaussian and CLT Warning: CS70: Jean Walrand: Lecture 36. Gaussian and CLT Warning: This lecture is also rated R. CS70: Jean Walrand: Lecture 36. Gaussian and


  1. Expectation of function of RV Definition The expectation of a function of a random variable is defined as

  2. Expectation of function of RV Definition The expectation of a function of a random variable is defined as � ∞ E [ h ( X )] = − ∞ h ( x ) f X ( x ) dx .

  3. Expectation of function of RV Definition The expectation of a function of a random variable is defined as � ∞ E [ h ( X )] = − ∞ h ( x ) f X ( x ) dx . Justification: Say X = n δ w.p. f X ( n δ ) δ .

  4. Expectation of function of RV Definition The expectation of a function of a random variable is defined as � ∞ E [ h ( X )] = − ∞ h ( x ) f X ( x ) dx . Justification: Say X = n δ w.p. f X ( n δ ) δ . Then, E [ h ( X )] = ∑ h ( n δ ) Pr [ X = n δ ] n

  5. Expectation of function of RV Definition The expectation of a function of a random variable is defined as � ∞ E [ h ( X )] = − ∞ h ( x ) f X ( x ) dx . Justification: Say X = n δ w.p. f X ( n δ ) δ . Then, E [ h ( X )] = ∑ h ( n δ ) Pr [ X = n δ ] = ∑ h ( n δ ) f X ( n δ ) δ n n

  6. Expectation of function of RV Definition The expectation of a function of a random variable is defined as � ∞ E [ h ( X )] = − ∞ h ( x ) f X ( x ) dx . Justification: Say X = n δ w.p. f X ( n δ ) δ . Then, � ∞ E [ h ( X )] = ∑ h ( n δ ) Pr [ X = n δ ] = ∑ h ( n δ ) f X ( n δ ) δ = − ∞ h ( x ) f X ( x ) dx . n n � g ( x ) dx ≈ ∑ n g ( n δ ) δ with g ( x ) = h ( x ) f X ( x ) . Indeed,

  7. Expectation of function of RV Definition The expectation of a function of a random variable is defined as � ∞ E [ h ( X )] = − ∞ h ( x ) f X ( x ) dx . Justification: Say X = n δ w.p. f X ( n δ ) δ . Then, � ∞ E [ h ( X )] = ∑ h ( n δ ) Pr [ X = n δ ] = ∑ h ( n δ ) f X ( n δ ) δ = − ∞ h ( x ) f X ( x ) dx . n n � g ( x ) dx ≈ ∑ n g ( n δ ) δ with g ( x ) = h ( x ) f X ( x ) . Indeed, Fact Expectation is linear.

  8. Variance Definition: The variance of a continuous random variable X is defined as

  9. Variance Definition: The variance of a continuous random variable X is defined as E (( X − E ( X )) 2 ) var [ X ] =

  10. Variance Definition: The variance of a continuous random variable X is defined as E (( X − E ( X )) 2 ) = E ( X 2 ) − ( E ( X )) 2 var [ X ] =

  11. Variance Definition: The variance of a continuous random variable X is defined as E (( X − E ( X )) 2 ) = E ( X 2 ) − ( E ( X )) 2 var [ X ] = � ∞ � � ∞ � 2 − ∞ x 2 f ( x ) dx − = − ∞ xf ( x ) dx .

  12. Motivation for Gaussian Distribution Key fact:

  13. Motivation for Gaussian Distribution Key fact: The sum of many small independent RVs has a Gaussian distribution.

  14. Motivation for Gaussian Distribution Key fact: The sum of many small independent RVs has a Gaussian distribution. This is the Central Limit Theorem.

  15. Motivation for Gaussian Distribution Key fact: The sum of many small independent RVs has a Gaussian distribution. This is the Central Limit Theorem. (See later.)

  16. Motivation for Gaussian Distribution Key fact: The sum of many small independent RVs has a Gaussian distribution. This is the Central Limit Theorem. (See later.) Examples: Binomial and Poisson suitably scaled.

  17. Motivation for Gaussian Distribution Key fact: The sum of many small independent RVs has a Gaussian distribution. This is the Central Limit Theorem. (See later.) Examples: Binomial and Poisson suitably scaled. This explains why the Gaussian distribution

  18. Motivation for Gaussian Distribution Key fact: The sum of many small independent RVs has a Gaussian distribution. This is the Central Limit Theorem. (See later.) Examples: Binomial and Poisson suitably scaled. This explains why the Gaussian distribution (the bell curve)

  19. Motivation for Gaussian Distribution Key fact: The sum of many small independent RVs has a Gaussian distribution. This is the Central Limit Theorem. (See later.) Examples: Binomial and Poisson suitably scaled. This explains why the Gaussian distribution (the bell curve) shows up everywhere.

  20. Normal Distribution. For any µ and σ , a normal (aka Gaussian )

  21. Normal Distribution. For any µ and σ , a normal (aka Gaussian ) random variable Y , which we write as Y = N ( µ , σ 2 ) , has pdf

  22. Normal Distribution. For any µ and σ , a normal (aka Gaussian ) random variable Y , which we write as Y = N ( µ , σ 2 ) , has pdf 1 2 πσ 2 e − ( y − µ ) 2 / 2 σ 2 . f Y ( y ) = √

  23. Normal Distribution. For any µ and σ , a normal (aka Gaussian ) random variable Y , which we write as Y = N ( µ , σ 2 ) , has pdf 1 2 πσ 2 e − ( y − µ ) 2 / 2 σ 2 . f Y ( y ) = √ Standard normal has µ = 0 and σ = 1 .

  24. Normal Distribution. For any µ and σ , a normal (aka Gaussian ) random variable Y , which we write as Y = N ( µ , σ 2 ) , has pdf 1 2 πσ 2 e − ( y − µ ) 2 / 2 σ 2 . f Y ( y ) = √ Standard normal has µ = 0 and σ = 1 .

  25. Normal Distribution. For any µ and σ , a normal (aka Gaussian ) random variable Y , which we write as Y = N ( µ , σ 2 ) , has pdf 1 2 πσ 2 e − ( y − µ ) 2 / 2 σ 2 . f Y ( y ) = √ Standard normal has µ = 0 and σ = 1 . Note: Pr [ | Y − µ | > 1 . 65 σ ] = 10 %;

  26. Normal Distribution. For any µ and σ , a normal (aka Gaussian ) random variable Y , which we write as Y = N ( µ , σ 2 ) , has pdf 1 2 πσ 2 e − ( y − µ ) 2 / 2 σ 2 . f Y ( y ) = √ Standard normal has µ = 0 and σ = 1 . Note: Pr [ | Y − µ | > 1 . 65 σ ] = 10 %; Pr [ | Y − µ | > 2 σ ] = 5 % .

  27. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) .

  28. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . √

  29. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . Now, √

  30. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . Now, √ f Y ( y ) dy = Pr [ Y ∈ [ y , y + dy ]] =

  31. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . Now, √ f Y ( y ) dy = Pr [ Y ∈ [ y , y + dy ]] = Pr [ µ + σ X ∈ [ y , y + dy ]]

  32. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . Now, √ f Y ( y ) dy = Pr [ Y ∈ [ y , y + dy ]] = Pr [ µ + σ X ∈ [ y , y + dy ]] = Pr [ σ X ∈ [ y − µ , y − µ + dy ]]

  33. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . Now, √ f Y ( y ) dy = Pr [ Y ∈ [ y , y + dy ]] = Pr [ µ + σ X ∈ [ y , y + dy ]] = Pr [ σ X ∈ [ y − µ , y − µ + dy ]] Pr [ X ∈ [ y − µ , y − µ + dy = σ ]] σ σ

  34. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . Now, √ f Y ( y ) dy = Pr [ Y ∈ [ y , y + dy ]] = Pr [ µ + σ X ∈ [ y , y + dy ]] = Pr [ σ X ∈ [ y − µ , y − µ + dy ]] Pr [ X ∈ [ y − µ , y − µ + dy = σ ]] σ σ f X ( y − µ ) dy = σ σ

  35. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . Now, √ f Y ( y ) dy = Pr [ Y ∈ [ y , y + dy ]] = Pr [ µ + σ X ∈ [ y , y + dy ]] = Pr [ σ X ∈ [ y − µ , y − µ + dy ]] Pr [ X ∈ [ y − µ , y − µ + dy = σ ]] σ σ f X ( y − µ ) dy σ = 1 σ f X ( y − µ = ) dy σ σ

  36. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . Now, √ f Y ( y ) dy = Pr [ Y ∈ [ y , y + dy ]] = Pr [ µ + σ X ∈ [ y , y + dy ]] = Pr [ σ X ∈ [ y − µ , y − µ + dy ]] Pr [ X ∈ [ y − µ , y − µ + dy = σ ]] σ σ f X ( y − µ ) dy σ = 1 σ f X ( y − µ = ) dy σ σ 2 πσ 2 exp {− ( y − µ ) 2 1 √ = } dy . 2 σ 2

  37. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ

  38. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 .

  39. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,....

  40. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,.... 2 π exp {− x 2 1 Thus, f X ( x ) = 2 } . √

  41. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,.... 2 π exp {− x 2 1 Thus, f X ( x ) = 2 } . √ First note that E [ X ] = 0 , by symmetry.

  42. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,.... 2 π exp {− x 2 1 Thus, f X ( x ) = 2 } . √ First note that E [ X ] = 0 , by symmetry. E [ X 2 ] var [ X ] =

  43. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,.... 2 π exp {− x 2 1 Thus, f X ( x ) = 2 } . √ First note that E [ X ] = 0 , by symmetry. exp {− x 2 1 � E [ X 2 ] = x 2 var [ X ] = √ 2 } dx 2 π

  44. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,.... 2 π exp {− x 2 1 Thus, f X ( x ) = 2 } . √ First note that E [ X ] = 0 , by symmetry. exp {− x 2 1 � E [ X 2 ] = x 2 var [ X ] = √ 2 } dx 2 π xd exp {− x 2 1 � √ = − 2 } 2 π

  45. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,.... 2 π exp {− x 2 1 Thus, f X ( x ) = 2 } . √ First note that E [ X ] = 0 , by symmetry. exp {− x 2 1 � E [ X 2 ] = x 2 var [ X ] = √ 2 } dx 2 π xd exp {− x 2 exp {− x 2 1 1 � � √ √ = − 2 } = 2 } dx by IBP 2 π 2 π

  46. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,.... 2 π exp {− x 2 1 Thus, f X ( x ) = 2 } . √ First note that E [ X ] = 0 , by symmetry. exp {− x 2 1 � E [ X 2 ] = x 2 var [ X ] = √ 2 } dx 2 π xd exp {− x 2 exp {− x 2 1 1 � � √ √ = − 2 } = 2 } dx by IBP 2 π 2 π � = f X ( x ) dx

  47. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,.... 2 π exp {− x 2 1 Thus, f X ( x ) = 2 } . √ First note that E [ X ] = 0 , by symmetry. exp {− x 2 1 � E [ X 2 ] = x 2 var [ X ] = √ 2 } dx 2 π xd exp {− x 2 exp {− x 2 1 1 � � √ √ = − 2 } = 2 } dx by IBP 2 π 2 π � = f X ( x ) dx = 1 .

  48. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.”

  49. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 .

  50. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n .

  51. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n . n = A n − µ Let A ′ σ / √ n .

  52. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n . n = A n − µ Let A ′ σ / √ n . E ( A ′ n )

  53. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n . n = A n − µ Let A ′ σ / √ n . 1 E ( A ′ n ) = σ / √ n ( E ( A n ) − µ )

  54. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n . n = A n − µ Let A ′ σ / √ n . 1 E ( A ′ n ) = σ / √ n ( E ( A n ) − µ ) = 0 .

  55. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n . n = A n − µ Let A ′ σ / √ n . 1 E ( A ′ n ) = σ / √ n ( E ( A n ) − µ ) = 0 . Var ( A ′ n )

  56. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n . n = A n − µ Let A ′ σ / √ n . 1 E ( A ′ n ) = σ / √ n ( E ( A n ) − µ ) = 0 . Var ( A ′ 1 n ) = σ 2 / n Var ( A n )

  57. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n . n = A n − µ Let A ′ σ / √ n . 1 E ( A ′ n ) = σ / √ n ( E ( A n ) − µ ) = 0 . Var ( A ′ 1 n ) = σ 2 / n Var ( A n ) = 1 .

  58. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n . n = A n − µ Let A ′ σ / √ n . 1 E ( A ′ n ) = σ / √ n ( E ( A n ) − µ ) = 0 . Var ( A ′ 1 n ) = σ 2 / n Var ( A n ) = 1 . Central limit theorem: As n goes to infinity the distribution of A ′ n approaches the standard normal distribution. � α 1 ∞ e − x 2 / 2 dx . Pr [ A ′ √ n ≤ α ] → 2 π

  59. Coins and normal.. Let X 1 , X 2 ,... be i.i.d. B ( p ) .

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend