eecs 70 lecture 27
play

EECS 70: Lecture 27. Joint and Conditional Distributions. EECS 70: - PowerPoint PPT Presentation

EECS 70: Lecture 27. Joint and Conditional Distributions. EECS 70: Lecture 27. Joint and Conditional Distributions. 1. Recap of variance of a random variable EECS 70: Lecture 27. Joint and Conditional Distributions. 1. Recap of variance of a


  1. Joint distribution. Two random variables, X and Y , in probability space: (Ω , P ) . What is ∑ x P [ X = x ] ? 1. What is ∑ y P [ Y = y ] ? 1. Let’s think about: P [ X = x , Y = y ] . What is ∑ x , y P [ X = x , Y = y ] ? Are the events “ X = x , Y = y ” disjoint?

  2. Joint distribution. Two random variables, X and Y , in probability space: (Ω , P ) . What is ∑ x P [ X = x ] ? 1. What is ∑ y P [ Y = y ] ? 1. Let’s think about: P [ X = x , Y = y ] . What is ∑ x , y P [ X = x , Y = y ] ? Are the events “ X = x , Y = y ” disjoint? Yes!

  3. Joint distribution. Two random variables, X and Y , in probability space: (Ω , P ) . What is ∑ x P [ X = x ] ? 1. What is ∑ y P [ Y = y ] ? 1. Let’s think about: P [ X = x , Y = y ] . What is ∑ x , y P [ X = x , Y = y ] ? Are the events “ X = x , Y = y ” disjoint? Yes! Y and X are functions on Ω .

  4. Joint distribution. Two random variables, X and Y , in probability space: (Ω , P ) . What is ∑ x P [ X = x ] ? 1. What is ∑ y P [ Y = y ] ? 1. Let’s think about: P [ X = x , Y = y ] . What is ∑ x , y P [ X = x , Y = y ] ? Are the events “ X = x , Y = y ” disjoint? Yes! Y and X are functions on Ω . Do they cover the entire sample space?

  5. Joint distribution. Two random variables, X and Y , in probability space: (Ω , P ) . What is ∑ x P [ X = x ] ? 1. What is ∑ y P [ Y = y ] ? 1. Let’s think about: P [ X = x , Y = y ] . What is ∑ x , y P [ X = x , Y = y ] ? Are the events “ X = x , Y = y ” disjoint? Yes! Y and X are functions on Ω . Do they cover the entire sample space? Yes!

  6. Joint distribution. Two random variables, X and Y , in probability space: (Ω , P ) . What is ∑ x P [ X = x ] ? 1. What is ∑ y P [ Y = y ] ? 1. Let’s think about: P [ X = x , Y = y ] . What is ∑ x , y P [ X = x , Y = y ] ? Are the events “ X = x , Y = y ” disjoint? Yes! Y and X are functions on Ω . Do they cover the entire sample space? Yes! X and Y are functions on Ω .

  7. Joint distribution. Two random variables, X and Y , in probability space: (Ω , P ) . What is ∑ x P [ X = x ] ? 1. What is ∑ y P [ Y = y ] ? 1. Let’s think about: P [ X = x , Y = y ] . What is ∑ x , y P [ X = x , Y = y ] ? Are the events “ X = x , Y = y ” disjoint? Yes! Y and X are functions on Ω . Do they cover the entire sample space? Yes! X and Y are functions on Ω . So,

  8. Joint distribution. Two random variables, X and Y , in probability space: (Ω , P ) . What is ∑ x P [ X = x ] ? 1. What is ∑ y P [ Y = y ] ? 1. Let’s think about: P [ X = x , Y = y ] . What is ∑ x , y P [ X = x , Y = y ] ? Are the events “ X = x , Y = y ” disjoint? Yes! Y and X are functions on Ω . Do they cover the entire sample space? Yes! X and Y are functions on Ω . So, ∑ x , y P [ X = x , Y = y ] = 1.

  9. Joint distribution. Two random variables, X and Y , in probability space: (Ω , P ) . What is ∑ x P [ X = x ] ? 1. What is ∑ y P [ Y = y ] ? 1. Let’s think about: P [ X = x , Y = y ] . What is ∑ x , y P [ X = x , Y = y ] ? Are the events “ X = x , Y = y ” disjoint? Yes! Y and X are functions on Ω . Do they cover the entire sample space? Yes! X and Y are functions on Ω . So, ∑ x , y P [ X = x , Y = y ] = 1. Joint Distribution: P [ X = x , Y = y ] .

  10. Joint distribution. Two random variables, X and Y , in probability space: (Ω , P ) . What is ∑ x P [ X = x ] ? 1. What is ∑ y P [ Y = y ] ? 1. Let’s think about: P [ X = x , Y = y ] . What is ∑ x , y P [ X = x , Y = y ] ? Are the events “ X = x , Y = y ” disjoint? Yes! Y and X are functions on Ω . Do they cover the entire sample space? Yes! X and Y are functions on Ω . So, ∑ x , y P [ X = x , Y = y ] = 1. Joint Distribution: P [ X = x , Y = y ] . Marginal Distributions: P [ X = x ] and P [ Y = y ] .

  11. Joint distribution. Two random variables, X and Y , in probability space: (Ω , P ) . What is ∑ x P [ X = x ] ? 1. What is ∑ y P [ Y = y ] ? 1. Let’s think about: P [ X = x , Y = y ] . What is ∑ x , y P [ X = x , Y = y ] ? Are the events “ X = x , Y = y ” disjoint? Yes! Y and X are functions on Ω . Do they cover the entire sample space? Yes! X and Y are functions on Ω . So, ∑ x , y P [ X = x , Y = y ] = 1. Joint Distribution: P [ X = x , Y = y ] . Marginal Distributions: P [ X = x ] and P [ Y = y ] . Important for inference.

  12. Two random variables, same outcome space. Experiment: pick a random person.

  13. Two random variables, same outcome space. Experiment: pick a random person. X = number of episodes of Games of Thrones they have seen.

  14. Two random variables, same outcome space. Experiment: pick a random person. X = number of episodes of Games of Thrones they have seen. Y = number of episodes of Westworld they have seen.

  15. Two random variables, same outcome space. Experiment: pick a random person. X = number of episodes of Games of Thrones they have seen. Y = number of episodes of Westworld they have seen. X 0 1 2 3 5 40 All P 0.3 0.05 0.05 0.05 0.05 0.1 0.4

  16. Two random variables, same outcome space. Experiment: pick a random person. X = number of episodes of Games of Thrones they have seen. Y = number of episodes of Westworld they have seen. X 0 1 2 3 5 40 All P 0.3 0.05 0.05 0.05 0.05 0.1 0.4 Is this a distribution?

  17. Two random variables, same outcome space. Experiment: pick a random person. X = number of episodes of Games of Thrones they have seen. Y = number of episodes of Westworld they have seen. X 0 1 2 3 5 40 All P 0.3 0.05 0.05 0.05 0.05 0.1 0.4 Is this a distribution? Yes! All the probabilities are non-negative and add up to 1.

  18. Two random variables, same outcome space. Experiment: pick a random person. X = number of episodes of Games of Thrones they have seen. Y = number of episodes of Westworld they have seen. X 0 1 2 3 5 40 All P 0.3 0.05 0.05 0.05 0.05 0.1 0.4 Is this a distribution? Yes! All the probabilities are non-negative and add up to 1. Y 0 1 5 10 P 0.3 0.1 0.1 0.5

  19. Joint distribution: Example. The joint distribution of X and Y is: Y/X 0 1 2 3 5 40 All 0 0.15 0 0 0 0 0.1 0.05 =0.3 1 0 0.05 0.05 0 0 0 0 =0.1 5 0 0 0 0.05 0.05 0 0 =0.1 10 0.15 0 0 0 0 0 0.35 =0.5 =0.3 =0.05 =0.05 =0.05 =0.05 =0.1 =0.4

  20. Joint distribution: Example. The joint distribution of X and Y is: Y/X 0 1 2 3 5 40 All 0 0.15 0 0 0 0 0.1 0.05 =0.3 1 0 0.05 0.05 0 0 0 0 =0.1 5 0 0 0 0.05 0.05 0 0 =0.1 10 0.15 0 0 0 0 0 0.35 =0.5 =0.3 =0.05 =0.05 =0.05 =0.05 =0.1 =0.4 Is this a valid distribution?

  21. Joint distribution: Example. The joint distribution of X and Y is: Y/X 0 1 2 3 5 40 All 0 0.15 0 0 0 0 0.1 0.05 =0.3 1 0 0.05 0.05 0 0 0 0 =0.1 5 0 0 0 0.05 0.05 0 0 =0.1 10 0.15 0 0 0 0 0 0.35 =0.5 =0.3 =0.05 =0.05 =0.05 =0.05 =0.1 =0.4 Is this a valid distribution? Yes!

  22. Joint distribution: Example. The joint distribution of X and Y is: Y/X 0 1 2 3 5 40 All 0 0.15 0 0 0 0 0.1 0.05 =0.3 1 0 0.05 0.05 0 0 0 0 =0.1 5 0 0 0 0.05 0.05 0 0 =0.1 10 0.15 0 0 0 0 0 0.35 =0.5 =0.3 =0.05 =0.05 =0.05 =0.05 =0.1 =0.4 Is this a valid distribution? Yes! Notice that P [ X = a ] and P [ Y = b ] are (marginal) distributions!

  23. Joint distribution: Example. The joint distribution of X and Y is: Y/X 0 1 2 3 5 40 All 0 0.15 0 0 0 0 0.1 0.05 =0.3 1 0 0.05 0.05 0 0 0 0 =0.1 5 0 0 0 0.05 0.05 0 0 =0.1 10 0.15 0 0 0 0 0 0.35 =0.5 =0.3 =0.05 =0.05 =0.05 =0.05 =0.1 =0.4 Is this a valid distribution? Yes! Notice that P [ X = a ] and P [ Y = b ] are (marginal) distributions! But now we have more information!

  24. Joint distribution: Example. The joint distribution of X and Y is: Y/X 0 1 2 3 5 40 All 0 0.15 0 0 0 0 0.1 0.05 =0.3 1 0 0.05 0.05 0 0 0 0 =0.1 5 0 0 0 0.05 0.05 0 0 =0.1 10 0.15 0 0 0 0 0 0.35 =0.5 =0.3 =0.05 =0.05 =0.05 =0.05 =0.1 =0.4 Is this a valid distribution? Yes! Notice that P [ X = a ] and P [ Y = b ] are (marginal) distributions! But now we have more information! For example, if I tell you someone watched 5 episodes of Westworld, they definitely didn’t watch all the episodes of GoT.

  25. Independent random variables. Definition: Independence

  26. Independent random variables. Definition: Independence The random variables X and Y are independent if and only if P [ Y = b | X = a ] = P [ Y = b ] , for all a and b .

  27. Independent random variables. Definition: Independence The random variables X and Y are independent if and only if P [ Y = b | X = a ] = P [ Y = b ] , for all a and b . Fact:

  28. Independent random variables. Definition: Independence The random variables X and Y are independent if and only if P [ Y = b | X = a ] = P [ Y = b ] , for all a and b . Fact: X , Y are independent if and only if P [ X = a , Y = b ] = P [ X = a ] P [ Y = b ] , for all a and b .

  29. Independent random variables. Definition: Independence The random variables X and Y are independent if and only if P [ Y = b | X = a ] = P [ Y = b ] , for all a and b . Fact: X , Y are independent if and only if P [ X = a , Y = b ] = P [ X = a ] P [ Y = b ] , for all a and b . Don’t need a huge table of probabilities like the previous slide.

  30. Independence: examples. Example 1 Roll two dices. X , Y = number of pips on the two dice. X , Y are independent.

  31. Independence: examples. Example 1 Roll two dices. X , Y = number of pips on the two dice. X , Y are independent. Indeed: P [ X = a , Y = b ] = 1 / 36 , P [ X = a ] = P [ Y = b ] = 1 / 6 .

  32. Independence: examples. Example 1 Roll two dices. X , Y = number of pips on the two dice. X , Y are independent. Indeed: P [ X = a , Y = b ] = 1 / 36 , P [ X = a ] = P [ Y = b ] = 1 / 6 . Example 2 Roll two dices. X = total number of pips, Y = number of pips on die 1 minus number on die 2. X and Y are

  33. Independence: examples. Example 1 Roll two dices. X , Y = number of pips on the two dice. X , Y are independent. Indeed: P [ X = a , Y = b ] = 1 / 36 , P [ X = a ] = P [ Y = b ] = 1 / 6 . Example 2 Roll two dices. X = total number of pips, Y = number of pips on die 1 minus number on die 2. X and Y are not independent.

  34. Independence: examples. Example 1 Roll two dices. X , Y = number of pips on the two dice. X , Y are independent. Indeed: P [ X = a , Y = b ] = 1 / 36 , P [ X = a ] = P [ Y = b ] = 1 / 6 . Example 2 Roll two dices. X = total number of pips, Y = number of pips on die 1 minus number on die 2. X and Y are not independent. Indeed: P [ X = 12 , Y = 1 ] = 0 � = P [ X = 12 ] P [ Y = 1 ] > 0.

  35. Independence: examples. Example 1 Roll two dices. X , Y = number of pips on the two dice. X , Y are independent. Indeed: P [ X = a , Y = b ] = 1 / 36 , P [ X = a ] = P [ Y = b ] = 1 / 6 . Example 2 Roll two dices. X = total number of pips, Y = number of pips on die 1 minus number on die 2. X and Y are not independent. Indeed: P [ X = 12 , Y = 1 ] = 0 � = P [ X = 12 ] P [ Y = 1 ] > 0.

  36. Mean of product of independent RVs. Theorem Let X , Y be independent RVs. Then E [ XY ] = E [ X ] E [ Y ] .

  37. Mean of product of independent RVs. Theorem Let X , Y be independent RVs. Then E [ XY ] = E [ X ] E [ Y ] . Proof: Recall that E [ g ( X , Y )] = ∑ x , y g ( x , y ) P [ X = x , Y = y ] .

  38. Mean of product of independent RVs. Theorem Let X , Y be independent RVs. Then E [ XY ] = E [ X ] E [ Y ] . Proof: Recall that E [ g ( X , Y )] = ∑ x , y g ( x , y ) P [ X = x , Y = y ] . Hence, = ∑ E [ XY ] xyP [ X = x , Y = y ] x , y

  39. Mean of product of independent RVs. Theorem Let X , Y be independent RVs. Then E [ XY ] = E [ X ] E [ Y ] . Proof: Recall that E [ g ( X , Y )] = ∑ x , y g ( x , y ) P [ X = x , Y = y ] . Hence, = ∑ xyP [ X = x , Y = y ] = ∑ E [ XY ] xyP [ X = x ] P [ Y = y ] x , y x , y

  40. Mean of product of independent RVs. Theorem Let X , Y be independent RVs. Then E [ XY ] = E [ X ] E [ Y ] . Proof: Recall that E [ g ( X , Y )] = ∑ x , y g ( x , y ) P [ X = x , Y = y ] . Hence, = ∑ xyP [ X = x , Y = y ] = ∑ E [ XY ] xyP [ X = x ] P [ Y = y ] , by ind. x , y x , y � � = ∑ ∑ xyP [ X = x ] P [ Y = y ] x y

  41. Mean of product of independent RVs. Theorem Let X , Y be independent RVs. Then E [ XY ] = E [ X ] E [ Y ] . Proof: Recall that E [ g ( X , Y )] = ∑ x , y g ( x , y ) P [ X = x , Y = y ] . Hence, = ∑ xyP [ X = x , Y = y ] = ∑ E [ XY ] xyP [ X = x ] P [ Y = y ] , by ind. x , y x , y � � = ∑ ∑ xyP [ X = x ] P [ Y = y ] x y = ∑ � � �� ∑ xP [ X = x ] yP [ Y = y ] x y

  42. Mean of product of independent RVs. Theorem Let X , Y be independent RVs. Then E [ XY ] = E [ X ] E [ Y ] . Proof: Recall that E [ g ( X , Y )] = ∑ x , y g ( x , y ) P [ X = x , Y = y ] . Hence, = ∑ xyP [ X = x , Y = y ] = ∑ E [ XY ] xyP [ X = x ] P [ Y = y ] , by ind. x , y x , y � � = ∑ ∑ xyP [ X = x ] P [ Y = y ] x y = ∑ � � �� ∑ xP [ X = x ] yP [ Y = y ] x y = ∑ xP [ X = x ] E [ Y ] x

  43. Mean of product of independent RVs. Theorem Let X , Y be independent RVs. Then E [ XY ] = E [ X ] E [ Y ] . Proof: Recall that E [ g ( X , Y )] = ∑ x , y g ( x , y ) P [ X = x , Y = y ] . Hence, = ∑ xyP [ X = x , Y = y ] = ∑ E [ XY ] xyP [ X = x ] P [ Y = y ] , by ind. x , y x , y � � = ∑ ∑ xyP [ X = x ] P [ Y = y ] x y = ∑ � � �� ∑ xP [ X = x ] yP [ Y = y ] x y = ∑ xP [ X = x ] E [ Y ] = E [ X ] E [ Y ] . x

  44. Variance of sum of two independent random variables

  45. Variance of sum of two independent random variables Theorem: If X and Y are independent, then Var ( X + Y ) = Var ( X )+ Var ( Y ) .

  46. Variance of sum of two independent random variables Theorem: If X and Y are independent, then Var ( X + Y ) = Var ( X )+ Var ( Y ) . Proof: Since shifting the random variables does not change their variance, let us subtract their means.

  47. Variance of sum of two independent random variables Theorem: If X and Y are independent, then Var ( X + Y ) = Var ( X )+ Var ( Y ) . Proof: Since shifting the random variables does not change their variance, let us subtract their means. That is, we assume that E ( X ) = 0 and E ( Y ) = 0.

  48. Variance of sum of two independent random variables Theorem: If X and Y are independent, then Var ( X + Y ) = Var ( X )+ Var ( Y ) . Proof: Since shifting the random variables does not change their variance, let us subtract their means. That is, we assume that E ( X ) = 0 and E ( Y ) = 0. Then, by independence, E ( XY ) = E ( X ) E ( Y ) = 0 .

  49. Variance of sum of two independent random variables Theorem: If X and Y are independent, then Var ( X + Y ) = Var ( X )+ Var ( Y ) . Proof: Since shifting the random variables does not change their variance, let us subtract their means. That is, we assume that E ( X ) = 0 and E ( Y ) = 0. Then, by independence, E ( XY ) = E ( X ) E ( Y ) = 0 . Hence, E (( X + Y ) 2 ) var ( X + Y ) =

  50. Variance of sum of two independent random variables Theorem: If X and Y are independent, then Var ( X + Y ) = Var ( X )+ Var ( Y ) . Proof: Since shifting the random variables does not change their variance, let us subtract their means. That is, we assume that E ( X ) = 0 and E ( Y ) = 0. Then, by independence, E ( XY ) = E ( X ) E ( Y ) = 0 . Hence, E (( X + Y ) 2 ) = E ( X 2 + 2 XY + Y 2 ) var ( X + Y ) =

  51. Variance of sum of two independent random variables Theorem: If X and Y are independent, then Var ( X + Y ) = Var ( X )+ Var ( Y ) . Proof: Since shifting the random variables does not change their variance, let us subtract their means. That is, we assume that E ( X ) = 0 and E ( Y ) = 0. Then, by independence, E ( XY ) = E ( X ) E ( Y ) = 0 . Hence, E (( X + Y ) 2 ) = E ( X 2 + 2 XY + Y 2 ) var ( X + Y ) = E ( X 2 )+ 2 E ( XY )+ E ( Y 2 ) =

  52. Variance of sum of two independent random variables Theorem: If X and Y are independent, then Var ( X + Y ) = Var ( X )+ Var ( Y ) . Proof: Since shifting the random variables does not change their variance, let us subtract their means. That is, we assume that E ( X ) = 0 and E ( Y ) = 0. Then, by independence, E ( XY ) = E ( X ) E ( Y ) = 0 . Hence, E (( X + Y ) 2 ) = E ( X 2 + 2 XY + Y 2 ) var ( X + Y ) = E ( X 2 )+ 2 E ( XY )+ E ( Y 2 ) = E ( X 2 )+ E ( Y 2 ) =

  53. Variance of sum of two independent random variables Theorem: If X and Y are independent, then Var ( X + Y ) = Var ( X )+ Var ( Y ) . Proof: Since shifting the random variables does not change their variance, let us subtract their means. That is, we assume that E ( X ) = 0 and E ( Y ) = 0. Then, by independence, E ( XY ) = E ( X ) E ( Y ) = 0 . Hence, E (( X + Y ) 2 ) = E ( X 2 + 2 XY + Y 2 ) var ( X + Y ) = E ( X 2 )+ 2 E ( XY )+ E ( Y 2 ) = E ( X 2 )+ E ( Y 2 ) = = var ( X )+ var ( Y ) .

  54. Examples. (1) Assume that X , Y , Z are (pairwise) independent, with E [ X ] = E [ Y ] = E [ Z ] = 0 and E [ X 2 ] = E [ Y 2 ] = E [ Z 2 ] = 1.

  55. Examples. (1) Assume that X , Y , Z are (pairwise) independent, with E [ X ] = E [ Y ] = E [ Z ] = 0 and E [ X 2 ] = E [ Y 2 ] = E [ Z 2 ] = 1. Then E [( X + 2 Y + 3 Z ) 2 ] = E [ X 2 + 4 Y 2 + 9 Z 2 + 4 XY + 12 YZ + 6 XZ ]

  56. Examples. (1) Assume that X , Y , Z are (pairwise) independent, with E [ X ] = E [ Y ] = E [ Z ] = 0 and E [ X 2 ] = E [ Y 2 ] = E [ Z 2 ] = 1. Then E [( X + 2 Y + 3 Z ) 2 ] = E [ X 2 + 4 Y 2 + 9 Z 2 + 4 XY + 12 YZ + 6 XZ ] = 1 + 4 + 9 + 4 × 0 + 12 × 0 + 6 × 0

  57. Examples. (1) Assume that X , Y , Z are (pairwise) independent, with E [ X ] = E [ Y ] = E [ Z ] = 0 and E [ X 2 ] = E [ Y 2 ] = E [ Z 2 ] = 1. Then E [( X + 2 Y + 3 Z ) 2 ] = E [ X 2 + 4 Y 2 + 9 Z 2 + 4 XY + 12 YZ + 6 XZ ] = 1 + 4 + 9 + 4 × 0 + 12 × 0 + 6 × 0 = 14 .

  58. Examples. (1) Assume that X , Y , Z are (pairwise) independent, with E [ X ] = E [ Y ] = E [ Z ] = 0 and E [ X 2 ] = E [ Y 2 ] = E [ Z 2 ] = 1. Then E [( X + 2 Y + 3 Z ) 2 ] = E [ X 2 + 4 Y 2 + 9 Z 2 + 4 XY + 12 YZ + 6 XZ ] = 1 + 4 + 9 + 4 × 0 + 12 × 0 + 6 × 0 = 14 . (2) Let X , Y be independent and U { 1 , 2 ,..., n } . Then

  59. Examples. (1) Assume that X , Y , Z are (pairwise) independent, with E [ X ] = E [ Y ] = E [ Z ] = 0 and E [ X 2 ] = E [ Y 2 ] = E [ Z 2 ] = 1. Then E [( X + 2 Y + 3 Z ) 2 ] = E [ X 2 + 4 Y 2 + 9 Z 2 + 4 XY + 12 YZ + 6 XZ ] = 1 + 4 + 9 + 4 × 0 + 12 × 0 + 6 × 0 = 14 . (2) Let X , Y be independent and U { 1 , 2 ,..., n } . Then E [ X 2 + Y 2 − 2 XY ] = 2 E [ X 2 ] − 2 E [ X ] 2 E [( X − Y ) 2 ] =

  60. Examples. (1) Assume that X , Y , Z are (pairwise) independent, with E [ X ] = E [ Y ] = E [ Z ] = 0 and E [ X 2 ] = E [ Y 2 ] = E [ Z 2 ] = 1. Then E [( X + 2 Y + 3 Z ) 2 ] = E [ X 2 + 4 Y 2 + 9 Z 2 + 4 XY + 12 YZ + 6 XZ ] = 1 + 4 + 9 + 4 × 0 + 12 × 0 + 6 × 0 = 14 . (2) Let X , Y be independent and U { 1 , 2 ,..., n } . Then E [ X 2 + Y 2 − 2 XY ] = 2 E [ X 2 ] − 2 E [ X ] 2 E [( X − Y ) 2 ] = 1 + 3 n + 2 n 2 − ( n + 1 ) 2 = . 3 2

  61. Variance: binomial. n � n � E [ X 2 ] i 2 p i ( 1 − p ) n − i . ∑ = i i = 0

  62. Variance: binomial. n � n � E [ X 2 ] i 2 p i ( 1 − p ) n − i . ∑ = i i = 0 =

  63. Variance: binomial. n � n � E [ X 2 ] i 2 p i ( 1 − p ) n − i . ∑ = i i = 0 = Really???!!##... Too hard!

  64. Variance: binomial. n � n � E [ X 2 ] i 2 p i ( 1 − p ) n − i . ∑ = i i = 0 = Really???!!##... Too hard! Ok..

  65. Variance: binomial. n � n � E [ X 2 ] i 2 p i ( 1 − p ) n − i . ∑ = i i = 0 = Really???!!##... Too hard! Ok.. fine.

  66. Variance: binomial. n � n � E [ X 2 ] i 2 p i ( 1 − p ) n − i . ∑ = i i = 0 = Really???!!##... Too hard! Ok.. fine. Let’s do something else.

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend