lecture 2 probability and distributions
play

Lecture 2: Probability and Distributions Ani Manichaikul - PowerPoint PPT Presentation

Lecture 2: Probability and Distributions Ani Manichaikul amanicha@jhsph.edu 17 April 2007 1 / 65 Probability: Why do we care? Probability helps us by: Allowing us to translate scientific questions info mathematical notation Providing a


  1. Lecture 2: Probability and Distributions Ani Manichaikul amanicha@jhsph.edu 17 April 2007 1 / 65

  2. Probability: Why do we care? Probability helps us by: Allowing us to translate scientific questions info mathematical notation Providing a framework for answering scientific questions Later, we will see how some common statistical methods in the scientific literature are actually probability concepts in disguise 2 / 65

  3. What is Probability? Probability is a measure of uncertainty about the occurrence of events Two definitions of probability Classical definition Relative frequency definition 3 / 65

  4. Classical Definition P ( E ) = m N If an event can occur in N equally likely and mutually exclusive ways, and if m of these ways possess the characteristic E , then the probability of E is m N 4 / 65

  5. Example: Coin toss Flip one coin Tails and heads equally likely N = 2 possible events Let H=Heads and T=Tails We are interested in the probability of tails: P (Tails)= P ( T ) = 1 2 5 / 65

  6. Relative Frequency Definition P ( E ) = m n If an experiment is repeated n times, and characteristic E occurs m of those times, then the relative frequency of E is m n , and it is approximately equal to the probability of E 6 / 65

  7. Example: Multiple coin tosses I Flip 100 coins Outcome Frequency T = Tails 53 H = Heads 47 Total 100 P (Tails) = P ( T ) ≈ 53 100 = 0 . 53 ≈ 0 . 50 7 / 65

  8. Example: Multiple coin tosses II What happens if we flip 10,000 coins? Outcome Frequency T = Tails 5063 H = Heads 4937 Total 10000 P (Tails) = P ( T ) ≈ 5063 10000 = 0 . 51 ≈ 0 . 50 8 / 65

  9. Relative frequency intuition The probability of T is the limit of the relative frequency of T, as the sample size n goes to infinity “The long run relative frequency” 9 / 65

  10. Outcome characteristics Statistical independence Mutually exclusive 10 / 65

  11. Statistical Independence Two events are statistically independent if the joint probability of both events occurring is the product of the probabilities of each event occuring: P ( A and B ) = P ( A ) × P ( B ) 11 / 65

  12. Example Let A = first born child is female Let B = second child is female P(A and B) = probability that first and second children are both female: Assuming independence: P ( A and B ) = P ( A ) × P ( B ) = 1 2 × 1 2 = 1 4 12 / 65

  13. Statistical independence: comment “In a study where we are selecting patients at random from a population of interest, we assume that the outcomes we observe are independent...” In what situations would this assumption be violated 13 / 65

  14. Mutually exclusive Two events are mutually exclusive if the joint probability of both events occuring is 0: P ( A and B ) = 0 Ex: A = first child is female, B = first child is male 14 / 65

  15. Probability rules 1 The probability of any event is non-negative, and no greater than 1: 0 ≤ P ( E ) ≤ 1 2 Given n mutually exclusive events, E 1 , E 2 , · · · , E n covering the sample space, the sum of the probabilities of events is 1: n � P ( E i ) = P ( E 1 ) + P ( E 2 ) + · · · + P ( E n ) = 1 i =1 3 If E i and E j are mutually exclusive events, then the probability that either E i or E j occur is: P ( E i ∪ E j ) = P ( E i ) + P ( E j ) 15 / 65

  16. Set notation A set is a group of disjoint objects An element of a set is an object in the set The union if two sets, A and B, is a larger set that contains all elements in either A, B or both Notation: A ∪ B The intersection if two sets, A and B, is the set containing all elements found in both A and B Notation: A ∩ B 16 / 65

  17. The addition rule If two events, A and B, are not mutually exclusive, then the probability that event A or event B occurs is: P ( A ∪ B ) = P ( A ) + P ( B ) − P ( A ∩ B ) where P ( A ∩ B ) is the probability that both events occur 17 / 65

  18. Conditional probability The conditional probability of an event A given an event B is: P ( A | B ) = P ( A ∩ B ) P ( B ) where P ( B ) � = 0 18 / 65

  19. The multiplication rule In general: P ( A ∩ B ) = P ( B ) × P ( A | B ) When events A and B are independent, P ( A | B ) = P ( A ) and: P ( A ∩ B ) = P ( A ) × P ( B ) 19 / 65

  20. Bayes rule Useful for computing P ( B | A ) if P ( A | B ) and P ( A | B c ) are known Ex: Screening We know P (test positive | true positive) We want P (true positive | test positive) Ex: Bayesian statistics uses assumptions about P (data | state of the world) to derive statements about P (state of the world | data) The rule: P ( A | B ) · P ( B ) P ( B | A ) = P ( A | B ) · P ( B ) + P ( A | B c ) · P ( B c ) where B c denotes “the complement of B” or “not B” 20 / 65

  21. Example: Sex and Age I Age Age Young ( B 1 ) Older ( B 2 ) Total Male ( A 1 ) 30 20 50 Female ( A 2 ) 40 10 50 Total 70 30 100 21 / 65

  22. Example: Sex and Age II A 1 = { all males } , A 2 = { all females } B 1 = { all young } , B 2 = { all older } A 1 ∪ A 2 = { all people } = B 1 ∪ B 2 A 1 ∩ A 2 = { no people } = ∅ = B 1 ∩ B 2 A 1 ∪ B 1 = { all males and young females } A 1 ∪ B 2 = { all males and older females } A 2 ∩ B 2 = { older females } 22 / 65

  23. Example: Sex and Age III P (male) = 50 P ( A 1 ) = 100 = 0 . 5 P (female) = 50 P ( A 2 ) = 100 = 0 . 5 P (young) = 70 P ( B 1 ) = 100 = 0 . 7 P (older) = 30 P ( B 2 ) = 100 = 0 . 3 23 / 65

  24. Example: Sex and Age IV P (older and female) = 10 P ( A 2 ∩ B 2 ) = 100 = 0 . 1 P ( A 1 ∪ B 1 ) = P (young or male) P ( A 1 ) + P ( B 1 ) − P ( A 1 ∩ B 1 ) = 100 + 70 50 100 − 30 = 100 90 = 100 = 0 . 9 24 / 65

  25. Example: Sex and Age V P ( B 2 | A 2 ) P (older | female) = P ( B 2 ∩ A 2 ) = 10 / 100 50 / 100 = 10 = 50 = 0 . 2 P ( A 2 ) P ( B 2 | A 1 ) P (older | male) = P ( B 2 ∩ A 1 ) = 20 / 100 50 / 100 = 20 = 50 = 0 . 4 P ( A 1 ) P (older) = 30 P ( B 2 ) = 100 = 0 . 3 P ( B 2 | A 2 ) � = P ( B 2 | A 1 ) � = P ( B 2 ) → In this group, sex and age are not independent 25 / 65

  26. Example: Sex and Age VI P ( A 1 ∪ A 2 ) = P ( B 1 ∪ B 2 ) = P ( A 2 | B 2 ) = 26 / 65

  27. Example: Blood Groups I Sex Blood group Male Female Total O 113 170 283 A 103 155 258 B 25 37 62 AB 10 15 25 Total 251 377 628 27 / 65

  28. Example: Blood Groups II 1 − P (female) = 251 628 ≈ 0 . 4 P (male) = 283 628 ≈ 0 . 45 P ( O ) = 258 P ( A ) = 628 ≈ 0 . 41 62 628 ≈ 0 . 10 P ( B ) = 25 628 ≈ 0 . 04 P ( AB ) = 28 / 65

  29. Example: Blood Groups III Question: Are sex and blood group independent? 113 P ( O | male) = 251 ≈ 0 . 45 170 P ( O | female) 377 ≈ 0 . 45 = 283 628 ≈ 0 . 45 same as P ( O ) = Can show same equalities for all blood types → Yes, sex and blood group appear to be independent of each other in this sample 29 / 65

  30. Example: Disease in the population I For patients with Disease X, suppose we knew the age proportions per sex, as well as the sex distribution. Question: Could we compute the sex proportions in each age group (young / older)? Answer: Use Bayes Rule 30 / 65

  31. Example: Disease in the population II P ( A 1 ) = P ( A 2 ) = 0 . 5 P ( B 2 | A 2 ) = 0 . 2 P ( B 2 | A 1 ) = 0 . 4 P ( B 2 | A 2 ) · P ( A 2 ) P ( A 2 | B 2 ) = P ( B 2 | A 2 ) · P ( A 2 ) + P ( B 2 | A 1 ) · P ( A 1 ) 31 / 65

  32. Probability Distributions Often, we assume a true underlying distribution Ex: P (tails) = 1 2 , P (heads) = 1 2 This distribution is characterized by a mathermatical formula and a set of possible outcomes Two types of distributions: Discrete Continuous 32 / 65

  33. Most Commonly Used Discrete Distributions Binomial – two possible outcomes Underlies much of statistical applications to epidemiology Basic model for logistic regression Poisson – uses counts of events of rates Basis for log-linear models 33 / 65

  34. Most Commonly Used Continuous Distributions Normal – bell shaped curve Many characteristics are normally distributed or approximately normally distributed Basic model for linear regression Exponential – useful in describing growth 34 / 65

  35. Counting techniques Factorials: counts the number of ways to arrange things Permutations: counts the number of possible ordered arrangements of subsets Combinations: counts the number of possible unordered arrangements of subsets 35 / 65

  36. Factorials Notation: n! (“n factorial”) Number of possible arrangements of n objects n! = n(n-1)(n-2)(n-3) · · · (3)(2)(1) 36 / 65

  37. Permutations Ordered arrangement of n objects, taken r at a time n ! = n P r ( n − r )! n ( n − 1) · · · ( n − r + 1)( n − r ) · · · 1 = ( n − r )( n − r − 1) · · · 1 n ( n − 1)( n − 2) · · · ( n − r + 1) = 37 / 65

  38. Combinations An arragement of n objects taken r at a time without regard to order � n � n ! = “n choose r” = r !( n − r )! r � 4 � 2!(4 − 2)! = 4 · 3 · 2 · 1 4! = 12 = 2 = 6 2! · 2! 2 Note: the number of combinations is less than or equal to the number of permutations. 38 / 65

  39. The Binomial Distribution You’ve seen it before: 2 x 2 tables and applications Proportions: CIs and tests Sensitivity and Specificity Odds ratio and relative risk Logistic regression 39 / 65

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend