statistical preliminaries
play

Statistical Preliminaries Stony Brook University CSE545, Fall 2016 - PowerPoint PPT Presentation

Statistical Preliminaries Stony Brook University CSE545, Fall 2016 Random Variables X : A mapping from to that describes the question we care about in practice. 2 Random Variables X : A mapping from to that describes the question


  1. Statistical Preliminaries Stony Brook University CSE545, Fall 2016

  2. Random Variables X : A mapping from Ω to ℝ that describes the question we care about in practice. 2

  3. Random Variables X : A mapping from Ω to ℝ that describes the question we care about in practice. Example: Ω = 5 coin tosses = {<HHHHH>, <HHHHT>, <HHHTH>, <HHHTH>…} We may just care about how many tails? Thus, X(<HHHHH>) = 0 X(<HHHTH>) = 1 X(<TTTHT>) = 4 X(<HTTTT>) = 4 X only has 6 possible values: 0, 1, 2, 3, 4, 5 What is the probability that we end up with k = 4 tails? P (X = k ) := P ( {ω : X(ω) = k} ) where ω ∊ Ω 3

  4. Random Variables X : A mapping from Ω to ℝ that describes the question we care about in practice. Example: Ω = 5 coin tosses = {<HHHHH>, <HHHHT>, <HHHTH>, <HHHTH>…} We may just care about how many tails? Thus, X(<HHHHH>) = 0 X(<HHHTH>) = 1 X(<TTTHT>) = 4 X(<HTTTT>) = 4 X only has 6 possible values: 0, 1, 2, 3, 4, 5 What is the probability that we end up with k = 4 tails? P (X = k ) := P ( {ω : X(ω) = k} ) where ω ∊ Ω X(ω) = 4 for 5 out of 32 sets in Ω . Thus, assuming a fair coin, P (X = 4) = 5/32 (Not a variable, but a function that we end up notating a lot like a variable) 4

  5. Random Variables X : A mapping from Ω to ℝ that describes the question we care about in practice. Example: Ω = 5 coin tosses = { <HHHHH>, <HHHHT>, <HHHTH>, <HHHTH> …} We may just care about how many tails? Thus, X(<HHHHH>) = 0 X is a discrete random variable X(<HHHTH>) = 1 if it takes only a countable X(<TTTHT>) = 4 number of values. X(<HTTTT>) = 4 X only has 6 possible values: 0, 1, 2, 3, 4, 5 What is the probability that we end up with k = 4 tails? P (X = k ) := P ( {ω : X(ω) = k} ) where ω ∊ Ω X(ω) = 4 for 5 out of 32 sets in Ω . Thus, assuming a fair coin, P (X = 4) = 5/32 (Not a variable, but a function that we end up notating a lot like a variable) 5

  6. Random Variables X : A mapping from Ω to ℝ that describes the question we care about in practice. X is a continuous random variable if it X is a discrete random variable can take on an infinite number of if it takes only a countable values between any two given values. number of values. 6

  7. Random Variables X : A mapping from Ω to ℝ that describes the question we care about in practice. Example: Ω = inches of snowfall = [0, ∞) ⊆ ℝ X amount of inches in a snowstorm X is a continuous random variable if it can take on an infinite number of X (ω) = ω values between any two given values. What is the probability we receive (at least) a inches? P (X ≥ a) := P ( {ω : X(ω) ≥ a} ) What is the probability we receive between a and b inches? P (a ≤ X ≤ b) := P ( {ω : a ≤ X(ω) ≤ b} ) 7

  8. Random Variables X : A mapping from Ω to ℝ that describes the question we care about in practice. Example: Ω = inches of snowfall = [0, ∞) ⊆ ℝ X amount of inches in a snowstorm X is a continuous random variable if it can take on an infinite number of X (ω) = ω P (X = i) := 0, for all i ∊ Ω values between any two given values. (probability of receiving exactly i What is the probability we receive (at least) a inches? inches of snowfall is zero) P (X ≥ a) := P ( {ω : X(ω) ≥ a} ) What is the probability we receive between a and b inches? P (a ≤ X ≤ b) := P ( {ω : a ≤ X(ω) ≤ b} ) 8

  9. Random Variables, Revisited X : A mapping from Ω to ℝ that describes the question we care about in practice. Example: Ω = inches of snowfall = [0, ∞) ⊆ ℝ X amount of inches in a snowstorm X is a continuous random variable if it can take on an infinite number of X (ω) = ω P (X = i) := 0, for all i ∊ Ω values between any two given values. (probability of receiving exactly i What is the probability we receive (at least) a inches? inches of snowfall is zero) P (X ≥ a) := P ( {ω : X(ω) ≥ a} ) How to model? What is the probability we receive between a and b inches? P (a ≤ X ≤ b) := P ( {ω : a ≤ X(ω) ≥ b} ) 9

  10. Continuous Random Variables Discretize them! (group into discrete bins) How to model? 10

  11. Continuous Random Variables P (bin=8) = .32 P (bin=12) = .08 But aren’t we throwing away information? 11

  12. Continuous Random Variables 12

  13. Continuous Random Variables X is a continuous random variable if it can take on an infinite number of values between any two given values. X is a continuous random variable if there exists a function fx such that: 13

  14. Continuous Random Variables X is a continuous random variable if it can take on an infinite number of values between any two given values. X is a continuous random variable if there exists a function fx such that: fx : “probability density function” (pdf) 14

  15. Continuous Random Variables 15

  16. Continuous Random Variables 16

  17. Continuous Random Variables Common Trap ● does not yield a probability ○ does ○ � may be anything ( ℝ ) ■ thus, may be > 1 17

  18. Continuous Random Variables A Common Probability Density Function 18

  19. Continuous Random Variables Common pdf s: Normal( μ , σ 2 ) = 19

  20. Continuous Random Variables Common pdf s: Normal( μ , σ 2 ) = μ : mean (or “center”) = expectation σ 2 : variance, σ : standard deviation 20

  21. Continuous Random Variables Common pdf s: Normal( μ , σ 2 ) Credit: Wikipedia = μ : mean (or “center”) = expectation σ 2 : variance, σ : standard deviation 21

  22. Continuous Random Variables Common pdf s: Normal( μ , σ 2 ) X ~ Normal( μ , σ 2 ), examples: ● height ● intelligence/ability ● measurement error ● averages (or sum) of lots of random variables 22

  23. Continuous Random Variables Common pdf s: Normal(0, 1) (“standard normal”) How to “standardize” any normal distribution: ● subtract the mean, μ (aka “mean centering”) ● divide by the standard deviation, σ z = (x - μ ) / σ , (aka “z score”) Credit: MIT Open Courseware: Probability and Statistics 23

  24. Continuous Random Variables Common pdf s: Normal(0, 1) Credit: MIT Open Courseware: Probability and Statistics 24

  25. Cumulative Distribution Function For a given random variable X, the cumulative distribution function (CDF), Uniform Fx: ℝ → [0, 1] , is defined by: Normal 25

  26. Cumulative Distribution Function For a given random variable X, the cumulative distribution function (CDF), Uniform Fx: ℝ → [0, 1] , is defined by: Pro: yields a probability! Exponential Con: Not intuitively interpretable. Normal 26

  27. Random Variables, Revisited X : A mapping from Ω to ℝ that describes the question we care about in practice. X is a continuous random variable if it X is a discrete random variable can take on an infinite number of if it takes only a countable values between any two given values. number of values. 27

  28. Discrete Random Variables For a given random variable X, the cumulative distribution function (CDF), Fx: ℝ → [0, 1] , is defined by: X is a discrete random variable if it takes only a countable number of values. 28

  29. Discrete Random Variables For a given random variable X, the cumulative distribution function (CDF), Fx: ℝ → [0, 1] , is defined by: X is a discrete random variable if it takes only a countable number of values. Binomial (n, p) (like normal) 29

  30. Discrete Random Variables Binomial (n, p) For a given random variable X, the cumulative distribution function (CDF), Fx: ℝ → [0, 1] , is defined by: X is a discrete random variable if it takes only a countable number of values. For a given discrete random variable X, probability mass function ( pmf ), fx: ℝ → [0, 1] , is defined by: 30

  31. Discrete Random Variables Binomial (n, p) Two Common Discrete Random Variables ● Binomial(n, p) example: number of heads after n coin flips (p, probability of heads) ● Bernoulli(p) = Binomial(1, p) example: one trial of success or failure 31

  32. Hypothesis Testing Hypothesis -- something one asserts to be true. Classical Approach: H 0 : null hypothesis -- some “default” value; “null” => nothing changes H 1 : the alternative -- the opposite of the null => a change or a difference

  33. Hypothesis Testing Hypothesis -- something one asserts to be true. Classical Approach: H 0 : null hypothesis -- some “default” value; “null” => nothing changes H 1 : the alternative -- the opposite of the null => a change or a difference Goal: Use probability to determine if we can “reject the null”( H 0 ) in favor of H 1 . “ There is less than a 5% chance that the null is true” (i.e. 95% alternative is true). Example: Hypothesize a coin is biased. H 0 : the coin is not biased (i.e. flipping n times results in a Binomial(n, 0.5))

  34. Hypothesis Testing Hypothesis -- something one asserts to be true. H 0 : null hypothesis -- some “default” value (usually that one’s hypothesis is false) Classical Approach: H 1 : the alternative -- usually that one’s “hypothesis” is true H 0 : null hypothesis -- some “default” value (usually that one’s hypothesis is false) More formally: Let X be a random variable and let R be the range of X. R reject ⊂ R is the rejection region. If X ∊ R reject then we reject the null.

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend