UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability - - PowerPoint PPT Presentation

uq stat2201 2017 lectures 3 and 4 unit 3 probability
SMART_READER_LITE
LIVE PREVIEW

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability - - PowerPoint PPT Presentation

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions. 1 Random Variables 2 A random variable X is a numerical (integer, real, complex, vector etc.) summary of the outcome of the random experiment. The range or support of


slide-1
SLIDE 1

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 – Probability Distributions.

1

slide-2
SLIDE 2

Random Variables

2

slide-3
SLIDE 3

A random variable X is a numerical (integer, real, complex, vector etc.) summary of the outcome of the random experiment. The range or support of the random variable is the set of possible values that it may take. Random variables are usually denoted by capital letters.

3

slide-4
SLIDE 4

A discrete random variable is an integer/real-valued random variable with a finite (or countably infinite) range. A continuous random variable is a real valued random variable with an interval (either finite or infinite) of real numbers for its range.

4

slide-5
SLIDE 5

Experiment ⇒ Outcome, ω, from the sample space. X(ω) ≡ Random Variable (function of the outcome). P

  • X ∈ U
  • = P
  • {ω | X(ω) ∈ U}
  • .

5

slide-6
SLIDE 6

Example: Dig a hole searching for gold. Ω ≡ all possible outcomes (many ways to define this). X ≡ Weight of gold found in grams. P

  • X > 20) = P
  • {ω | X(ω) ∈ U}
  • with U = {x : x > 20}.

6

slide-7
SLIDE 7

Probability Distributions

7

slide-8
SLIDE 8

The probability distribution of a random variable X is a description of the probabilities associated with the possible values

  • f X.

There are several common alternative ways to describe the probability distribution, with some differences between discrete and continuous random variables.

8

slide-9
SLIDE 9

While not the most popular in practice, a unified way to describe the distribution of any scalar valued random variable X (real or integer) is the cumulative distribution function, F(x) = P(X ≤ x). It holds that (1) 0 ≤ F(x) ≤ 1. (2) limx→−∞ F(x) = 0. (3) limx→∞ F(x) = 1. (3) If x ≤ y, then F(x) ≤ F(y). That is, F(·) is non-decreasing.

9

slide-10
SLIDE 10

Examples to understand: F(x) =      0, x < −1, 0.3, −1 ≤ x < 1, 1, 1 ≤ x. F(x) =      0, x < 0, x, 0 ≤ x ≤ 1, 1, 1 ≤ x.

10

slide-11
SLIDE 11

Distributions are often summarised by numbers such as the mean, µ, variance, σ2, or moments. These numbers, in general do not identify the distribution, but hint at the general location, spread and shape. The standard deviation of X is σ = √ σ2 and is particularly useful when working with the Normal distribution. More on these soon.

11

slide-12
SLIDE 12

Discrete Random Variables

12

slide-13
SLIDE 13

Given a discrete random variable X with possible values x1, x2, . . . , xn, the probability mass function of X is, p(x) = P(X = x). Note: In [MonRun2014] and many other sources, the notation used is f (x) (as a pdf of a continuous random variable).

13

slide-14
SLIDE 14

A probability mass function, p(x) satisfies: (1) p(xi) ≥ 0. (2)

n

  • i=1

p(xi) = 1. The cumulative distribution function of a discrete random variable X, denoted as F(x), is F(x) =

  • xi≤x

p(xi).

14

slide-15
SLIDE 15

P(X = xi) can be determined from the jump at the value of x. More specifically p(xi) = P(X = xi) = F(xi) − lim

x↑xi F(xi).

15

slide-16
SLIDE 16

Back to the example: F(x) =      0, x < −1, 0.3, −1 ≤ x < 1, 1, 1 ≤ x. What is the pmf?

16

slide-17
SLIDE 17

The mean or expected value of a discrete random variable X, is µ = E(X) =

  • x

x p(x).

17

slide-18
SLIDE 18

The expected value of h(X) for some function h(·) is: E

  • h(X)
  • =
  • x

h(x) p(x).

18

slide-19
SLIDE 19

The k’th moment of X is, E(X k) =

  • x

xk p(x).

19

slide-20
SLIDE 20

The variance of X, is σ2 = V (X) = E(X − µ)2 =

  • x

(x − µ)2 p(x) =

  • x

x2 p(x) − µ2.

20

slide-21
SLIDE 21

The Discrete Uniform Distribution

21

slide-22
SLIDE 22

A random variable X has a discrete uniform distribution if each

  • f the n values in its range, x1, x2, . . . , xn, has equal probability. I.e.

p(xi) = 1/n.

22

slide-23
SLIDE 23

Suppose that X is a discrete uniform random variable on the consecutive integers a, a + 1, a + 2, . . . , b, for a ≤ b. The mean and variance of X are E(X) = b + a 2 and V (X) = (b − a + 1)2 − 1 12 .

23

slide-24
SLIDE 24

To compute the mean and variance of the discrete uniform, use:

n

  • k=1

k = n(n + 1) 2 ,

n

  • k=1

k2 = n(n + 1)(2n + 1) 6

24

slide-25
SLIDE 25

E(X) = b

k=a k 1 b−a+1 =

25

slide-26
SLIDE 26

E(X 2) = b

k=a k2 1 b−a+1 =

26

slide-27
SLIDE 27

The Binomial Distribution

27

slide-28
SLIDE 28

The setting of n independent and identical Bernoulli trials is as follows: (1) There are n trials. (1) The trials are independent. (2) Each trial results in only two possible outcomes, labelled as “success” and “failure”. (3) The probability of a success in each trial denoted as p is the same for all trials.

28

slide-29
SLIDE 29

Binomial Example: Number of digs finding gold. n = 5 digs in different spots. p = 0.1 chance of finding gold in each spot.

29

slide-30
SLIDE 30

The random variable X that equals the number of trials that result in a success is a binomial random variable with parameters 0 ≤ p ≤ 1 and n = 1, 2, . . . . The probability mass function of X is p(x) = n x

  • px(1 − p)n−x,

x = 0, 1, . . . , n.

30

slide-31
SLIDE 31

Useful to remember from algebra: the binomial expansion for constants a and b is (a + b)n =

n

  • k=0

n k

  • akbn−k.

31

slide-32
SLIDE 32

If X is a binomial random variable with parameters p and n, then, E(X) = n p and V (X) = n p (1 − p).

32

slide-33
SLIDE 33

Example (cont.): Number of digs finding gold (n = 5, p = 0.1):

33

slide-34
SLIDE 34

Continuous Random Variables

34

slide-35
SLIDE 35

Given a continuous random variable X, the probability density function (pdf) is a function, f (x) such that, (1) f (x) ≥ 0. (2) f (x) = 0 for x not in the range. (3)

  • −∞

f (x) dx = 1. (4) For small ∆x, f (x) ∆x ≈ P(X ∈ [x, x + ∆ x)). (5) P(a ≤ X ≤ b) =

b

  • a

f (x)dx = area under f (x) from a to b.

35

slide-36
SLIDE 36

Given the pdf, f (x) we can get the cdf as follows: F(x) = P(X ≤ x) =

x

  • −∞

f (u)du for − ∞ < x < ∞.

36

slide-37
SLIDE 37

Given the cdf, F(x) we can get the pdf: f (x) = d dx F(x).

37

slide-38
SLIDE 38

The mean or expected value of a continous random variable X, is µ = E(X) =

  • −∞

x f (x)dx. The expected value of h(X) for some function h(·) is: E

  • h(X)
  • =

  • −∞

h(x)f (x) dx. The k’th moment of X is, E(X k) =

  • −∞

xk f (x) dx. The variance of X, is σ2 = V (X) =

  • −∞

(x − µ)2f (x)dx = ∞

−∞

x2f (x) dx − µ2.

38

slide-39
SLIDE 39

Continuous Uniform Distribution

39

slide-40
SLIDE 40

A continuous random variable X with probability density function f (x) = 1 b − a, a ≤ x ≤ b. is a continuous uniform random variable or “uniform random variable” for short.

40

slide-41
SLIDE 41

If X is a continuous uniform random variable over a ≤ x ≤ b, the mean and variance are: µ = E(X) = a + b 2 and σ2 = V (X) = (b − a)2 12 .

41

slide-42
SLIDE 42

The Normal Distribution

42

slide-43
SLIDE 43

A random variable X with probability density function f (x) = 1 σ √ 2π e

−(x−µ)2 2σ2

, −∞ < x < ∞, is a normal random variable with parameters µ where −∞ < µ < ∞, and σ > 0. For this distribution, the parameters map directly to the mean and variance, E(X) = µ and V (X) = σ2. The notation N(µ, σ2) is used to denote the distribution. Note that some authors and software packages use σ for the second parameter and not σ2.

43

slide-44
SLIDE 44

A normal random variable with a mean and variance of: µ = 0 and σ2 = 1 is called a standard normal random variable and is denoted as

  • Z. The cumulative distribution function of a standard normal

random variable is denoted as Φ(z) = FZ(z) = P(Z ≤ z), and is tabulated in a table.

44

slide-45
SLIDE 45

It is very common to compute P(a < X < b) for X ∼ N(µ, σ2). This is the typical way: P(a < X < b) = P(a − µ < X − µ < b − µ) = P a − µ σ < X − µ σ < b − µ σ

  • = P

a − µ σ < Z < b − µ σ

  • = Φ

b − µ σ

  • − Φ

a − µ σ

  • .

We get: FX(b) − FX(a) = FZ b − µ σ

  • − FZ

a − µ σ

  • .

45

slide-46
SLIDE 46

The Exponential Distribution

46

slide-47
SLIDE 47

The exponential distribution with parameter λ > 0 is given by the survival function, F(x) = 1 − F(x) = P(X > x) = e−λx. The random variable X represents the distance between successive events from a Poisson process with mean number of events per unit interval λ > 0.

47

slide-48
SLIDE 48

The probability density function of X is f (x) = λe−λx for 0 ≤ x < ∞. Note that sometimes a different parameterisation, θ = 1/λ is used (e.g. in the Julia Distributions package).

48

slide-49
SLIDE 49

The mean and variance are: µ = E(X) = 1 λ and σ2 = V (X) = 1 λ2

49

slide-50
SLIDE 50

The exponential distribution is the only continuous distribution with range [0, ∞) exhibiting the lack of memory property. For an exponential random variable X, P(X > t + s | X > t) = P(X > s).

50

slide-51
SLIDE 51

Monte Carlo Random Variable Generation

51

slide-52
SLIDE 52

Monte Carlo simulation makes use of methods to transform a uniform random variable in a manner where it follows an arbitrary given given distribution. One example of this is if U ∼ Uniform(0, 1) then X = − 1

λ log(U) is exponentially

distributed with parameter λ.

52