Graphical Models Graphical Models Review of probability theory - - PowerPoint PPT Presentation

graphical models graphical models
SMART_READER_LITE
LIVE PREVIEW

Graphical Models Graphical Models Review of probability theory - - PowerPoint PPT Presentation

Graphical Models Graphical Models Review of probability theory Review of probability theory Siamak Ravanbakhsh Winter 2018 Learning objectives Learning objectives Probability distribution and density functions Random variable Bayes' rule


slide-1
SLIDE 1

Graphical Models Graphical Models

Review of probability theory Review of probability theory

Siamak Ravanbakhsh

Winter 2018

slide-2
SLIDE 2

Learning objectives Learning objectives

Probability distribution and density functions Random variable Bayes' rule Conditional independence Expectation and Variance

slide-3
SLIDE 3

Sample space Sample space

: the set of all possible outcomes (a.k.a. outcome space)

Ω = {ω}

Ω = {hhh, hht, hth, … , ttt}

Example1: three tosses of a coin

Ω

image: http://web.mnstate.edu/peil/MDEV102/U3/S25/Cartesian3.PNG

slide-4
SLIDE 4

Ω = {(1, 1), … , (6, 6)}

Example 2: two dice

Image source: http://www.stat.ualberta.ca/people/schmu/preprints/article/Article.htm

: the set of all possible outcomes (a.k.a. outcome space)

Ω = {ω}

Sample space Sample space Ω

slide-5
SLIDE 5

∣Ω∣ = = = 1431 ( 2

54) 2!52! 54!

Example 3: 2 cards from a deck (assuming order doesn't matter)

A A 2 2 3 3 4 4 5 5 6 6 7 7 8 8 9 9 10 10 J J Q Q K K A A 2 2 3 3 4 4 5 5 6 6 7 7 8 8 9 9 10 10 J J Q Q K K A A 2 2 3 3 4 4 5 5 6 6 7 7 8 8 9 9 10 10 J J Q Q K K A A 2 2 3 3 4 4 5 5 6 6 7 7 8 8 9 9 10 10 J J Q Q K K

: the set of all possible outcomes (a.k.a. outcome space)

Ω = {ω}

Sample space Sample space Ω

slide-6
SLIDE 6

Event Event space space

event space is a set of events

S ⊆ 2Ω

S

F ⊆ Ω

An event is a set of outcomes

slide-7
SLIDE 7

Event Event space space

event space is a set of events

S ⊆ 2Ω

Example: Event: at least two heads Event: pair of aces F = {hht, thh, hth, hhh}

∣F∣ = 6

S

F ⊆ Ω

An event is a set of outcomes

slide-8
SLIDE 8

A ∈ S → Ω − A ∈ S

A, B ∈ S → A ∩ B ∈ S

Requirements for event space: Complement of an event is also an event Intersection of events is also an event Example:

at least one head ∈ S → no heads ∈ S at least one head, at least one tail ∈ S → at least one head and one tail ∈ S

Event Event space space S

slide-9
SLIDE 9

Probability distribution Probability distribution

Assigns a real value to each event Probability axioms (Kolmogorov axioms) Probability is non-negative The probability of disjoint events is additive P : S → ℜ

P(A) ≥ 0

A ∩ B = ∅ → P(A ∪ B) = P(A) + P(B)

P(Ω) = 1

slide-10
SLIDE 10

Probability distribution Probability distribution

Probability axioms (Kolmogorov axioms) Probability is non-negative disjoint events are additive:

P(A) ≥ 0

A ∩ B = ∅ → P(A ∪ B) = P(A) + P(B)

P(Ω) = 1

Derivatives:

union bound:

P(∅) = 0

P(A ∪ B) = P(A) + P(B) − P(A ∩ B) P(A ∪ B) ≤ P(A) + P(B)

.

P(Ω\A) = 1 − P(A) P(A ∩ B) ≤ min{P(A), P(B)}

slide-11
SLIDE 11

Probability distribution; Probability distribution; examples examples

S = {∅, Ω}

Ω = {1, 2, 3, 4, 5, 6}

P(∅) = 0, P(Ω) = 1

(a minimal choice of event space)

slide-12
SLIDE 12

Probability distribution; Probability distribution; examples examples

S = {∅, Ω}

Ω = {1, 2, 3, 4, 5, 6}

P(∅) = 0, P(Ω) = 1 S = 2Ω

P(A) =

6 ∣A∣

(a maximal choice of event space)

P({1, 3}) = 6

2

that is (a minimal choice of event space)

(any other consistent assignment is acceptable)

slide-13
SLIDE 13

Conditional Conditional probability probability

Probability of an event A after observing the event B

P(A ∣ B) =

P(B) P(A∩B)

slide-14
SLIDE 14

Conditional Conditional probability probability

Probability of an event A after observing the event B

P(A ∣ B) =

P(B) P(A∩B)

P(B) > 0

slide-15
SLIDE 15

Conditional Conditional probability probability

Probability of an event A after observing the event B

P(A ∣ B) =

P(B) P(A∩B)

Example: three coin tosses

P(at least one head ∣ at least one tail) =

P(at least one tail) P(at least one head and one tail)

P(B) > 0

slide-16
SLIDE 16

Chain Chain rule rule

P(A ∣ B) =

P(B) P(A∩B)

slide-17
SLIDE 17

Chain Chain rule rule

P(A ∣ B) =

P(B) P(A∩B)

Chain rule: P(A ∩ B) = P(B)P(A ∣ B)

slide-18
SLIDE 18

Chain Chain rule rule

P(A ∣ B) =

P(B) P(A∩B)

Chain rule: P(A ∩ B) = P(B)P(A ∣ B) B = C ∩ D

and

slide-19
SLIDE 19

Chain Chain rule rule

P(A ∣ B) =

P(B) P(A∩B)

Chain rule: P(A ∩ B) = P(B)P(A ∣ B) B = C ∩ D

and

P(A ∩ C ∩ D) = P(C ∩ D)P(A ∣ C ∩ D)

slide-20
SLIDE 20

Chain Chain rule rule

P(A ∣ B) =

P(B) P(A∩B)

Chain rule: P(A ∩ B) = P(B)P(A ∣ B) B = C ∩ D

and

P(A ∩ C ∩ D) = P(C ∩ D)P(A ∣ C ∩ D) P(A ∩ C ∩ D) = P(D)P(C ∣ D)P(A ∣ C ∩ D)

slide-21
SLIDE 21

Chain Chain rule rule

P(A ∣ B) =

P(B) P(A∩B)

Chain rule: P(A ∩ B) = P(B)P(A ∣ B) B = C ∩ D

and

P(A ∩ C ∩ D) = P(C ∩ D)P(A ∣ C ∩ D) P(A ∩ C ∩ D) = P(D)P(C ∣ D)P(A ∣ C ∩ D) More generally:

P(A ∩ … ∩ A ) = P(A )P(A ∣ A ) … P(A ∣ A ∩ … ∩ A )

1 n 1 2 1 n 1 n−1

slide-22
SLIDE 22

Bayes Bayes' rule ' rule

P(A ∣ B) =

P(B) P(B∣A)P(A)

Reasoning about the event A:

  • ur prior belief about A

likelihood of the event B if A were to happen

  • ur posterior belief about A after
  • bserving B
slide-23
SLIDE 23

Bayes' rule; Bayes' rule; example example

P(A ∣ B) =

P(B) P(B∣A)P(A) prior

likelihood

posterior

1% of the population has cancer cancer test False positive 10% False negative 5% chance of having cancer given a positive test result? sample space? events A, B? prior? lilkelihood? {TP, TN, FP, FN} A = {TP, FN}, B = {TP, TN} P(A) = .01, P(B|A) = .9

slide-24
SLIDE 24

Bayes' rule; Bayes' rule; example example

P(A ∣ B) =

P(B) P(B∣A)P(A) prior

likelihood

posterior

1% of the population has cancer cancer test False positive 10% False negative 5% chance of having cancer given a positive test result? sample space? events A, B? prior? lilkelihood? {TP, TN, FP, FN} A = {TP, FN}, B = {TP, TN} P(A) = .01, P(B|A) = .9 P(B) is not trivial

P(cancer ∣ +) ∝ P(+ ∣ cancer)P(cancer) = .009 P(cancer ∣ −) ∝ P(+ ∣ cancer)P(cancer) = .99 × .1 = .099

P(cancer ∣ +) = ≈ .08

.009+.099 .009

slide-25
SLIDE 25

Independence Independence

Observing A does not change P(B)

P(A ∩ B) = P(A)P(B)

Events A and B are independent iff

P ⊨ (A ⊥ B)

slide-26
SLIDE 26

Independence Independence

Observing A does not change P(B)

P(A ∩ B) = P(A)P(B)

Events A and B are independent iff

P(A ∩ B) = P(A)P(B ∣ A)

using Equivalent definition: or

P(B) = P(B ∣ A)

P(A) = 0

P ⊨ (A ⊥ B)

slide-27
SLIDE 27

Independence; Independence; example example

Are A and B independent?

Ω A B

slide-28
SLIDE 28

Independence; Independence; example example

Example 1:

P(h * * ∣ * t *) = P(h * *) = 2

1

P(hhh) = P(hht) … = P(ttt) = 8

1

equivalently: P(h t *) = P(* t *)P(h * *) = 4

1

slide-29
SLIDE 29

Independence; Independence; example example

Example 1:

P(h * * ∣ * t *) = P(h * *) = 2

1

Example 2: are these two events independent? P(hhh) = P(hht) … = P(ttt) = 8

1

P({ht, hh}) = .3, P({th}) = .1 equivalently: P(h t *) = P(* t *)P(h * *) = 4

1

slide-30
SLIDE 30

Conditional Conditional independence independence

P(A ∩ B ∣ C) = P(A ∣ C)P(B ∣ C)

P ⊨ (A ⊥ B ∣ C)

a more common phenomenon:

slide-31
SLIDE 31

Conditional Conditional independence independence

P(A ∩ B ∣ C) = P(A ∣ C)P(B ∣ C)

P(A ∩ B ∣ C) = P(A ∣ C)P(B ∣ A ∩ C) using

P ⊨ (A ⊥ B ∣ C)

a more common phenomenon:

slide-32
SLIDE 32

Conditional Conditional independence independence

P(A ∩ B ∣ C) = P(A ∣ C)P(B ∣ C)

P(A ∩ B ∣ C) = P(A ∣ C)P(B ∣ A ∩ C) using Equivalent definition: P(A ∩ C) = 0 P(B ∣ C) = P(B ∣ A ∩ C) or

P ⊨ (A ⊥ B ∣ C)

a more common phenomenon:

slide-33
SLIDE 33

Conditional independence; Conditional independence; example example

P(A ∩ B ∣ C) = P(A ∣ C)P(B ∣ C)

Generalization of independence:

P ⊨ (R ⊥ B ∣ Y )

Ω

from: wikipedia

slide-34
SLIDE 34

Summary Summary

Outcome space: a set Event: a subset of outcomes Event space: a set of events Probability dist. is associated with events Conditional probability: based on intersection of events Chain rule follows from conditional probability (Conditional) independence: relevance of some events to others Basics of probability

slide-35
SLIDE 35

Random Variable Random Variable

is an attribute associated with each outcome

X : Ω → V al(X)

a formalism to define events

P(X = x) ≜ P({ω ∈ Ω ∣ X(ω) = x})

intensity of a pixel head/tail value of the first coin in multiple coin tosses first draw from a deck is larger than the second

slide-36
SLIDE 36

Random Variable Random Variable

is an attribute associated with each outcome

X : Ω → V al(X)

a formalism to define events

P(X = x) ≜ P({ω ∈ Ω ∣ X(ω) = x})

intensity of a pixel head/tail value of the first coin in multiple coin tosses first draw from a deck is larger than the second Example: three tosses of coin number of heads number of heads in the first two trials at least one head

X : Ω → {0, 1, 2, 3}

1

X : Ω → {0, 1, 2}

2

X : Ω → {True, False}

3

slide-37
SLIDE 37

Random Variable ( Random Variable (RV RV)

is an attribute associated with each outcome a formalism to define events

P(X = x) ≜ P({ω ∈ Ω ∣ X(ω) = x})

Multiple RVs:

  • utcomes that we care about:

cannonical outcome space: X = x , … , X = x

1 1 n n

X , … , X

1 n

X : Ω → V al(X)

Ω ≜ V al(X ) × … × V al(X )

c 1 n

slide-38
SLIDE 38

Random Variable ( Random Variable (RV RV)

is an attribute associated with each outcome a formalism to define events

P(X = x) ≜ P({ω ∈ Ω ∣ X(ω) = x})

Multiple RVs:

  • utcomes that we care about:

cannonical outcome space: X = x , … , X = x

1 1 n n

X , … , X

1 n

P(X = x , … , X = x ) ≜ P(X = x ∩ … ∩ X = x )

1 1 n n 1 1 n n

X : Ω → V al(X)

Ω ≜ V al(X ) × … × V al(X )

c 1 n

slide-39
SLIDE 39

Random Variable ( Random Variable (RV RV)

is an attribute associated with each outcome a formalism to define events

P(X = x) ≜ P({ω ∈ Ω ∣ X(ω) = x})

Multiple RVs:

  • utcomes that we care about:

cannonical outcome space: X = x , … , X = x

1 1 n n

X , … , X

1 n

P(X = x , … , X = x ) ≜ P(X = x ∩ … ∩ X = x )

1 1 n n 1 1 n n

P(X = x ) = P(X = x , … , X = x )

1 1

∑x ,…,x

2 n

1 1 n n

X : Ω → V al(X)

Ω ≜ V al(X ) × … × V al(X )

c 1 n

slide-40
SLIDE 40

Random Variable ( Random Variable (RV RV)

is an attribute associated with each outcome a formalism to define events

P(X = x) ≜ P({ω ∈ Ω ∣ X(ω) = x})

Multiple RVs:

  • utcomes that we care about:

cannonical outcome space: joint probability: X = x , … , X = x

1 1 n n

X , … , X

1 n

P(X = x , … , X = x ) ≜ P(X = x ∩ … ∩ X = x )

1 1 n n 1 1 n n

P(X = x ) = P(X = x , … , X = x )

1 1

∑x ,…,x

2 n

1 1 n n

X : Ω → V al(X)

Ω ≜ V al(X ) × … × V al(X )

c 1 n

slide-41
SLIDE 41

Random Variable ( Random Variable (RV RV)

is an attribute associated with each outcome a formalism to define events

P(X = x) ≜ P({ω ∈ Ω ∣ X(ω) = x})

Multiple RVs:

  • utcomes that we care about:

cannonical outcome space: joint probability: marginal probability: X = x , … , X = x

1 1 n n

X , … , X

1 n

P(X = x , … , X = x ) ≜ P(X = x ∩ … ∩ X = x )

1 1 n n 1 1 n n

P(X = x ) = P(X = x , … , X = x )

1 1

∑x ,…,x

2 n

1 1 n n

X : Ω → V al(X)

Ω ≜ V al(X ) × … × V al(X )

c 1 n

slide-42
SLIDE 42

Random Variable; Random Variable; example example

a joint probability

Three tosses of coin

1 2 3 P(X2) True .1 .1 .4 .05 .65 False .2 .01 .09 .05 .35 P(X1) .3 .11 .49 .1

number of heads first trial is a head

X : Ω → {0, 1, 2, 3}

1

X : Ω → {True, False}

2

Cannonical outcome space:

Ω = {(0, True), … , (3, False)}

c

atomic outcome

marginal probability

slide-43
SLIDE 43

Conditional independence Conditional independence for RVs for RVs

Given random variables X, Y, Z iff

P ⊨ (X ⊥ Y ∣ Z) P ⊨ (X = x ⊥ Y = y ∣ Z = z) ∀x, y, z

Therefore iff

P ⊨ (X ⊥ Y ∣ Z)

P(X, Y ∣ Z) = P(X ∣ Z)P(Y ∣ Z) P(X ∣ Y , Z) = P(X ∣ Z)

OR Marginal independence: P ⊨ (X ⊥ Y ∣ ∅)

slide-44
SLIDE 44

Continuous Continuous domain domain

probability density function (pdf)

p : V al(X) → [0, +∞) s.t. p(x)dx = 1 ∫V al(X)

P(X ≤ a) ≜ p(x)dx ∫−∞

a

the cumulative distribution function (cdf)

F(a) :

p(x)

slide-45
SLIDE 45

Continuous Continuous domain domain

probability density function (pdf)

p : V al(X) → [0, +∞) s.t. p(x)dx = 1 ∫V al(X)

note that often can be larger than 1 it is not a probability distribution

P(X ≤ a) ≜ p(x)dx ∫−∞

a

the cumulative distribution function (cdf)

F(a) :

P(X = x) = 0

p(x)

P(a ≤ X ≤ b) = F(b) − F(a)

p(x)

slide-46
SLIDE 46

Continuous Continuous domain domain

probability density function (pdf)

p : V al(X) → [0, +∞) s.t. p(x)dx = 1 ∫V al(X)

for discrete domains: probability mass function (pmf) p(x) ≜ P(X = x) s.t. p(x) = 1 ∑V al(X)

slide-47
SLIDE 47

Continuous Continuous domain; domain; multivariate multivariate case

case

Joint density of multipe RVs: (same conditions)

P(X ≤ a , … , X ≤ a ) ≜ … p(x , … , x )dx … dx

1 1 n n

∫−∞

a1

∫−∞

an 1 n n 1

joint CDF

F(a , … , a ) :

1 n

slide-48
SLIDE 48

Continuous Continuous domain; domain; multivariate multivariate case

case

Joint density of multipe RVs: (same conditions)

P(X ≤ a , … , X ≤ a ) ≜ … p(x , … , x )dx … dx

1 1 n n

∫−∞

a1

∫−∞

an 1 n n 1

joint CDF

F(a , … , a ) :

1 n

Marginal density: marginal CDF p(x ) = … p(x , … , x )dx … dx

1

∫−∞

a2

∫−∞

an 1 n n 2

F(x ) = lim F(x , … , x )

1 x ,…,x →∞

2 n

1 n

slide-49
SLIDE 49

Continuous Continuous domain; domain; conditional conditional case

case

Conditional distribution:

zero measure!

P(X ∣ Y = y) =

P(Y =y) P(X,Y =y)

Take the limit in:

P(X ≤ a ∣ y − ϵ ≤ Y ≤ y + ϵ) =

p(y+e)de ∫e=−ϵ

ϵ

p(x,y+e)dedx ∫−∞

a

∫e=−ϵ

ϵ

ϵ → 0

slide-50
SLIDE 50

Continuous Continuous domain; domain; conditional conditional case

case

Conditional distribution:

zero measure!

P(X ∣ Y = y) =

P(Y =y) P(X,Y =y)

Take the limit in:

P(X ≤ a ∣ y − ϵ ≤ Y ≤ y + ϵ) =

p(y+e)de ∫e=−ϵ

ϵ

p(x,y+e)dedx ∫−∞

a

∫e=−ϵ

ϵ

ϵ → 0

using

f(y + e)de = 2ϵf(y) + O(ϵ ) ∫e=−ϵ

ϵ 2

P(X ≤ a ∣ y − ϵ ≤ Y ≤ y + ϵ) ≈

p(y) p(x,y)dx ∫−∞

a

slide-51
SLIDE 51

Continuous Continuous domain; domain; conditional conditional case

case

Conditional distribution:

zero measure!

P(X ∣ Y = y) =

P(Y =y) P(X,Y =y)

Conditional density of is

p(x ∣ y) =

p(y) p(x,y)

Take the limit in:

P(X ≤ a ∣ y − ϵ ≤ Y ≤ y + ϵ) =

p(y+e)de ∫e=−ϵ

ϵ

p(x,y+e)dedx ∫−∞

a

∫e=−ϵ

ϵ

ϵ → 0

using

f(y + e)de = 2ϵf(y) + O(ϵ ) ∫e=−ϵ

ϵ 2

P(X ≤ a ∣ y − ϵ ≤ Y ≤ y + ϵ) ≈

p(y) p(x,y)dx ∫−∞

a

P(X ∣ Y = y)

extends Bayes' rule and chain rule and conditional independence to densities

slide-52
SLIDE 52

Functions Functions of random variables

  • f random variables

RV is a function of the outcome therefore is an RV itself E.g.,

X : Ω → V al(X)

g(X) = g(X(ω))

Y = X + X

1 2

slide-53
SLIDE 53

Expectation Expectation & Variance & Variance

Expectation: linearity:

X:# heads, Y:#heads in the first trial (X&Y are not independent)

for independent X & Y

E[X] ≜ xp(x) ∑x∈V al(X) E[X] ≜ xp(x)dx ∫x∈V al(X)

OR

E[X + aY ] = E[X] + aE[Y ]

E[XY ] = p(x, y)(xy) = p(x)p(y)(xy) ∑x,y∈V al(X)×V al(Y ) ∑x,y∈V al(X)×V al(Y )

= ( xp(x))( yp(y)) = E[X]E[Y ] ∑x∈V al(X) ∑y∈V al(Y )

slide-54
SLIDE 54

Expectation & Expectation & Variance Variance

Variance: V ar[X] ≜ E[(X − E[X]) ]

2

= E[X + E[X] − 2XE[X]] = E[X ] + E[X] − 2E[X]E[X] = E[X ] − E[X]

2 2 2 2 2 2

for independent X and Y if not independent Covariance: generalizes variance symmetric & bilinear

V ar[X + Y ] = V ar[X] + V ar[Y ] V ar[X + Y ] = V ar[X] + V ar[Y ] + 2 Cov[X, Y ]

Cov[X, Y ] ≜ E[XY − E[XY ]] = E[XY − E[X]E[Y ] Cov[X, X] = V ar[X]

Cov[aX, bY ] = abCov[Y , X]

slide-55
SLIDE 55

Classical members of exponential family of distribution

Gaussian Bernoulli Binomial Multinomial Gamma Exponential Poisson Beta Dirichlet

Examples Examples of probability dists.

  • f probability dists.

more on this later

slide-56
SLIDE 56

Bernoulli: discrete distribution with

P(X = 1; μ) = μ 0 ≤ μ ≤ 1

V al(X) = {0, 1} p(x; μ) = μ (1 − μ)

x 1−x

Binomial:

  • dist. over the number of ones in n independent Bernoulli trials

number of heads in n coin toss

V al(X) = {0, … , n} P(X = k; μ, n) = μ (1 − μ) (k

n) k n−k

OR

Examples Examples of probability dists.

  • f probability dists.
slide-57
SLIDE 57

Categorical (aka. multinulli) : fully parameterized discrete distribution with

V al(X) = {0 … , L}

P(X = l; μ) = μ where μ = 1

l

∑l

l

Multinomial distribution:

  • dist. over the number of different outcomes in n

independent categorial trials

P(X = x , … , X = x ; μ, n) = I( x = n) μ

1 1 L L

∑l

l x ! ∏l

l

n!

∏l

l xl

Examples Examples of probability dists.

  • f probability dists.
slide-58
SLIDE 58

Uniform:

CONTINUOUS

p(x)

DISCRETE

max-entropy discrete distribution P(X = j) = n

1

V al(X) = [a, b] V al(X) = {a, a + 1, … , b}

Examples Examples of probability dists.

  • f probability dists.
slide-59
SLIDE 59

Gaussian: motivated by central limit theorem max-entropy dist. with a fixed variance

p(x; μ, σ) = e

√ 2πσ2 1 −

2σ2 (x−μ)2

Examples Examples of probability dists.

  • f probability dists.
slide-60
SLIDE 60

Summary Summary

Random variable: assigns a value to each outcome

Event (using RV): set of outcomes with a particular attribute

  • Prob. dist., cond. prob., chain rule, indep. ... are all extended to RVs

Continuous domains: same definition of probability, event, RV etc.

Specifying the prob. dist. using density function

Adding random variables

slide-61
SLIDE 61

Summary Summary

random variable variable PDF, PMF probability distribution domain of an RV Notation

X, Y , Z

X = [X , … , X ]

1 n

p(x), p(x), p(x, y) x, y, z P(X), P(x) ≜ P(X = x) V al(X), V al(X, Y , Z)

use interchangeably

slide-62
SLIDE 62

bonus slides

slide-63
SLIDE 63

Properties Properties of conditional independence

  • f conditional independence

Symmetry: Decomposition: Weak union: Contraction: Intersection: if P is positive

(X ⊥ Y ∣ Z) ⇒ (Y ⊥ X ∣ Z)

image: Pearl's book

(X ⊥ Y , W ∣ Z) ⇒ (X ⊥ Y ∣ Z)

(X ⊥ Y , W ∣ Z) ⇒ (X ⊥ Y ∣ W, Z)

(X ⊥ W ∣ Y , Z)&(X ⊥ Y ∣ Z) ⇒ (X ⊥ Y , W ∣ Z)

(X ⊥ Y ∣ W, Z)&(X ⊥ W ∣ Y , Z) ⇒ (X ⊥ Y , W ∣ Z)

slide-64
SLIDE 64

Poisson: frequency of rare events events are assumed independent p(x; λ) = where λ > 0

x! λ e

x −λ

is the mean frequency

(rate parameter)

V al(X) = Z+

similar to binomial with large number of trials (λ ≈ nμ)

Examples Examples of probability dists.

  • f probability dists.
slide-65
SLIDE 65

Exponential: time between events in Poisson dist. memoryless property p(x; λ) = λe where λ > 0

−λx

V al(X) = R+

Geometric: number of Bernoulli trials until success memoryless property

V al(X) = N

p(x, k; μ) = (1 − μ) μ where 0 < μ < 1

k−1

(1 − μ) ≡ e−λ

Examples Examples of probability dists.

  • f probability dists.