Markov Processes in Isabelle/HOL Applications Probabilistic - - PowerPoint PPT Presentation

markov processes in isabelle hol
SMART_READER_LITE
LIVE PREVIEW

Markov Processes in Isabelle/HOL Applications Probabilistic - - PowerPoint PPT Presentation

CPP 2017 Johannes Hlzl (Technical University of Munich) Markov Processes in Isabelle/HOL Applications Probabilistic programming, Continuous-time Markov Example proc x randomised walk on proc stream m proc x do x Normal x 0 1 proc x


slide-1
SLIDE 1

Markov Processes in Isabelle/HOL

Johannes Hölzl (Technical University of Munich) CPP 2017

slide-2
SLIDE 2

Formalization for Modelling Stochastic Processes

Applications Probabilistic programming, Continuous-time Markov chains (queuing theory, biological processes, …), physical processes with errors, … Example proc x — randomised walk on Next step is Normal distributed with variance 0 1. proc

m

stream proc x do x Normal x 0 1 proc x return x Wanted A general method to construct processes: proc x do y K x proc y return y

1

slide-3
SLIDE 3

Formalization for Modelling Stochastic Processes

Applications Probabilistic programming, Continuous-time Markov chains (queuing theory, biological processes, …), physical processes with errors, … Example proc x — randomised walk on Next step is Normal distributed with variance 0 1. proc

m

stream proc x do x Normal x 0 1 proc x return x Wanted A general method to construct processes: proc x do y K x proc y return y

1

slide-4
SLIDE 4

Formalization for Modelling Stochastic Processes

Applications Probabilistic programming, Continuous-time Markov chains (queuing theory, biological processes, …), physical processes with errors, … Example proc x — randomised walk on Next step is Normal distributed with variance 0 1. proc

m

stream proc x do x Normal x 0 1 proc x return x Wanted A general method to construct processes: proc x do y K x proc y return y

1

slide-5
SLIDE 5

Formalization for Modelling Stochastic Processes

Applications Probabilistic programming, Continuous-time Markov chains (queuing theory, biological processes, …), physical processes with errors, … Example proc x — randomised walk on Next step is Normal distributed with variance 0 1. proc

m

stream proc x do x Normal x 0 1 proc x return x Wanted A general method to construct processes: proc x do y K x proc y return y

1

slide-6
SLIDE 6

Formalization for Modelling Stochastic Processes

Applications Probabilistic programming, Continuous-time Markov chains (queuing theory, biological processes, …), physical processes with errors, … Example proc x — randomised walk on R Next step is Normal distributed with variance 0 1. proc

m

stream proc x do x Normal x 0 1 proc x return x Wanted A general method to construct processes: proc x do y K x proc y return y

1

slide-7
SLIDE 7

Formalization for Modelling Stochastic Processes

Applications Probabilistic programming, Continuous-time Markov chains (queuing theory, biological processes, …), physical processes with errors, … Example proc x — randomised walk on R Next step is Normal distributed with variance 0.1. proc

m

stream proc x do x Normal x 0 1 proc x return x Wanted A general method to construct processes: proc x do y K x proc y return y

1

slide-8
SLIDE 8

Formalization for Modelling Stochastic Processes

Applications Probabilistic programming, Continuous-time Markov chains (queuing theory, biological processes, …), physical processes with errors, … Example proc x — randomised walk on R Next step is Normal distributed with variance 0.1. proc ∈ R →m Pr(stream(R)) proc x = do { x′ ← Normal (x, 0.1) ω ← proc x′ return x′·ω } Wanted A general method to construct processes: proc x do y K x proc y return y

1

slide-9
SLIDE 9

Formalization for Modelling Stochastic Processes

Applications Probabilistic programming, Continuous-time Markov chains (queuing theory, biological processes, …), physical processes with errors, … Example proc x — randomised walk on R Next step is Normal distributed with variance 0.1. proc ∈ R →m Pr(stream(R)) proc x = do { x′ ← Normal (x, 0.1) ω ← proc x′ return x′·ω } Wanted A general method to construct processes: proc x = do {y ← K x; ω ← proc y; return (y·ω)}

1

slide-10
SLIDE 10

Formalization for Modelling Stochastic Processes

Applications Probabilistic programming, Continuous-time Markov chains (queuing theory, biological processes, …), physical processes with errors, … Example proc x — randomised walk on R Next step is Normal distributed with variance 0.1. proc ∈ R →m Pr(stream(R)) proc x = do { x′ ← Normal (x, 0.1) ω ← proc x′ return x′·ω } Wanted A general method to construct processes: proc x = do {y ← K x; ω ← proc y; return (y·ω)}

1

slide-11
SLIDE 11

Overview

  • Formalize Markov Processes
  • Transition Function
  • Construction Method
  • Basic Properties
  • Application: Continuous-Time Markov Chains
  • Transition Rates
  • Construction Method
  • Properties
  • Discussion

2

slide-12
SLIDE 12

Markov Processes

2

slide-13
SLIDE 13

Giry Monad

Monad on probability spaces Pr(S) for a measurable space S Monad Combinators Map map

m m

Bind bind

m

Return return

m

Note: Functions are regular HOL functions, theorems have measurability assumptions. We use do notation

3

slide-14
SLIDE 14

Giry Monad

Monad on probability spaces Pr(S) for a measurable space S Monad Combinators Map map ∈ (S →m T) → (Pr(S) →m Pr(T)) Bind bind

m

Return return

m

Note: Functions are regular HOL functions, theorems have measurability assumptions. We use do notation

3

slide-15
SLIDE 15

Giry Monad

Monad on probability spaces Pr(S) for a measurable space S Monad Combinators Map map ∈ (S →m T) → (Pr(S) →m Pr(T)) Bind bind ∈ Pr(S) → (S →m Pr(T)) → Pr(T) Return return

m

Note: Functions are regular HOL functions, theorems have measurability assumptions. We use do notation

3

slide-16
SLIDE 16

Giry Monad

Monad on probability spaces Pr(S) for a measurable space S Monad Combinators Map map ∈ (S →m T) → (Pr(S) →m Pr(T)) Bind bind ∈ Pr(S) → (S →m Pr(T)) → Pr(T) Return return ∈ S →m Pr(S) Note: Functions are regular HOL functions, theorems have measurability assumptions. We use do notation

3

slide-17
SLIDE 17

Giry Monad

Monad on probability spaces Pr(S) for a measurable space S Monad Combinators Map map ∈ (S →m T) → (Pr(S) →m Pr(T)) Bind bind ∈ Pr(S) → (S →m Pr(T)) → Pr(T) Return return ∈ S →m Pr(S) Note: Functions are regular HOL functions, theorems have measurability assumptions. We use do {. . .} notation

3

slide-18
SLIDE 18

Markov Kernels (a.k.a. stochastic relations)

Transition functions for Markov chains on state spaces S traditional T : S → S → R

  • nonnegative & rows sum up to 1

coalgebraic T S S

  • S are the discrete distributions on S

generalized K

m

  • is a measurable space on S,
  • are probability distributions on

4

slide-19
SLIDE 19

Markov Kernels (a.k.a. stochastic relations)

Transition functions for Markov chains on state spaces S traditional T : S → S → R

  • nonnegative & rows sum up to 1

coalgebraic T : S → D(S)

  • D(S) are the discrete distributions on S

generalized K

m

  • is a measurable space on S,
  • are probability distributions on

4

slide-20
SLIDE 20

Markov Kernels (a.k.a. stochastic relations)

Transition functions for Markov chains on state spaces S traditional T : S → S → R

  • nonnegative & rows sum up to 1

coalgebraic T : S → D(S)

  • D(S) are the discrete distributions on S

generalized K ∈ S →m Pr(S)

  • S is a measurable space on S,
  • Pr(S) are probability distributions on S

4

slide-21
SLIDE 21

Extension Theorem by Ionescu-Tulcea

There exists proc x ∈ S →m Pr(stream(S)) where do { y1 ← K x y2 ← K y1 do { y3 ← K y2 y ← K x proc x = y4 ← K y3 = ω ← proc y y5 ← K y4 return (y·ω) y6 ← K y5 } . . . return (y1·y2·y3·y4·y5· · ·) }

5

slide-22
SLIDE 22

Uniqueness

Ionescu-Tulcea proves existence. Is it unique?

Yes: Bisimulation implies equality

Bisimulation relation R stream stream R M N exists K , and M N

m

stream s.t.

  • M

do y K M y return y ,

  • N

do y K N y return y , and

  • K y R M y

N y 1. Bisimulation implies equality (a.k.a coinduction rule for equality) R bisimulation relation: R M N M N.

6

slide-23
SLIDE 23

Uniqueness

Ionescu-Tulcea proves existence. Is it unique?

Yes: Bisimulation implies equality

Bisimulation relation R stream stream R M N exists K , and M N

m

stream s.t.

  • M

do y K M y return y ,

  • N

do y K N y return y , and

  • K y R M y

N y 1. Bisimulation implies equality (a.k.a coinduction rule for equality) R bisimulation relation: R M N M N.

6

slide-24
SLIDE 24

Uniqueness

Ionescu-Tulcea proves existence. Is it unique?

Yes: Bisimulation implies equality

Bisimulation relation R : Pr(stream(S)) → Pr(stream(S)) → B R M N exists K ∈ Pr(S), and M′, N′ ∈ S →m Pr(stream(S)) s.t.

  • M = do {y ← K; ω ← M′ y; return (y·ω)},
  • N = do {y ← K; ω ← N′ y; return (y·ω)}, and
  • PrK {y | R (M′ y) (N′ y)} = 1.

Bisimulation implies equality (a.k.a coinduction rule for equality) R bisimulation relation: R M N M N.

6

slide-25
SLIDE 25

Uniqueness

Ionescu-Tulcea proves existence. Is it unique?

Yes: Bisimulation implies equality

Bisimulation relation R : Pr(stream(S)) → Pr(stream(S)) → B R M N exists K ∈ Pr(S), and M′, N′ ∈ S →m Pr(stream(S)) s.t.

  • M = do {y ← K; ω ← M′ y; return (y·ω)},
  • N = do {y ← K; ω ← N′ y; return (y·ω)}, and
  • PrK {y | R (M′ y) (N′ y)} = 1.

Bisimulation implies equality (a.k.a coinduction rule for equality) R bisimulation relation: R M N = ⇒ M = N.

6

slide-26
SLIDE 26

Markov Property

a b c

0.5 0.5 0.33 0.67 0.5 0.5

time a a b c a

7

slide-27
SLIDE 27

Markov Property

a b c

0.5 0.5 0.33 0.67 0.5 0.5

time a a b c a

7

slide-28
SLIDE 28

Markov Property

a b c

0.5 0.5 0.33 0.67 0.5 0.5

time a a b c a

7

slide-29
SLIDE 29

Markov Property

a b c

0.5 0.5 0.33 0.67 0.5 0.5

time a a b c a

7

slide-30
SLIDE 30

Markov Property

a b c

0.5 0.5 0.33 0.67 0.5 0.5

time a a b c a

7

slide-31
SLIDE 31

Markov Property

a b c

0.5 0.5 0.33 0.67 0.5 0.5

time a a b c a

7

slide-32
SLIDE 32

Markov Property

a b c

0.5 0.5 0.33 0.67 0.5 0.5

time a a b c ∑ = a

7

slide-33
SLIDE 33

Markov Property

a b c

0.5 0.5 0.33 0.67 0.5 0.5

time a a b c ∑ = a

7

slide-34
SLIDE 34

Markov Property

a b c

0.5 0.5 0.33 0.67 0.5 0.5

time a a b c ∑ = a

7

slide-35
SLIDE 35

Markov Property

a b c

0.5 0.5 0.33 0.67 0.5 0.5

time a a b c ∑ = a

7

slide-36
SLIDE 36

Markov Property

a b c

0.5 0.5 0.33 0.67 0.5 0.5

time a a b c ∑ = a = =

7

slide-37
SLIDE 37

Markov Property

Lemma markov_prop: proc x = do { ω ← proc x ω′ ← proc ω[n] return (take n ω)·ω′ } Proof: Induction on n, or bisimulation. Can we generalize n?

8

slide-38
SLIDE 38

Markov Property

Lemma markov_prop: proc x = do { ω ← proc x ω′ ← proc ω[n] return (take n ω)·ω′ } Proof: Induction on n, or bisimulation. Can we generalize n?

8

slide-39
SLIDE 39

Strong Markov Property

Stopping time t ∈ stream(S) →m N ∪ {∞} If t ω = n, then take n ω′ = take n ω = ⇒ t ω′ = n. Examples

  • First time the state x is in
  • Natural number in state

0,

  • 3rd occurence of a failure
  • max, min compositions of stopping times

Not a stopping time The last occurence of a state is not a stopping time. n time  t n t n

9

slide-40
SLIDE 40

Strong Markov Property

Stopping time t ∈ stream(S) →m N ∪ {∞} If t ω = n, then take n ω′ = take n ω = ⇒ t ω′ = n. Examples

  • First time the state x is in
  • Natural number in state

0,

  • 3rd occurence of a failure
  • max, min compositions of stopping times

Not a stopping time The last occurence of a state is not a stopping time. n time ω  t n t n

9

slide-41
SLIDE 41

Strong Markov Property

Stopping time t ∈ stream(S) →m N ∪ {∞} If t ω = n, then take n ω′ = take n ω = ⇒ t ω′ = n. Examples

  • First time the state x is in
  • Natural number in state

0,

  • 3rd occurence of a failure
  • max, min compositions of stopping times

Not a stopping time The last occurence of a state is not a stopping time. n time ω  t ω = n t n

9

slide-42
SLIDE 42

Strong Markov Property

Stopping time t ∈ stream(S) →m N ∪ {∞} If t ω = n, then take n ω′ = take n ω = ⇒ t ω′ = n. Examples

  • First time the state x is in
  • Natural number in state

0,

  • 3rd occurence of a failure
  • max, min compositions of stopping times

Not a stopping time The last occurence of a state is not a stopping time. n time ω  t ω = n ω′ t n

9

slide-43
SLIDE 43

Strong Markov Property

Stopping time t ∈ stream(S) →m N ∪ {∞} If t ω = n, then take n ω′ = take n ω = ⇒ t ω′ = n. Examples

  • First time the state x is in
  • Natural number in state

0,

  • 3rd occurence of a failure
  • max, min compositions of stopping times

Not a stopping time The last occurence of a state is not a stopping time. n time ω  t ω = n ω′ t ω′ = n

9

slide-44
SLIDE 44

Strong Markov Property

Stopping time t ∈ stream(S) →m N ∪ {∞} If t ω = n, then take n ω′ = take n ω = ⇒ t ω′ = n. Examples

  • First time the state x is in ω
  • Natural number in state ω0,
  • 3rd occurence of a failure
  • max, min compositions of stopping times

Not a stopping time The last occurence of a state is not a stopping time. n time  t n t n

9

slide-45
SLIDE 45

Strong Markov Property

Stopping time t ∈ stream(S) →m N ∪ {∞} If t ω = n, then take n ω′ = take n ω = ⇒ t ω′ = n. Examples

  • First time the state x is in ω
  • Natural number in state ω0,
  • 3rd occurence of a failure
  • max, min compositions of stopping times

Not a stopping time The last occurence of a state is not a stopping time. n time  t n t n

9

slide-46
SLIDE 46

Strong Markov Property

Lemma strong_markov_prop: Given: t is stopping time proc x = do { ω ← proc x case t ω of | ∞ ⇒ return ω | n ⇒ ω′ ← proc ω[n] return (take n ω)·ω′ } Proof: Bisimulation.

10

slide-47
SLIDE 47

Continuous-Time Markov Chain

10

slide-48
SLIDE 48

Queuing Example: Client-Server exchange

State n: active requests, c: client request rate, s: server reponse rate 1 2 3 4 · · · c s c s c s c s c s t n 1 2 3 4

11

slide-49
SLIDE 49

Queuing Example: Client-Server exchange

State n: active requests, c: client request rate, s: server reponse rate 1 2 3 4 · · · c s c s c s c s c s t n 1 2 3 4

11

slide-50
SLIDE 50

Exponential Distribution

X ∼ Exp(r) — Exponentially distributed with rate r 1 t Pr(X > t) t t t t Pr(X > t) = exp(−t · r) Exponential distribution is memoryless t t X t X t X t t

12

slide-51
SLIDE 51

Exponential Distribution

X ∼ Exp(r) — Exponentially distributed with rate r 1 t Pr(X > t) t t t t Pr(X > t) = exp(−t · r) Exponential distribution is memoryless t′ ⩾ t = ⇒ Pr(X > t′ | X > t) = Pr(X > t′ − t)

12

slide-52
SLIDE 52

Exponential Distribution

X ∼ Exp(r) — Exponentially distributed with rate r 1 t Pr(X > t) t t t t Pr(X > t) = exp(−t · r) Exponential distribution is memoryless t′ ⩾ t = ⇒ Pr(X > t′ | X > t) = Pr(X > t′ − t)

12

slide-53
SLIDE 53

Exponential Distribution

X ∼ Exp(r) — Exponentially distributed with rate r 1 t Pr(X > t) t t′ t t Pr(X > t) = exp(−t · r) Exponential distribution is memoryless t′ ⩾ t = ⇒ Pr(X > t′ | X > t) = Pr(X > t′ − t)

12

slide-54
SLIDE 54

Exponential Distribution

X ∼ Exp(r) — Exponentially distributed with rate r 1 t Pr(X > t) t t′ t′ − t Pr(X > t) = exp(−t · r) Exponential distribution is memoryless t′ ⩾ t = ⇒ Pr(X > t′ | X > t) = Pr(X > t′ − t)

12

slide-55
SLIDE 55

Parallel Choices

s s1 s2 s3 r1 r2 r3 time t J S J i

ri

i ri

Exp

i ri

J do t

iExp ri

i THE i j ti tj return ti i

13

slide-56
SLIDE 56

Parallel Choices

s s1 s2 s3 r1 r2 r3 time t J S J i

ri

i ri

Exp

i ri

J do t

iExp ri

i THE i j ti tj return ti i

13

slide-57
SLIDE 57

Parallel Choices

s s1 s2 s3 r1 r2 r3 time t J S J i

ri

i ri

Exp

i ri

J do t

iExp ri

i THE i j ti tj return ti i

13

slide-58
SLIDE 58

Parallel Choices

s s1 s2 s3 r1 r2 r3 time t J S J i

ri

i ri

Exp

i ri

J do t

iExp ri

i THE i j ti tj return ti i

13

slide-59
SLIDE 59

Parallel Choices

s s1 s2 s3 r1 r2 r3 time t J ∈ D(S) J {i} :=

ri ∑

i ri

Exp(∑

i ri) × J

= do { t ← ΠiExp(ri) i := THE i. ∀j. ti < tj return (ti, i) }

13

slide-60
SLIDE 60

Kernel for CTMCs

Transition rates R : S → S → R The rate to go from state x to state y is R x y. Nonnegative R x y ⩾ 0 Finite and Positive 0 < ∑

y R x y < ∞

Zero Diagonal R x x = 0 Markov Kernel K K maps current state and jump time to next state and jump time. K S

m

S K t x map t Exp

y R x y

Jx CTMC for rate R ctmc S

m

stream S ctmc t x procK t x

14

slide-61
SLIDE 61

Kernel for CTMCs

Transition rates R : S → S → R The rate to go from state x to state y is R x y. Nonnegative R x y ⩾ 0 Finite and Positive 0 < ∑

y R x y < ∞

Zero Diagonal R x x = 0 Markov Kernel K K maps current state and jump time to next state and jump time. K ∈ R × S →m Pr(R × S) K (t, x) := ( map (+t) Exp(∑

y R x y)

) × Jx CTMC for rate R ctmc S

m

stream S ctmc t x procK t x

14

slide-62
SLIDE 62

Kernel for CTMCs

Transition rates R : S → S → R The rate to go from state x to state y is R x y. Nonnegative R x y ⩾ 0 Finite and Positive 0 < ∑

y R x y < ∞

Zero Diagonal R x x = 0 Markov Kernel K K maps current state and jump time to next state and jump time. K ∈ R × S →m Pr(R × S) K (t, x) := ( map (+t) Exp(∑

y R x y)

) × Jx CTMC for rate R ctmc ∈ R × S →m Pr(stream (R × S)) ctmc (t, x) := procK (t, x)

14

slide-63
SLIDE 63

Construct CTMCs

Markov Property ctmc (t, x) = do { ω ← ctmc (t, x) t ⩽ t′ = ⇒ ω′ ← ctmc (t′, state_at ω t′) return (merge ω t′ ω′) } t 1 t 1

15

slide-64
SLIDE 64

Construct CTMCs

Markov Property ctmc (t, x) = do { ω ← ctmc (t, x) t ⩽ t′ = ⇒ ω′ ← ctmc (t′, state_at ω t′) return (merge ω t′ ω′) } t 1 ω t 1

15

slide-65
SLIDE 65

Construct CTMCs

Markov Property ctmc (t, x) = do { ω ← ctmc (t, x) t ⩽ t′ = ⇒ ω′ ← ctmc (t′, state_at ω t′) return (merge ω t′ ω′) } t 1 ω t′ 1

15

slide-66
SLIDE 66

Construct CTMCs

Markov Property ctmc (t, x) = do { ω ← ctmc (t, x) t ⩽ t′ = ⇒ ω′ ← ctmc (t′, state_at ω t′) return (merge ω t′ ω′) } t 1 ω t′ 1 ω′

15

slide-67
SLIDE 67

Properties of CTMCs

Transition probability p p x y t := Pr(CTMC started in x is in y at time t) Chapman–Kolmogorov equation t1 t2 p x y t1 t2

x

p x x t1 p x y t2 p is the solution of a differential equation t p x y t

x

R x x p x y t p x y t

16

slide-68
SLIDE 68

Properties of CTMCs

Transition probability p p x y t := Pr(CTMC started in x is in y at time t) Chapman–Kolmogorov equation t1, t2 ⩾ 0 = ⇒ p x y (t1 + t2) = ∑

x′

p x x′ t1 · p x′ y t2 p is the solution of a differential equation t p x y t

x

R x x p x y t p x y t

16

slide-69
SLIDE 69

Properties of CTMCs

Transition probability p p x y t := Pr(CTMC started in x is in y at time t) Chapman–Kolmogorov equation t1, t2 ⩾ 0 = ⇒ p x y (t1 + t2) = ∑

x′

p x x′ t1 · p x′ y t2 p is the solution of a differential equation t′ > 0 = ⇒ p′ x y t = ∑

x′

R x x′ · (p x′ y t − p x y t)

16

slide-70
SLIDE 70

Discussion

16

slide-71
SLIDE 71

Difference to Traditional Probability Theory

Traditional statement for the Markov property: Pr(A | ∀t′ ⩽ t. Xt′ = xt′) = Pr(A | Xt = xt).

  • Restricted to probability of set A

Not obvious to derive rule for integral and with probability 1

  • Requires to massage A and xt′ into right form

Advantage of working on the measure level

  • Deriving rules for probability, integral, and with probability 1:

Equations for monadic operations

  • Nice algebraic rewrite rules for monad operations
  • Most statements expressible as equations on measures
  • Allows bisimulation as a powerful proof method

17

slide-72
SLIDE 72

Difference to Traditional Probability Theory

Traditional statement for the Markov property: Pr(A | ∀t′ ⩽ t. Xt′ = xt′) = Pr(A | Xt = xt).

  • Restricted to probability of set A

Not obvious to derive rule for integral and with probability 1

  • Requires to massage A and xt′ into right form

Advantage of working on the measure level

  • Deriving rules for probability, integral, and with probability 1:

Equations for monadic operations

  • Nice algebraic rewrite rules for monad operations
  • Most statements expressible as equations on measures
  • Allows bisimulation as a powerful proof method

17

slide-73
SLIDE 73

Difference to Traditional Probability Theory

Traditional statement for the Markov property: Pr(A | ∀t′ ⩽ t. Xt′ = xt′) = Pr(A | Xt = xt).

  • Restricted to probability of set A

Not obvious to derive rule for integral and with probability 1

  • Requires to massage A and xt′ into right form

Advantage of working on the measure level

  • Deriving rules for probability, integral, and with probability 1:

Equations for monadic operations

  • Nice algebraic rewrite rules for monad operations
  • Most statements expressible as equations on measures
  • Allows bisimulation as a powerful proof method

17

slide-74
SLIDE 74

Difference to Traditional Probability Theory

Traditional statement for the Markov property: Pr(A | ∀t′ ⩽ t. Xt′ = xt′) = Pr(A | Xt = xt).

  • Restricted to probability of set A

Not obvious to derive rule for integral and with probability 1

  • Requires to massage A and xt′ into right form

Advantage of working on the measure level

  • Deriving rules for probability, integral, and with probability 1:

Equations for monadic operations

  • Nice algebraic rewrite rules for monad operations
  • Most statements expressible as equations on measures
  • Allows bisimulation as a powerful proof method

17

slide-75
SLIDE 75

Difference to Traditional Probability Theory

Traditional statement for the Markov property: Pr(A | ∀t′ ⩽ t. Xt′ = xt′) = Pr(A | Xt = xt).

  • Restricted to probability of set A

Not obvious to derive rule for integral and with probability 1

  • Requires to massage A and xt′ into right form

Advantage of working on the measure level

  • Deriving rules for probability, integral, and with probability 1:

Equations for monadic operations

  • Nice algebraic rewrite rules for monad operations
  • Most statements expressible as equations on measures
  • Allows bisimulation as a powerful proof method

17

slide-76
SLIDE 76

Related Work

  • Giry monad

(discrete: Audebaud and Paulin-Mohring [MPC 2006], and more)

  • Daniel-Kolmogorov (stronger than I-T, restricted to Borel spaces)

(Immler, Masters thesis 2012)

  • On arbitrary measurable spaces

(Eberl, Hölzl, and Nipkow [ESOP 2015])

  • Markov Kernels in Isabelle/HOL:

(Backes, Berg, and Unruh [LPAR 2008]) but missing proofs Newly developed

  • Theory of Markov Kernels & Processes
  • incl. Extension Theorem by Ionescu-Tulcea
  • Equality by Bisimulation on Measures, and
  • CTMCs

18

slide-77
SLIDE 77

Related Work

  • Giry monad

(discrete: Audebaud and Paulin-Mohring [MPC 2006], and more)

  • Daniel-Kolmogorov (stronger than I-T, restricted to Borel spaces)

(Immler, Masters thesis 2012)

  • On arbitrary measurable spaces

(Eberl, Hölzl, and Nipkow [ESOP 2015])

  • Markov Kernels in Isabelle/HOL:

(Backes, Berg, and Unruh [LPAR 2008]) but missing proofs Newly developed

  • Theory of Markov Kernels & Processes
  • incl. Extension Theorem by Ionescu-Tulcea
  • Equality by Bisimulation on Measures, and
  • CTMCs

18

slide-78
SLIDE 78

Related Work

  • Giry monad

(discrete: Audebaud and Paulin-Mohring [MPC 2006], and more)

  • Daniel-Kolmogorov (stronger than I-T, restricted to Borel spaces)

(Immler, Masters thesis 2012)

  • On arbitrary measurable spaces

(Eberl, Hölzl, and Nipkow [ESOP 2015])

  • Markov Kernels in Isabelle/HOL:

(Backes, Berg, and Unruh [LPAR 2008]) but missing proofs Newly developed

  • Theory of Markov Kernels & Processes
  • incl. Extension Theorem by Ionescu-Tulcea
  • Equality by Bisimulation on Measures, and
  • CTMCs

18

slide-79
SLIDE 79

Related Work

  • Giry monad

(discrete: Audebaud and Paulin-Mohring [MPC 2006], and more)

  • Daniel-Kolmogorov (stronger than I-T, restricted to Borel spaces)

(Immler, Masters thesis 2012)

  • On arbitrary measurable spaces

(Eberl, Hölzl, and Nipkow [ESOP 2015])

  • Markov Kernels in Isabelle/HOL:

(Backes, Berg, and Unruh [LPAR 2008]) but missing proofs Newly developed

  • Theory of Markov Kernels & Processes
  • incl. Extension Theorem by Ionescu-Tulcea
  • Equality by Bisimulation on Measures, and
  • CTMCs

18

slide-80
SLIDE 80

Related Work

  • Giry monad

(discrete: Audebaud and Paulin-Mohring [MPC 2006], and more)

  • Daniel-Kolmogorov (stronger than I-T, restricted to Borel spaces)

(Immler, Masters thesis 2012)

  • On arbitrary measurable spaces

(Eberl, Hölzl, and Nipkow [ESOP 2015])

  • Markov Kernels in Isabelle/HOL:

(Backes, Berg, and Unruh [LPAR 2008]) but missing proofs Newly developed

  • Theory of Markov Kernels & Processes
  • incl. Extension Theorem by Ionescu-Tulcea
  • Equality by Bisimulation on Measures, and
  • CTMCs

18

slide-81
SLIDE 81

Related Work

  • Giry monad

(discrete: Audebaud and Paulin-Mohring [MPC 2006], and more)

  • Daniel-Kolmogorov (stronger than I-T, restricted to Borel spaces)

(Immler, Masters thesis 2012)

  • On arbitrary measurable spaces

(Eberl, Hölzl, and Nipkow [ESOP 2015])

  • Markov Kernels in Isabelle/HOL:

(Backes, Berg, and Unruh [LPAR 2008]) but missing proofs Newly developed

  • Theory of Markov Kernels & Processes
  • incl. Extension Theorem by Ionescu-Tulcea
  • Equality by Bisimulation on Measures, and
  • CTMCs

18

slide-82
SLIDE 82

Related Work

  • Giry monad

(discrete: Audebaud and Paulin-Mohring [MPC 2006], and more)

  • Daniel-Kolmogorov (stronger than I-T, restricted to Borel spaces)

(Immler, Masters thesis 2012)

  • On arbitrary measurable spaces

(Eberl, Hölzl, and Nipkow [ESOP 2015])

  • Markov Kernels in Isabelle/HOL:

(Backes, Berg, and Unruh [LPAR 2008]) but missing proofs Newly developed

  • Theory of Markov Kernels & Processes
  • incl. Extension Theorem by Ionescu-Tulcea
  • Equality by Bisimulation on Measures, and
  • CTMCs

18

slide-83
SLIDE 83

Related Work

  • Giry monad

(discrete: Audebaud and Paulin-Mohring [MPC 2006], and more)

  • Daniel-Kolmogorov (stronger than I-T, restricted to Borel spaces)

(Immler, Masters thesis 2012)

  • On arbitrary measurable spaces

(Eberl, Hölzl, and Nipkow [ESOP 2015])

  • Markov Kernels in Isabelle/HOL:

(Backes, Berg, and Unruh [LPAR 2008]) but missing proofs Newly developed

  • Theory of Markov Kernels & Processes
  • incl. Extension Theorem by Ionescu-Tulcea
  • Equality by Bisimulation on Measures, and
  • CTMCs

18

slide-84
SLIDE 84

Related Work

  • Giry monad

(discrete: Audebaud and Paulin-Mohring [MPC 2006], and more)

  • Daniel-Kolmogorov (stronger than I-T, restricted to Borel spaces)

(Immler, Masters thesis 2012)

  • On arbitrary measurable spaces

(Eberl, Hölzl, and Nipkow [ESOP 2015])

  • Markov Kernels in Isabelle/HOL:

(Backes, Berg, and Unruh [LPAR 2008]) but missing proofs Newly developed

  • Theory of Markov Kernels & Processes
  • incl. Extension Theorem by Ionescu-Tulcea
  • Equality by Bisimulation on Measures, and
  • CTMCs

18

slide-85
SLIDE 85

Questions?

18