Compositionality for Markov Reward Chains with Fast Transitions J. - - PowerPoint PPT Presentation

compositionality for markov reward chains with fast
SMART_READER_LITE
LIVE PREVIEW

Compositionality for Markov Reward Chains with Fast Transitions J. - - PowerPoint PPT Presentation

Compositionality for Markov Reward Chains with Fast Transitions J. Markovski, A. Sokolova, N. Tr cka, E. P. de Vink presented by Elias Pschernig January 24, 2008 Outline Introduction Motivation Recapitulation: Markov Chains Aggregation


slide-1
SLIDE 1

Compositionality for Markov Reward Chains with Fast Transitions

  • J. Markovski, A. Sokolova, N. Trˇ

cka, E. P. de Vink presented by Elias Pschernig January 24, 2008

slide-2
SLIDE 2

Outline

Introduction Motivation Recapitulation: Markov Chains Aggregation methods Discontinuous Markov reward chains

Ordinary lumping Reduction

Markov reward chains with fast transitions

τ-lumping τ-reduction

Relational properties Parallel composition

slide-3
SLIDE 3

Markov Reward Chains

◮ Among most important and wide-spread analytical

performance models

◮ Ever growing complexity of Markov reward chain systems ◮ Compositional generation: Composing a big system from

several small components

◮ State space explosion: Result size is product of sizes of

components

◮ Need aggregation methods... ◮ ...and they should be compositional ◮ We consider special models of Markov reward chains:

Discontinuous Markov reward chains and Markov reward chains with fast transitions

slide-4
SLIDE 4

Markov Reward Chains

◮ Among most important and wide-spread analytical

performance models

◮ Ever growing complexity of Markov reward chain systems ◮ Compositional generation: Composing a big system from

several small components

◮ State space explosion: Result size is product of sizes of

components

◮ Need aggregation methods... ◮ ...and they should be compositional ◮ We consider special models of Markov reward chains:

Discontinuous Markov reward chains and Markov reward chains with fast transitions

slide-5
SLIDE 5

Recapitulation: Discrete time Markov chains

  • 1

0.3

  • 0.3
  • 0.4
  • 2

0.1

  • 0.9
  • 3

0.9

  • 0.1
  • Transition probability ma-

trix: 1 2 3 1 0.4 0.3 0.3 2 0.1 0.9 3 0.9 0.1

◮ Graphs with nodes

representing states

◮ Outgoing arrows

determine stochastic behavior of each state

◮ Probabilities only

depend on current state

slide-6
SLIDE 6

Continuous time Markov chains

  • 1

2

  • 1
  • 2

2

  • 3

1

  • ◮ Generator matrix:

1 2 3 1

  • 3

1 2 2

  • 2

2 3 1

  • 1

=Q

slide-7
SLIDE 7

Continuous time Markov reward chains

  • 1

0.8 r0 λ

  • τ
  • 2

0.1 r1 µ

  • 3

0.1 r2 ν

  • ◮ P = (σ, Q, ρ)

◮ σ is a stochastic row initial probability vector (0.8, 0.1, 0.1) ◮ ρ is a state reward vector (r0, r1, r2) ◮ Transition probability matrix

P(t) =

  • n=0

Qntn n! = eQt

◮ Rewards are used to measure performance (application

dependent).

slide-8
SLIDE 8

Discontinuous Markov reward chains

◮ Markov chains with instantaneous transitions → discontinuous

Markov chains

◮ Discontinuous Markov reward chain: P = (σ, Π, Q, ρ) ◮ Intuition: Π[i, j] denotes probability that a process occupies

two states via an instantaneous transition.

◮ Π = I leads to a standard Markov chain → generalization

  • 1
  • 2
  • 3
  • 4
slide-9
SLIDE 9

Discontinuous Markov reward chains

Aggregation for discontinuous Markov reward chains

◮ Ordinary lumping ◮ Reduction

slide-10
SLIDE 10

Ordinary lumping

◮ We lump P = (σ, Π, Q, ρ) to P = (σ, Π, Q, ρ) ◮ Partition L is an ordinary lumping ◮ P L

→ P

slide-11
SLIDE 11

Ordinary lumping

◮ P L

→ P

◮ Partition of the state space into classes ◮ States lumped together form a class ◮ Equivalent transition behavior to other classes (intuitively:

probability of class is sum of probabilities of states)

◮ All states in a class have the same reward, total reward is

preserved

slide-12
SLIDE 12

Example

◮ P L

→ P

  • 1

0.5 r1 λ

  • µ
  • 2

0.2 r2

1 2 λ

  • 1

2 λ

  • 3

0.3 r2 λ

  • 4

r3 ρ

  • 5

r3 ρ

  • {{1},{2,3},{4,5}}
  • 1

0.5 r1 λ+µ

  • 2, 3

0.5 r2 λ

  • 4, 5

r3 ρ

slide-13
SLIDE 13

Reduction

◮ We reduce P = (σ, Π, Q, ρ) to P = (σ, I, Q, ρ) ◮ P →r P ◮ Result is unique up to state permutation. ◮ Canonical product decomposition of Π ◮ Reduced states are given by ergodic classes of the original

process (ergodic = each state can be reached from each other state in finite time)

◮ Total reward is preserved

slide-14
SLIDE 14

Markov reward chains with fast transitions

Markov reward chains with fast transitions

◮ Definition ◮ Aggregation

slide-15
SLIDE 15

Markov reward chains with fast transitions

◮ Adds parameterized (“fast”) transitions to a standard Markov

reward chain.

◮ Uses two generator matrixes Qs and Qf , for slow and fast

transitions.

◮ P = (σ, Qs, Qf , ρ) is a function... ◮ ...where to each τ > 0 a Markov reward chain

Pτ = (σ, I, Qs + τQf , ρ) is assigned

◮ The limit τ → ∞ makes fast transitions instantaneous, and

we end up with a discontinuous Markov reward chain.

slide-16
SLIDE 16

Markov reward chains with fast transitions

Aggregation for Markov reward chains with fast transitions

◮ τ-lumping ◮ τ-reduction

slide-17
SLIDE 17

τ-lumping

◮ We τ-lump P = (σ, Qs, Qf , ρ) to P = (σ, Qs, Qf , ρ) ◮ Can define it using the limiting discontinuous Markov reward

chain.

◮ P L

P

◮ Not unique

slide-18
SLIDE 18

τ-lumping

P (fast transitions)

  • L
  • P

(lumped fast transitions)

  • Q

(discontinuous)

L

  • Q

(lumped discontinuous)

slide-19
SLIDE 19

τ-reduction

◮ We τ-reduce P = (σ, Qs, Qf , ρ) to R = (σ, I, Q, ρ) ◮ P r R

slide-20
SLIDE 20

Example

◮ P r R

  • 1

λ

  • 2

  • 3

µ

  • 4

ρ

  • 5

τ-reduction

  • 1

a a+b λ

  • b

a+b λ

  • 2, 3

µ

  • 2, 4

ρ

  • 5
slide-21
SLIDE 21

τ-reduction

P (fast transitions)

  • r
  • Q

(discontinuous)

r

  • R = R′

(continuous)

◮ if P r R ◮ and P →∞ Q →r R′ ◮ then R = R′

slide-22
SLIDE 22

Relational properties of ordinary lumping and τ-lumping

◮ Reduction works in one step, so no need to look at details of

its relational properties. Lumping:

◮ Need transitivity and strong confluence... ◮ ...to ensure that iterative application yields a uniquely

determined process.

◮ Repeated application of ordinary lumping... ◮ ...can be replaced by single application of composition of

individual lumpings.

◮ For τ-lumping, only the limit is uniquely determined.

slide-23
SLIDE 23

Relational properties of ordinary lumping and τ-lumping

◮ Reduction works in one step, so no need to look at details of

its relational properties. Lumping:

◮ Need transitivity and strong confluence... ◮ ...to ensure that iterative application yields a uniquely

determined process.

◮ Repeated application of ordinary lumping... ◮ ...can be replaced by single application of composition of

individual lumpings.

◮ For τ-lumping, only the limit is uniquely determined.

slide-24
SLIDE 24

Example

  • 1

1 r1 aτ

  • 2

r2 bτ

  • 3

r3 λ

  • 4

r4 µ

  • {{1, 2},

{3}, {4}}

  • 1, 2

1 r2 bτ

  • 3

r3 λ

  • 4

r4 µ

  • {{{1, 2}, {3}},

{{4}}}

  • 1, 2, 3

1

cr2+br3 b+c b b+c λ

  • 4

r4 µ

  • {{1, 2, 3},

{4}}

slide-25
SLIDE 25

Parallel composition

◮ P1 ≥ P1, P2 ≥ P2 =

⇒ P1 P2 ≥ P1 P2

◮ Aggregate smaller components first... ◮ ...then combine them into the aggregated complete system. ◮ ≥ is semantic preorder. ◮ P ≥ P means P is an aggregated version of P. ◮ is a parallel composition.

slide-26
SLIDE 26

Composing discontinuous Markov reward chains

◮ Kronecker sum ⊕ and Kronecker product ⊗ ◮ Parallel composition P1 P2 =

(σ1 ⊗ σ2, Π1 ⊗ Π2, Q1 ⊗ Π2 + Π1 ⊗ Q2, ρ1 ⊗ 1|ρ2| + 1|ρ1| ⊗ ρ2)

◮ If P1 and P2 are discontinuous Markov reward chains, then so

is P1 P2

slide-27
SLIDE 27

Composing discontinuous Markov reward chains

◮ Both lumping and reduction are compositional with respect to

the parallel composition of discontinuous Markov reward chains

◮ If P1 L1

→ P1 and P2

L2

→ P2, then P1 P2

L1⊗L2

→ P1 P2.

◮ If P1 →r P1 and P2 →r P2, then P1 P2 →r P1 P2

slide-28
SLIDE 28

Composing Markov reward chains with fast transitions

◮ Parallel composition

P1 P2 = (σ1⊗σ2, Qs,1⊕Qs,2, Qf ,1⊕Qf ,2, ρ1⊗1|ρ2|+1|ρ1|⊗ρ2)

◮ If P1 L1

P1 and P2

L2

P2, then P1 P2

L1⊗L2

  • P1 P2.

◮ If P1 r P1 and P2 r P2, then P1 P2 r P1 P2

slide-29
SLIDE 29

Example of parallel composition

  • 1

1 r0 aτ

  • λ
  • 2

r1 µ

  • 3

r2 ν

  • 1

π rr bτ

  • 2

1−π r4 cτ

  • ξ
  • 3

= ⇒

  • 1

  • λ
  • 4

µ

  • 7

  • ν
  • 2

  • ξ
  • λ
  • 5

µ

  • ξ
  • 8

  • ξ
  • ν
  • 3

  • λ
  • 6

µ

  • 9

ν

slide-30
SLIDE 30

Aggregated version of composition

  • 1

1 r1 µ

  • 2

r2 ν

  • 1

1 πr3+(1−π)r4

b b+c ξ

  • 2

= ⇒

  • 1

1 r1+πr3+ (1−π)r4

b b+c ξ

  • µ
  • 3

r2+πr3+ (1−π)r4

b b+c ξ

  • ν
  • 2

r1 µ

  • 4

r2 ν

slide-31
SLIDE 31

Summary

P1

  • L1
  • P1

  • Q1

L1 Q1

P2

  • L2
  • P2

  • Q2

L2 Q2

= ⇒ P1 P2

  • L1⊗L2
  • P1 P2

  • Q1 Q2

L1⊗L2 Q1 Q2

P1

  • r
  • Q1

r R1

P2

  • r
  • Q2

r R2

= ⇒ P1 P2

  • r
  • Q1 Q2

r R1 R2

slide-32
SLIDE 32

Summary

P1

  • L1
  • P1

  • Q1

L1 Q1

P2

  • L2
  • P2

  • Q2

L2 Q2

= ⇒ P1 P2

  • L1⊗L2
  • P1 P2

  • Q1 Q2

L1⊗L2 Q1 Q2

P1

  • r
  • Q1

r R1

P2

  • r
  • Q2

r R2

= ⇒ P1 P2

  • r
  • Q1 Q2

r R1 R2

slide-33
SLIDE 33

End

◮ Questions?