SLIDE 1 Compositionality for Markov Reward Chains with Fast Transitions
- J. Markovski, A. Sokolova, N. Trˇ
cka, E. P. de Vink presented by Elias Pschernig January 24, 2008
SLIDE 2
Outline
Introduction Motivation Recapitulation: Markov Chains Aggregation methods Discontinuous Markov reward chains
Ordinary lumping Reduction
Markov reward chains with fast transitions
τ-lumping τ-reduction
Relational properties Parallel composition
SLIDE 3
Markov Reward Chains
◮ Among most important and wide-spread analytical
performance models
◮ Ever growing complexity of Markov reward chain systems ◮ Compositional generation: Composing a big system from
several small components
◮ State space explosion: Result size is product of sizes of
components
◮ Need aggregation methods... ◮ ...and they should be compositional ◮ We consider special models of Markov reward chains:
Discontinuous Markov reward chains and Markov reward chains with fast transitions
SLIDE 4
Markov Reward Chains
◮ Among most important and wide-spread analytical
performance models
◮ Ever growing complexity of Markov reward chain systems ◮ Compositional generation: Composing a big system from
several small components
◮ State space explosion: Result size is product of sizes of
components
◮ Need aggregation methods... ◮ ...and they should be compositional ◮ We consider special models of Markov reward chains:
Discontinuous Markov reward chains and Markov reward chains with fast transitions
SLIDE 5 Recapitulation: Discrete time Markov chains
0.3
0.1
0.9
- 0.1
- Transition probability ma-
trix: 1 2 3 1 0.4 0.3 0.3 2 0.1 0.9 3 0.9 0.1
◮ Graphs with nodes
representing states
◮ Outgoing arrows
determine stochastic behavior of each state
◮ Probabilities only
depend on current state
SLIDE 6 Continuous time Markov chains
2
2
1
1 2 3 1
1 2 2
2 3 1
=Q
SLIDE 7 Continuous time Markov reward chains
0.8 r0 λ
0.1 r1 µ
0.1 r2 ν
◮ σ is a stochastic row initial probability vector (0.8, 0.1, 0.1) ◮ ρ is a state reward vector (r0, r1, r2) ◮ Transition probability matrix
P(t) =
∞
Qntn n! = eQt
◮ Rewards are used to measure performance (application
dependent).
SLIDE 8 Discontinuous Markov reward chains
◮ Markov chains with instantaneous transitions → discontinuous
Markov chains
◮ Discontinuous Markov reward chain: P = (σ, Π, Q, ρ) ◮ Intuition: Π[i, j] denotes probability that a process occupies
two states via an instantaneous transition.
◮ Π = I leads to a standard Markov chain → generalization
SLIDE 9
Discontinuous Markov reward chains
Aggregation for discontinuous Markov reward chains
◮ Ordinary lumping ◮ Reduction
SLIDE 10
Ordinary lumping
◮ We lump P = (σ, Π, Q, ρ) to P = (σ, Π, Q, ρ) ◮ Partition L is an ordinary lumping ◮ P L
→ P
SLIDE 11
Ordinary lumping
◮ P L
→ P
◮ Partition of the state space into classes ◮ States lumped together form a class ◮ Equivalent transition behavior to other classes (intuitively:
probability of class is sum of probabilities of states)
◮ All states in a class have the same reward, total reward is
preserved
SLIDE 12 Example
◮ P L
→ P
0.5 r1 λ
0.2 r2
1 2 λ
2 λ
0.3 r2 λ
r3 ρ
r3 ρ
0.5 r1 λ+µ
0.5 r2 λ
r3 ρ
SLIDE 13
Reduction
◮ We reduce P = (σ, Π, Q, ρ) to P = (σ, I, Q, ρ) ◮ P →r P ◮ Result is unique up to state permutation. ◮ Canonical product decomposition of Π ◮ Reduced states are given by ergodic classes of the original
process (ergodic = each state can be reached from each other state in finite time)
◮ Total reward is preserved
SLIDE 14
Markov reward chains with fast transitions
Markov reward chains with fast transitions
◮ Definition ◮ Aggregation
SLIDE 15
Markov reward chains with fast transitions
◮ Adds parameterized (“fast”) transitions to a standard Markov
reward chain.
◮ Uses two generator matrixes Qs and Qf , for slow and fast
transitions.
◮ P = (σ, Qs, Qf , ρ) is a function... ◮ ...where to each τ > 0 a Markov reward chain
Pτ = (σ, I, Qs + τQf , ρ) is assigned
◮ The limit τ → ∞ makes fast transitions instantaneous, and
we end up with a discontinuous Markov reward chain.
SLIDE 16
Markov reward chains with fast transitions
Aggregation for Markov reward chains with fast transitions
◮ τ-lumping ◮ τ-reduction
SLIDE 17
τ-lumping
◮ We τ-lump P = (σ, Qs, Qf , ρ) to P = (σ, Qs, Qf , ρ) ◮ Can define it using the limiting discontinuous Markov reward
chain.
◮ P L
P
◮ Not unique
SLIDE 18 τ-lumping
P (fast transitions)
∞
(lumped fast transitions)
∞
(discontinuous)
L
(lumped discontinuous)
SLIDE 19
τ-reduction
◮ We τ-reduce P = (σ, Qs, Qf , ρ) to R = (σ, I, Q, ρ) ◮ P r R
SLIDE 20 Example
◮ P r R
λ
aτ
µ
ρ
τ-reduction
a a+b λ
a+b λ
µ
ρ
SLIDE 21 τ-reduction
P (fast transitions)
∞
(discontinuous)
r
(continuous)
◮ if P r R ◮ and P →∞ Q →r R′ ◮ then R = R′
SLIDE 22
Relational properties of ordinary lumping and τ-lumping
◮ Reduction works in one step, so no need to look at details of
its relational properties. Lumping:
◮ Need transitivity and strong confluence... ◮ ...to ensure that iterative application yields a uniquely
determined process.
◮ Repeated application of ordinary lumping... ◮ ...can be replaced by single application of composition of
individual lumpings.
◮ For τ-lumping, only the limit is uniquely determined.
SLIDE 23
Relational properties of ordinary lumping and τ-lumping
◮ Reduction works in one step, so no need to look at details of
its relational properties. Lumping:
◮ Need transitivity and strong confluence... ◮ ...to ensure that iterative application yields a uniquely
determined process.
◮ Repeated application of ordinary lumping... ◮ ...can be replaced by single application of composition of
individual lumpings.
◮ For τ-lumping, only the limit is uniquely determined.
SLIDE 24 Example
1 r1 aτ
r2 bτ
r3 λ
r4 µ
{3}, {4}}
1 r2 bτ
r3 λ
r4 µ
{{4}}}
1
cr2+br3 b+c b b+c λ
r4 µ
{4}}
SLIDE 25
Parallel composition
◮ P1 ≥ P1, P2 ≥ P2 =
⇒ P1 P2 ≥ P1 P2
◮ Aggregate smaller components first... ◮ ...then combine them into the aggregated complete system. ◮ ≥ is semantic preorder. ◮ P ≥ P means P is an aggregated version of P. ◮ is a parallel composition.
SLIDE 26
Composing discontinuous Markov reward chains
◮ Kronecker sum ⊕ and Kronecker product ⊗ ◮ Parallel composition P1 P2 =
(σ1 ⊗ σ2, Π1 ⊗ Π2, Q1 ⊗ Π2 + Π1 ⊗ Q2, ρ1 ⊗ 1|ρ2| + 1|ρ1| ⊗ ρ2)
◮ If P1 and P2 are discontinuous Markov reward chains, then so
is P1 P2
SLIDE 27
Composing discontinuous Markov reward chains
◮ Both lumping and reduction are compositional with respect to
the parallel composition of discontinuous Markov reward chains
◮ If P1 L1
→ P1 and P2
L2
→ P2, then P1 P2
L1⊗L2
→ P1 P2.
◮ If P1 →r P1 and P2 →r P2, then P1 P2 →r P1 P2
SLIDE 28 Composing Markov reward chains with fast transitions
◮ Parallel composition
P1 P2 = (σ1⊗σ2, Qs,1⊕Qs,2, Qf ,1⊕Qf ,2, ρ1⊗1|ρ2|+1|ρ1|⊗ρ2)
◮ If P1 L1
P1 and P2
L2
P2, then P1 P2
L1⊗L2
◮ If P1 r P1 and P2 r P2, then P1 P2 r P1 P2
SLIDE 29 Example of parallel composition
1 r0 aτ
r1 µ
r2 ν
π rr bτ
1−π r4 cτ
= ⇒
aτ
µ
bτ
cτ
µ
cτ
aτ
µ
ν
SLIDE 30 Aggregated version of composition
1 r1 µ
r2 ν
1 πr3+(1−π)r4
b b+c ξ
= ⇒
1 r1+πr3+ (1−π)r4
b b+c ξ
r2+πr3+ (1−π)r4
b b+c ξ
r1 µ
r2 ν
SLIDE 31 Summary
P1
∞
∞
L1 Q1
P2
∞
∞
L2 Q2
= ⇒ P1 P2
∞
∞
L1⊗L2 Q1 Q2
P1
∞
r R1
P2
∞
r R2
= ⇒ P1 P2
∞
r R1 R2
SLIDE 32 Summary
P1
∞
∞
L1 Q1
P2
∞
∞
L2 Q2
= ⇒ P1 P2
∞
∞
L1⊗L2 Q1 Q2
P1
∞
r R1
P2
∞
r R2
= ⇒ P1 P2
∞
r R1 R2
SLIDE 33
End
◮ Questions?