Minimizing Markov chains Beyond Bisimilarity* Giovanni Bacci, - - PowerPoint PPT Presentation

minimizing markov chains beyond bisimilarity
SMART_READER_LITE
LIVE PREVIEW

Minimizing Markov chains Beyond Bisimilarity* Giovanni Bacci, - - PowerPoint PPT Presentation

Minimizing Markov chains Beyond Bisimilarity* Giovanni Bacci, Giorgio Bacci, Kim G. Larsen , Radu Mardare Aalborg University, Denmark 22 April 2017 - Uppsala, Sweden SynCoP + PV 2017 (*) On the Metric-based Approximate Minimization of Markov


slide-1
SLIDE 1

Minimizing Markov chains Beyond Bisimilarity*

Giovanni Bacci, Giorgio Bacci, Kim G. Larsen, Radu Mardare Aalborg University, Denmark

22 April 2017 - Uppsala, Sweden

SynCoP + PV 2017

(*) On the Metric-based Approximate Minimization of Markov Chains - accepted for ICALP 2017

slide-2
SLIDE 2

1 1

m1 m3

1

m4 m2 m5 m0

1/2 1/2 1/2 1/2 1/2 1/6 1/3

MC(5)

Best Approximant & Parameter Synthesis

slide-3
SLIDE 3

1 1

m1 m3

1

m4 m2 m5 m0

1/2 1/2 1/2 1/2 1/2 1/6 1/3

1 1

m12

m3

1

m4 m5 m0

? ? ? ? ?

MC(5)

Best Approximant & Parameter Synthesis

slide-4
SLIDE 4

1 1

m1 m3

1

m4 m2 m5 m0

1/2 1/2 1/2 1/2 1/2 1/6 1/3

1 1

m12

m3

1

m4 m5 m0

? ? ? ? ? 1

m12

1

m4 m5 m0

? ? ? ? ?

m12

1

MC(5)

Best Approximant & Parameter Synthesis

slide-5
SLIDE 5

1 1

m1 m3

1

m4 m2 m5 m0

1/2 1/2 1/2 1/2 1/2 1/6 1/3

MC(5)

Best Approximant & Parameter Synthesis

1 1

m12

m3

1

m4 m5 m0

1/6 1/2 1/2 1/2 1/3

slide-6
SLIDE 6

1 1

m1 m3

1

m4 m2 m5 m0

1/2 1/2 1/2 1/2 1/2 1/6 1/3

1

m12

1

m4 m5 m0

? ? ? ? ?

m12

1

MC(5)

Best Approximant & Parameter Synthesis

1 1

m12

m3

1

m4 m5 m0

1/6 1/2 1/2 1/2 1/3

slide-7
SLIDE 7

1 1

m1 m3

1

m4 m2 m5 m0

1/2 1/2 1/2 1/2 1/2 1/6 1/3

1 1

m12

m3

1

m4 m5 m0

1/6 1/2 1/2 1/2 1/3 1

m12

1

m4 m5 m0

1/6 1/3 1/2 1/2 1/2

m12

1

MC(5)

Best Approximant & Parameter Synthesis

1/6 4/9

slide-8
SLIDE 8

Optimal parameters may be irrational

1

m1 m3 m4 m2 m0

79/100 21/100 79/100 79/100 21/100 1 21/100 1

n1 n2 n0

1 - x - y y 1 x

slide-9
SLIDE 9

Optimal parameters may be irrational

1

m1 m3 m4 m2 m0

79/100 21/100 79/100 79/100 21/100 1 21/100 1

n1 n2 n0

1 - x - y y 1 x

x = 1 30 ⇣ 10 + √ 163 ⌘ y = 21 200

O p t i m a l p a r a m e t e r s m a y b e i r r a t i

  • n

a l !

slide-10
SLIDE 10

Optimal parameters may be irrational

1

m1 m3 m4 m2 m0

79/100 21/100 79/100 79/100 21/100 1 21/100 1

n1 n2 n0

1 - x - y y 1 x

O p t i m a l d i s t a n c e i s i r r a t i

  • n

a l !

δ(m0, n0) = 436 675 − 163 √ 163 13500 ≈ 0.49

x = 1 30 ⇣ 10 + √ 163 ⌘ y = 21 200

O p t i m a l p a r a m e t e r s m a y b e i r r a t i

  • n

a l !

slide-11
SLIDE 11

The focus of the talk

  • Probabilistic Models (Markov chains)
  • Automatic verification (e.g., Model Checking)
  • state space explosion (even after model

reduction, symbolic tech., partial-order reduction)

  • Still too large: one needs to compromise in the

accuracy of the model (introduce an error)

  • Our proposal: metric-based state space reduction
slide-12
SLIDE 12

Probabilistic Bisimulation

m0 m1 m2

1/3 2/3 1 1

n1 n0 n2 n3

1/3 1 1 1/3 1/3 1

s

1/2 1/2

[Larsen & Skou’91]

slide-13
SLIDE 13

Probabilistic Bisimulation

m0 m1 m2

1/3 2/3 1 1

n1 n0 n2 n3

1/3 1 1 1/3 1/3 1

s

1/2 1/2

[Larsen & Skou’91]

slide-14
SLIDE 14

Probabilistic Bisimulation

m0 m1 m2

1/3 2/3 1 1

n1 n0 n2 n3

1/3 1 1 1/3 1/3 1

s

1/2 1/2

[Larsen & Skou’91]

slide-15
SLIDE 15

Probabilistic Bisimulation

m0 m1 m2

1/3 2/3 1 1

n1 n0 n2 n3

1/3 1 1 1/3 1/3 1

s

1/2 1/2

[Larsen & Skou’91]

slide-16
SLIDE 16

2/3

s

1

m0 m1 m2

1/3 1 1

Probabilistic Bisimulation

[Larsen & Skou’91]

slide-17
SLIDE 17

2/3

s

1

m0 m1 m2

1/3 1 1

Probabilistic Bisimulation

[Larsen & Skou’91]

O p t i m a l l u m p i n g [ K e m e n y & S n e l l ’ 6 ]

slide-18
SLIDE 18

2/3

s

1

m0 m1 m2

1/3 1 1

Probabilistic Bisimulation

[Larsen & Skou’91]

O p t i m a l l u m p i n g [ K e m e n y & S n e l l ’ 6 ] E f fi c i e n t t e c h n i q u e [ D e r i s a v i e t a l . ’ 3 ]

slide-19
SLIDE 19

2/3

s

1

m0 m1 m2

1/3 1 1

Probabilistic Bisimulation

[Larsen & Skou’91]

O p t i m a l l u m p i n g [ K e m e n y & S n e l l ’ 6 ] E f fi c i e n t t e c h n i q u e [ D e r i s a v i e t a l . ’ 3 ] …but small variations may prevent aggregation

slide-20
SLIDE 20

Probabilistic Bisimulation

m0 m1 m2

1/3 2/3 1 1

n1 n0 n2 n3

1 1 1/3 1/3-ɛ 1

s

1/2 1/2

[Larsen & Skou’91]

1/3+ɛ

…but small variations may prevent aggregation

slide-21
SLIDE 21

Probabilistic Bisimulation

m0 m1 m2

1/3 2/3 1 1

n1 n0 n2 n3

1 1 1/3 1/3-ɛ 1

s

1/2 1/2

[Larsen & Skou’91]

1/3+ɛ

…but small variations may prevent aggregation

slide-22
SLIDE 22

m0 m1 m2

1/3 2/3

n1 n0 n2 n3

1 1 1/3 1 1 1/3+ɛ 1/3-ɛ

𝓝 = (M, 𝜐,,m0) 𝓞= (N,θ,𝛽, n0)

Bisimilarity Distance

1

slide-23
SLIDE 23

1

m0 m1 m2 n1 n0 n2 n3

1/3+ɛ 1 1 1/3 1/3-ɛ 1 1 1/3 2/3

Bisimilarity Distance

𝓝 = (M, 𝜐,,m0) 𝓞= (N,θ,𝛽, n0)

slide-24
SLIDE 24

1

m0 m1 m2 n1 n0 n2 n3

1 1 1 1 1/3 1/3 1/3 2/3 1/3+ɛ 1/3-ɛ

Bisimilarity Distance

𝓝 = (M, 𝜐,,m0) 𝓞= (N,θ,𝛽, n0)

slide-25
SLIDE 25

1

m0 m1 m2 n1 n0 n2 n3

1 1 1 1 1/3 1/3-ɛ 1/3 1/3 2/3 1/3+ɛ 1/3-ɛ

Bisimilarity Distance

𝓝 = (M, 𝜐,,m0) 𝓞= (N,θ,𝛽, n0)

slide-26
SLIDE 26

1

m0 m1 m2 n1 n0 n2 n3

1 1 1 1 1/3 1/3 1/3-ɛ 1/3 1/3 2/3 1/3+ɛ 1/3-ɛ

Bisimilarity Distance

𝓝 = (M, 𝜐,,m0) 𝓞= (N,θ,𝛽, n0)

slide-27
SLIDE 27

1

m0 m1 m2 n1 n0 n2 n3

1 1 1 1 1/3 1/3 1/3-ɛ ɛ 1/3 1/3 2/3 1/3+ɛ 1/3-ɛ

Bisimilarity Distance

𝓝 = (M, 𝜐,,m0) 𝓞= (N,θ,𝛽, n0)

slide-28
SLIDE 28

Bisimilarity Distance

Given a parameter 𝜇∈(0,1], called discount factor, the bisimilarity distance 𝜀𝜇 is the smallest distance satisfying

(fixed point characterization by van Breugel & Worrell) 𝜀𝜇(m,n) = 1 if (m)≠𝛽(n) 𝜇⋅𝓛(𝜀𝜇)(𝜐(m),θ(n)) otherwise

slide-29
SLIDE 29

Bisimilarity Distance

Given a parameter 𝜇∈(0,1], called discount factor, the bisimilarity distance 𝜀𝜇 is the smallest distance satisfying

(fixed point characterization by van Breugel & Worrell) 𝜀𝜇(m,n) = 1 if (m)≠𝛽(n) 𝜇⋅𝓛(𝜀𝜇)(𝜐(m),θ(n)) otherwise

Kantorovich lifting

slide-30
SLIDE 30

Bisimilarity Distance

Given a parameter 𝜇∈(0,1], called discount factor, the bisimilarity distance 𝜀𝜇 is the smallest distance satisfying

(fixed point characterization by van Breugel & Worrell) 𝓛(d)(𝜐(m),θ(n)) = min ∑ d(u,v)⋅C(u,v) ∑u∈M C(u,v) = θ(n)(v) ∑v∈N C(u,v) = 𝜐(m)(u)

coupling

𝜀𝜇(m,n) = 1 if (m)≠𝛽(n) 𝜇⋅𝓛(𝜀𝜇)(𝜐(m),θ(n)) otherwise

Kantorovich lifting

slide-31
SLIDE 31

Bisimilarity Distance

Given a parameter 𝜇∈(0,1], called discount factor, the bisimilarity distance 𝜀𝜇 is the smallest distance satisfying

(fixed point characterization by van Breugel & Worrell) 𝓛(d)(𝜐(m),θ(n)) = min ∑ d(u,v)⋅C(u,v) ∑u∈M C(u,v) = θ(n)(v) ∑v∈N C(u,v) = 𝜐(m)(u)

coupling

𝜀𝜇(m,n) = 1 if (m)≠𝛽(n) 𝜇⋅𝓛(𝜀𝜇)(𝜐(m),θ(n)) otherwise

Kantorovich lifting discount at each step

slide-32
SLIDE 32

Remarkable properties

Theorem (Desharnais et. al 99)

m ~ n iff 𝜀𝜇(m,n) = 0

Theorem (Chen, van Breugel, Worrell 12) The probabilistic bisimilarity distance can be computed in polynomial time

(Jonsson & L 91) (Bacci, L, Mardare 13)

slide-33
SLIDE 33

Theorem (Chen et al. FoSSaCS’12, Bacci et al. ICTAC’15)

|P(𝓝)([φ]) - P(𝓞)([φ])| ≤ 𝜀1(𝓝,𝓞)

f

  • r

a l l L T L f

  • r

m u l a s !

Approximate verification

difference in the probability of satisfying φ

slide-34
SLIDE 34

Theorem (Chen et al. FoSSaCS’12, Bacci et al. ICTAC’15)

|P(𝓝)([φ]) - P(𝓞)([φ])| ≤ 𝜀1(𝓝,𝓞)

f

  • r

a l l L T L f

  • r

m u l a s !

Approximate verification

difference in the probability of satisfying φ

P(𝓝)([φ])

P(𝓞)([φ])

1 d d

approximate solution on φ

…imagine that |𝓝|≫|𝓞|, we can use 𝓞 in place of 𝓝

slide-35
SLIDE 35

Metric-based State Space Reduction

Closest Bounded Approximant (CBA) Minimum Significant Approximant Bound (MSAB)

slide-36
SLIDE 36

Metric-based State Space Reduction

Closest Bounded Approximant (CBA) Minimum Significant Approximant Bound (MSAB) MC(k) MC N M

d

minimize d

slide-37
SLIDE 37

Metric-based State Space Reduction

Closest Bounded Approximant (CBA) Minimum Significant Approximant Bound (MSAB) MC(k) MC N M

d

minimize d MC(k) MC N M

< 1

minimize k

slide-38
SLIDE 38

List of our Results

  • CBA as bilinear program
  • The CBA’s threshold problem is
  • NP-hard (complexity lower bound)
  • PSPACE (complexity upper bound)
  • The MSAB’s threshold problem is NP-complete
  • Expectation Maximization heuristic for CBA
slide-39
SLIDE 39

The CBA-λ problem

The Closest Bounded Approximant w.r.t. 𝜀𝜇 Instance: An MC M, and a positive integer k Ouput: An MC Ñ, with at most k states minimizing 𝜀𝜇(m0,ñ0)

𝜀𝜇(m0,ñ0) = inf { 𝜀𝜇(m0,n0) | N ∈ MC(k) }

we get a solution iff the infimum is a minimum

slide-40
SLIDE 40

The CBA-λ problem

The Closest Bounded Approximant w.r.t. 𝜀𝜇 Instance: An MC M, and a positive integer k Ouput: An MC Ñ, with at most k states minimizing 𝜀𝜇(m0,ñ0)

𝜀𝜇(m0,ñ0) = inf { 𝜀𝜇(m0,n0) | N ∈ MC(k) }

we get a solution iff the infimum is a minimum

generalization of bisimilarity quotient

slide-41
SLIDE 41

CBA-λ as a Bilinear Program

Lemma (Meaningful labels) For any N∈MC(k), there exists N’∈MC(k) with labels taken from M, such that 𝜀𝜇(M,N) ≥ 𝜀𝜇(M,N’)

slide-42
SLIDE 42

CBA-λ as a Bilinear Program

Lemma (Meaningful labels) For any N∈MC(k), there exists N’∈MC(k) with labels taken from M, such that 𝜀𝜇(M,N) ≥ 𝜀𝜇(M,N’)

mimimize dm0,n0 such that ⁄ q

(u,v)œM◊N cm,n u,v · du,v Æ dm,n

m œ M, n œ N 1 ≠ –n,l Æ dm,n Æ 1 n œ N, l œ L(M), ¸(m) ”= l –n,l · –n,lÕ = 0 n œ N, l, lÕ œ L(M), l ”= lÕ q

lœL(M) –n,l = 1

n œ N q

vœN cm,n u,v = ·(m)(u)

m, u œ M, n œ N q

uœM cm,n u,v = ◊n,v

m œ M, n, v œ N cm,n

u,v Ø 0

m, u œ M, n, v œ N

slide-43
SLIDE 43

CBA-λ as a Bilinear Program

Lemma (Meaningful labels) For any N∈MC(k), there exists N’∈MC(k) with labels taken from M, such that 𝜀𝜇(M,N) ≥ 𝜀𝜇(M,N’)

mimimize dm0,n0 such that ⁄ q

(u,v)œM◊N cm,n u,v · du,v Æ dm,n

m œ M, n œ N 1 ≠ –n,l Æ dm,n Æ 1 n œ N, l œ L(M), ¸(m) ”= l –n,l · –n,lÕ = 0 n œ N, l, lÕ œ L(M), l ”= lÕ q

lœL(M) –n,l = 1

n œ N q

vœN cm,n u,v = ·(m)(u)

m, u œ M, n œ N q

uœM cm,n u,v = ◊n,v

m œ M, n, v œ N cm,n

u,v Ø 0

m, u œ M, n, v œ N

slide-44
SLIDE 44

CBA-λ as a Bilinear Program

Lemma (Meaningful labels) For any N∈MC(k), there exists N’∈MC(k) with labels taken from M, such that 𝜀𝜇(M,N) ≥ 𝜀𝜇(M,N’)

mimimize dm0,n0 such that ⁄ q

(u,v)œM◊N cm,n u,v · du,v Æ dm,n

m œ M, n œ N 1 ≠ –n,l Æ dm,n Æ 1 n œ N, l œ L(M), ¸(m) ”= l –n,l · –n,lÕ = 0 n œ N, l, lÕ œ L(M), l ”= lÕ q

lœL(M) –n,l = 1

n œ N q

vœN cm,n u,v = ·(m)(u)

m, u œ M, n œ N q

uœM cm,n u,v = ◊n,v

m œ M, n, v œ N cm,n

u,v Ø 0

m, u œ M, n, v œ N

slide-45
SLIDE 45

CBA-λ as a Bilinear Program

Lemma (Meaningful labels) For any N∈MC(k), there exists N’∈MC(k) with labels taken from M, such that 𝜀𝜇(M,N) ≥ 𝜀𝜇(M,N’)

mimimize dm0,n0 such that ⁄ q

(u,v)œM◊N cm,n u,v · du,v Æ dm,n

m œ M, n œ N 1 ≠ –n,l Æ dm,n Æ 1 n œ N, l œ L(M), ¸(m) ”= l –n,l · –n,lÕ = 0 n œ N, l, lÕ œ L(M), l ”= lÕ q

lœL(M) –n,l = 1

n œ N q

vœN cm,n u,v = ·(m)(u)

m, u œ M, n œ N q

uœM cm,n u,v = ◊n,v

m œ M, n, v œ N cm,n

u,v Ø 0

m, u œ M, n, v œ N

slide-46
SLIDE 46

CBA-λ as a Bilinear Program

this characterization has two main consequences…

  • 1. CBA-λ admits always a solution
  • 2. CBA-λ is can be approximated up

to any precision

slide-47
SLIDE 47

Complexity of CBA-λ

“To study the complexity of an optimization problem

  • ne has to look at its decision variant”

(C. Papadimitriou)

slide-48
SLIDE 48

Complexity of CBA-λ

“To study the complexity of an optimization problem

  • ne has to look at its decision variant”

(C. Papadimitriou) Bounded Approximant threshold wrt dλ Instance: An MC M, a positive integer k, and a rational ε > 0 Ouput: yes iff there exists N with at most k states such that 𝜀𝜇(m0,n0) ≤ ε

slide-49
SLIDE 49

Complexity upper bound

BA-λ is in PSPACE

Theorem

Proof sketch: we can encode the question ⟨M,k,ε⟩∈BA-λ to that of checking the feasibility of a set of bilinear inequalities. This can be encoded as a decision problem for the existential theory of the reals, thus it can be solved in PSPACE [Canny—STOC88].

slide-50
SLIDE 50

Complexity lower bound

BA-λ is NP-hard

Theorem

Proof idea: we provide a reduction from VERTEX COVER.

slide-51
SLIDE 51

Complexity lower bound

BA-λ is NP-hard

Theorem unlikely to solve CBA as simple linear program

Proof idea: we provide a reduction from VERTEX COVER.

slide-52
SLIDE 52

The MSAB-λ problem

The Minimum Significant Approximant Bound wrt 𝜀𝜇 Instance: An MC M Output: The smallest k such that dλ(m0,n0)<1, for some N∈MC(k)

slide-53
SLIDE 53

The MSAB-λ problem

The Minimum Significant Approximant Bound wrt 𝜀𝜇 Instance: An MC M Output: The smallest k such that dλ(m0,n0)<1, for some N∈MC(k)

For λ<1, the MSAB-λ problem is trivial, because the solution is always k=1

slide-54
SLIDE 54

The MSAB-λ problem

The Minimum Significant Approximant Bound wrt 𝜀𝜇 Instance: An MC M Output: The smallest k such that dλ(m0,n0)<1, for some N∈MC(k)

For λ<1, the MSAB-λ problem is trivial, because the solution is always k=1

For λ=1, the same problem is surprisingly difficult…

slide-55
SLIDE 55

Complexity of MSAB-1

…as before we should look at its decision variant

slide-56
SLIDE 56

Complexity of MSAB-1

Significant Bounded Approximant wrt 𝜀1 Instance: An MC M and a positive k Output: yes iff there exists N with at most k states such that 𝜀1(m0,n0)<1. …as before we should look at its decision variant

slide-57
SLIDE 57

Complexity of MSAB-1

Significant Bounded Approximant wrt 𝜀1 Instance: An MC M and a positive k Output: yes iff there exists N with at most k states such that 𝜀1(m0,n0)<1. …as before we should look at its decision variant

SBA-1 is NP-complete

Theorem

slide-58
SLIDE 58

SBA-1 ⊆ NP

Lemma Assume M be maximally collapsed. Then, ⟨M,k⟩∈SBA-1 iff

C

m0 mn

G(M) =

BSCC

and h+|C| ≤ k

number of labels in m0…mn-1

slide-59
SLIDE 59

SBA-1 ⊆ NP

Lemma Assume M be maximally collapsed. Then, ⟨M,k⟩∈SBA-1 iff

C

m0 mn

G(M) =

BSCC

and h+|C| ≤ k

Proof sketch: compute with Tarjan’s algorithm all the SCCs of G(M). Then non deterministically choose a BSCC and a path to it. In poly- time we can count the number of labels in the path and the size of the BSCC.

number of labels in m0…mn-1

slide-60
SLIDE 60

SBA-1 is NP-hard

Proof sketch: by reduction to VERTEX COVER: ⟨G,h⟩∈VERTEX COVER iff ⟨MG, h+m+1⟩∈SBA-1

1 2 3 4

e1 e2 e3

e3 e2 e1 e0

1 2 3 2 3 4

1 1/2 1/2 1 1/2 1/2 1/2 1/2 1 1 1 1 1

sink

slide-61
SLIDE 61

SBA-1 is NP-hard

Proof sketch: by reduction to VERTEX COVER: ⟨G,h⟩∈VERTEX COVER iff ⟨MG, h+m+1⟩∈SBA-1

1 2 3 4

e1 e2 e3

e3 e2 e1 e0

1 2 3 2 3 4

1 1/2 1/2 1 1/2 1/2 1/2 1/2 1 1 1 1 1

sink

paths from e3 to e0 describes all vertex covers of G

slide-62
SLIDE 62

Towards an Algorithm…

slide-63
SLIDE 63

Towards an Algorithm…

  • CBA can be solved as a bilinear program.


Theoretically nice, but practically unfeasible!
 (our implementation in PENBMI can
 handle MCs with at most 5 states…)

slide-64
SLIDE 64

Towards an Algorithm…

  • CBA can be solved as a bilinear program.


Theoretically nice, but practically unfeasible!
 (our implementation in PENBMI can
 handle MCs with at most 5 states…)

  • We are happy with sub-optimal solutions if 


they can be obtained by a practical algorithm.

slide-65
SLIDE 65

EM-like Algorithm

  • Given the MC M and an initial approximant N0
  • it produces a sequence N0, …, Nh of approximants


having strictly decreasing distance from M

  • Nh may be a sub-optimal solution of CBA-λ

MC(k) MC N0 M

dh

N1 Nh

d0 > d1 > … > dh

slide-66
SLIDE 66

EM-like Algorithm

Algorithm 1 Expectation Maximization

Input: M = (M, ⌧, `), N0 = (N, ✓0, ↵), and h ∈ N.

  • 1. i ← 0
  • 2. repeat

3. i ← i + 1 4. compute C ∈ ⌦(M, Ni−1) such that λ(M, Ni−1) = C

λ(M, Ni−1)

5. ✓i ← UpdateTransition(✓i−1, C) 6. Ni ← (N, ✓i, ↵)

  • 7. until λ(M, Ni) > λ(M, Ni−1) or i ≥ h
  • 8. return Ni−1
slide-67
SLIDE 67

EM-like Algorithm

UpdateTransition assigns greater probability to transitions that are most representative of the behavior of M Intuitive Idea

Algorithm 1 Expectation Maximization

Input: M = (M, ⌧, `), N0 = (N, ✓0, ↵), and h ∈ N.

  • 1. i ← 0
  • 2. repeat

3. i ← i + 1 4. compute C ∈ ⌦(M, Ni−1) such that λ(M, Ni−1) = C

λ(M, Ni−1)

5. ✓i ← UpdateTransition(✓i−1, C) 6. Ni ← (N, ✓i, ↵)

  • 7. until λ(M, Ni) > λ(M, Ni−1) or i ≥ h
  • 8. return Ni−1
slide-68
SLIDE 68

Two update heuristics

  • Averaged Marginal (AM): given Nk we construct


Nk+1 by averaging the marginal of certain 
 “coupling variables” obtained by optimizing 
 the number of occurrences of the edges that
 are most likely to be seen in M.

  • Averaged Expectations (AE): similar to the above,


but now the Nk+1 looks only the expectation


  • f the number of occurrences of the edges 


likely to be found in M.

slide-69
SLIDE 69

IPv4 Zero Conf Protocol

0.8 0.2 1. 0.5 0.5 1. 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 1 4 2 3 2 2 2 2 2 2 2 2 2

M

Averaged Marginal (AM)

slide-70
SLIDE 70

IPv4 Zero Conf Protocol

0.8 0.2 1. 0.5 0.5 1. 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 1 4 2 3 2 2 2 2 2 2 2 2 2

M

0.2 0.8 1. 0.9 0.1 1. 0.9 0.1 1 4 2 3 2

𝜀0.9(M,N0) ≈ 0.67

Averaged Marginal (AM)

slide-71
SLIDE 71

IPv4 Zero Conf Protocol

0.8 0.2 1. 0.5 0.5 1. 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 1 4 2 3 2 2 2 2 2 2 2 2 2

M

0.2 0.8 1. 0.9 0.1 1. 0.9 0.1 1 4 2 3 2

𝜀0.9(M,N0) ≈ 0.67

Averaged Marginal (AM)

𝜀0.9(M,N1) ≈ 0.043

0.8 0.2 1. 0.526316 0.473684 1. 0.9 0.1 1 4 2 3 2
slide-72
SLIDE 72

IPv4 Zero Conf Protocol

0.8 0.2 1. 0.5 0.5 1. 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 1 4 2 3 2 2 2 2 2 2 2 2 2

M

0.2 0.8 1. 0.9 0.1 1. 0.9 0.1 1 4 2 3 2

𝜀0.9(M,N0) ≈ 0.67

Averaged Marginal (AM)

𝜀0.9(M,N1) ≈ 0.043

0.8 0.2 1. 0.526316 0.473684 1. 0.9 0.1 1 4 2 3 2

𝜀0.9(M,N2) ≈ 0.041

0.8 0.2 1. 0.5 0.5 1. 0.909091 0.0909091 1 4 2 3 2
slide-73
SLIDE 73

IPv4 Zero Conf Protocol

0.2 0.8 1. 0.9 0.1 1. 0.9 0.1 1 4 2 3 2

𝜀0.9(M,N0) ≈ 0.67

Averaged Expectations (AE)

0.8 0.2 1. 0.5 0.5 1. 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 1 4 2 3 2 2 2 2 2 2 2 2 2

M

slide-74
SLIDE 74

IPv4 Zero Conf Protocol

0.2 0.8 1. 0.9 0.1 1. 0.9 0.1 1 4 2 3 2

𝜀0.9(M,N0) ≈ 0.67

0.78728 0.21272 1. 0.872927 0.127073 1. 0.946363 0.0536368 1 4 2 3 2

𝜀0.9(M,N1) ≈ 0.08

Averaged Expectations (AE)

0.8 0.2 1. 0.5 0.5 1. 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 1 4 2 3 2 2 2 2 2 2 2 2 2

M

slide-75
SLIDE 75

IPv4 Zero Conf Protocol

0.2 0.8 1. 0.9 0.1 1. 0.9 0.1 1 4 2 3 2

𝜀0.9(M,N0) ≈ 0.67

0.78728 0.21272 1. 0.872927 0.127073 1. 0.946363 0.0536368 1 4 2 3 2

𝜀0.9(M,N1) ≈ 0.08

0.873866 0.126134 1. 0.875451 0.124549 1. 0.990499 0.00950111 1 4 2 3 2

𝜀0.9(M,N2) ≈ 0.11

Averaged Expectations (AE)

0.8 0.2 1. 0.5 0.5 1. 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 0.5 1 4 2 3 2 2 2 2 2 2 2 2 2

M

slide-76
SLIDE 76

Drunkard's Walk

Averaged Marginal (AM)

0.1 0.9 0.9 0.1 0.1 0.9 1. 1. 0.9 0.1 0.9 0.1 0.1 0.9 0.1 0.9 0.1 0.9 0.9 0.1 1 4 4 2 3 4 4 4 4 4 4

M

slide-77
SLIDE 77

Drunkard's Walk

Averaged Marginal (AM)

𝜀0.9(M,N0) ≈ 0.64

0.3 0.7 0.7 0.3 0.3 0.7 1. 1. 0.7 0.3 0.3 0.7 1 4 4 2 3 4 4

0.1 0.9 0.9 0.1 0.1 0.9 1. 1. 0.9 0.1 0.9 0.1 0.1 0.9 0.1 0.9 0.1 0.9 0.9 0.1 1 4 4 2 3 4 4 4 4 4 4

M

slide-78
SLIDE 78

Drunkard's Walk

Averaged Marginal (AM)

𝜀0.9(M,N0) ≈ 0.64

0.3 0.7 0.7 0.3 0.3 0.7 1. 1. 0.7 0.3 0.3 0.7 1 4 4 2 3 4 4

0.1 0.9 0.9 0.1 0.1 0.9 1. 1. 0.9 0.1 0.9 0.1 0.1 0.9 0.1 0.9 0.1 0.9 0.9 0.1 1 4 4 2 3 4 4 4 4 4 4

M

𝜀0.9(M,N1) ≈ 0.56

0.235318 0.764682 0.144928 0.855072 0.149254 0.850746 1. 1. 1. 0.844828 0.155172 1 4 4 2 3 4 4

slide-79
SLIDE 79

Drunkard's Walk

Averaged Marginal (AM)

𝜀0.9(M,N0) ≈ 0.64

0.3 0.7 0.7 0.3 0.3 0.7 1. 1. 0.7 0.3 0.3 0.7 1 4 4 2 3 4 4

0.1 0.9 0.9 0.1 0.1 0.9 1. 1. 0.9 0.1 0.9 0.1 0.1 0.9 0.1 0.9 0.1 0.9 0.9 0.1 1 4 4 2 3 4 4 4 4 4 4

M

𝜀0.9(M,N1) ≈ 0.56

0.235318 0.764682 0.144928 0.855072 0.149254 0.850746 1. 1. 1. 0.844828 0.155172 1 4 4 2 3 4 4

𝜀0.9(M,N2) ≈ 0.567

0.22602 0.77398 0.147059 0.852941 0.169492 0.830508 1. 1. 1. 0.842105 0.157895 1 4 4 2 3 4 4

slide-80
SLIDE 80

Drunkard's Walk

Averaged Expectations (AE)

𝜀0.9(M,N0) ≈ 0.64

0.3 0.7 0.7 0.3 0.3 0.7 1. 1. 0.7 0.3 0.3 0.7 1 4 4 2 3 4 4

0.1 0.9 0.9 0.1 0.1 0.9 1. 1. 0.9 0.1 0.9 0.1 0.1 0.9 0.1 0.9 0.1 0.9 0.9 0.1 1 4 4 2 3 4 4 4 4 4 4

M

slide-81
SLIDE 81

Drunkard's Walk

Averaged Expectations (AE)

𝜀0.9(M,N0) ≈ 0.64

0.3 0.7 0.7 0.3 0.3 0.7 1. 1. 0.7 0.3 0.3 0.7 1 4 4 2 3 4 4

0.257064 0.742936 0.523324 0.476676 0.114461 0.885539 1. 1. 1. 0.40076 0.59924 1 4 4 2 3 4 4

𝜀0.9(M,N1) ≈ 0.56

0.1 0.9 0.9 0.1 0.1 0.9 1. 1. 0.9 0.1 0.9 0.1 0.1 0.9 0.1 0.9 0.1 0.9 0.9 0.1 1 4 4 2 3 4 4 4 4 4 4

M

slide-82
SLIDE 82

Drunkard's Walk

Averaged Expectations (AE)

𝜀0.9(M,N0) ≈ 0.64

0.3 0.7 0.7 0.3 0.3 0.7 1. 1. 0.7 0.3 0.3 0.7 1 4 4 2 3 4 4

0.257064 0.742936 0.523324 0.476676 0.114461 0.885539 1. 1. 1. 0.40076 0.59924 1 4 4 2 3 4 4

𝜀0.9(M,N1) ≈ 0.56

0.194657 0.805343 0.377773 0.622227 0.0685134 0.931487 1. 1. 1. 0.46345 0.53655 1 4 4 2 3 4 4

𝜀0.9(M,N2) ≈ 0.543

0.1 0.9 0.9 0.1 0.1 0.9 1. 1. 0.9 0.1 0.9 0.1 0.1 0.9 0.1 0.9 0.1 0.9 0.9 0.1 1 4 4 2 3 4 4 4 4 4 4

M

slide-83
SLIDE 83

Drunkard's Walk

Averaged Expectations (AE)

𝜀0.9(M,N0) ≈ 0.64

0.3 0.7 0.7 0.3 0.3 0.7 1. 1. 0.7 0.3 0.3 0.7 1 4 4 2 3 4 4

0.257064 0.742936 0.523324 0.476676 0.114461 0.885539 1. 1. 1. 0.40076 0.59924 1 4 4 2 3 4 4

𝜀0.9(M,N1) ≈ 0.56

0.194657 0.805343 0.377773 0.622227 0.0685134 0.931487 1. 1. 1. 0.46345 0.53655 1 4 4 2 3 4 4

𝜀0.9(M,N2) ≈ 0.543

0.142708 0.857292 0.280029 0.719971 0.0441171 0.955883 1. 1. 1. 0.501941 0.498059 1 4 4 2 3 4 4

𝜀0.9(M,N3) ≈ 0.540

0.1 0.9 0.9 0.1 0.1 0.9 1. 1. 0.9 0.1 0.9 0.1 0.1 0.9 0.1 0.9 0.1 0.9 0.9 0.1 1 4 4 2 3 4 4 4 4 4 4

M

slide-84
SLIDE 84

Case |M| k λ = 1 λ = 0.8 δλ-init δλ-final # time δλ-init δλ-final # time IPv4 (AM) 23 5 0.775 0.054 3 4.8 0.576 0.025 3 4.8 53 5 0.856 0.062 3 25.7 0.667 0.029 3 25.9 103 5 0.923 0.067 3 116.3 0.734 0.035 3 116.5 53 6 0.757 0.030 3 39.4 0.544 0.011 3 39.4 103 6 0.837 0.032 3 183.7 0.624 0.017 3 182.7 203 6 – – – TO – – – TO IPv4 (AE) 23 5 0.775 0.109 2 2.7 0.576 0.049 3 4.2 53 5 0.856 0.110 2 14.2 0.667 0.049 3 21.8 103 5 0.923 0.110 2 67.1 0.734 0.049 3 100.4 53 6 0.757 0.072 2 21.8 0.544 0.019 3 33.0 103 6 0.837 0.072 2 105.9 0.624 0.019 3 159.5 203 6 – – – TO – – – TO DrkW (AM) 39 7 0.565 0.466 14 259.3 0.432 0.323 14 252.8 49 7 0.568 0.460 14 453.7 0.433 0.322 14 420.5 59 8 0.646 – – TO 0.423 – – TO DrkW (AE) 39 7 0.565 0.435 11 156.6 0.432 0.321 2 28.6 49 7 0.568 0.434 10 247.7 0.433 0.316 2 46.2 59 8 0.646 0.435 10 588.9 0.423 0.309 2 115.7 Table 1. Comparison of the performance of EM algorithm on the IPv4 zeroconf pro- tocol and the classic Drunkard’s Walk w.r.t. the heuristics AM and AE.

slide-85
SLIDE 85

What we have seen

Metric-based state space reduction for MCs

  • 1. Closest Bounded Approximant (CBA)


encoded as a bilinear program

  • 2. Bounded Approximant (BA)


PSPACE & NP-hard for all 𝜇∈(0,1]

  • 3. Significant Bounded Approximant (SBA)


NP-complete for 𝜇=1

Theoretical Results

We proposed an EM-like method to

  • btain a sub-optimal approximants

Practical Results

slide-86
SLIDE 86

Previous work

  • On-the-Fly Exact Computation of Bisimilarity Distances - TACAS

2013

  • The BisimDist Library: Efficient Computation of Bisimilarity

Distances for Markovian Models - QEST 2013

  • Computing Behavioral Distances, Compositionally - MFCS 2013
  • Converging from Branching to Linear Metrics on Markov Chains
  • ICTAC 2015
  • On the Metric-based Approximate Minimization of Markov

Chains - ICALP 2017

slide-87
SLIDE 87

Future Work

  • Is BA-λ SUM-OF-SQUARE-ROOTS-hard?
  • Can we obtain a real/better EM-heuristics?
  • What about different models/distances?