Strategy Synthesis for Quantitative Objectives Krishnendu Chatterjee - - PowerPoint PPT Presentation

strategy synthesis for quantitative objectives
SMART_READER_LITE
LIVE PREVIEW

Strategy Synthesis for Quantitative Objectives Krishnendu Chatterjee - - PowerPoint PPT Presentation

Strategy Synthesis for Quantitative Objectives Krishnendu Chatterjee 1 Mickael Randour 2 cois Raskin 3 Jean-Fran 1 IST Austria 2 UMONS 3 ULB 30.11.2011 EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion Aim of this


slide-1
SLIDE 1

Strategy Synthesis for Quantitative Objectives

Krishnendu Chatterjee1 Mickael Randour2 Jean-Fran¸ cois Raskin3

1 IST Austria 2 UMONS 3 ULB

30.11.2011

slide-2
SLIDE 2

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Aim of this work

Study games with (multi-dimensional) quantitative objectives: energy and mean-payoff. Address questions that revolve around strategies:

⊲ bounds on memory, ⊲ synthesis algorithm, ⊲ randomness

?

∼ memory.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 1 / 42

slide-3
SLIDE 3

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Results Overview

Strategy synthesis

MEGs MMPGs

  • ptimal

finite-memory optimal

  • ptimal

Memory exp. exp. infinite [CDHR10] Synthesis EXPTIME EXPTIME /

Randomness as a substitute for finite-memory

MEGs EPGs MMPGs MPBGs MPPGs 1-player × × √ √ √ (conj.) 2-player × × × √ √ (conj.)

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 2 / 42

slide-4
SLIDE 4

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

1 Classical energy and mean-payoff games 2 Extensions to multi-dimensions and parity 3 Strategy synthesis 4 Randomization as a substitute to finite-memory 5 Conclusion and ongoing work

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 3 / 42

slide-5
SLIDE 5

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

1 Classical energy and mean-payoff games 2 Extensions to multi-dimensions and parity 3 Strategy synthesis 4 Randomization as a substitute to finite-memory 5 Conclusion and ongoing work

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 4 / 42

slide-6
SLIDE 6

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Turn-based games

s0 s1 s2 s3 s4 s5

G = (S1, S2, sinit, E) S = S1 ∪ S2, S1 ∩ S2 = ∅, E ⊆ S × S P1 states = P2 states =

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 5 / 42

slide-7
SLIDE 7

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Turn-based games

s0 s1 s2 s3 s4 s5

G = (S1, S2, sinit, E) S = S1 ∪ S2, S1 ∩ S2 = ∅, E ⊆ S × S P1 states = P2 states = Play π = s0s1s2 . . . sn . . . s.t. s0 = sinit Prefix ρ = π(n) = s0s1s2 . . . sn

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 5 / 42

slide-8
SLIDE 8

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Pure strategies

s0 s1 s2 s3 s4 s5

Pure strategy for Pi λi ∈ Λi : Prefsi(G) → S s.t. for all ρ ∈ Prefsi(G), (Last(ρ), λi(ρ)) ∈ E Memoryless strategy λpm

i

∈ ΛPM

i

: Si → S Finite-memory strategy λfm

i

∈ ΛFM

i

: Prefsi(G) → S, and can be encoded as a deterministic Moore machine

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 5 / 42

slide-9
SLIDE 9

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Integer payoff function

s0 s1 s2 s3 s4 s5 −2 2 3 1 −2 1

G = (S1, S2, sinit, E, w) w : E → Z

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 5 / 42

slide-10
SLIDE 10

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Integer payoff function

s0 s1 s2 s3 s4 s5 −2 2 3 1 −2 1

G = (S1, S2, sinit, E, w) w : E → Z Energy level EL(ρ) = i=n−1

i=0

w(si, si+1) Mean-payoff MP(π) = lim infn→∞ 1

nEL(π(n))

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 5 / 42

slide-11
SLIDE 11

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Energy and mean-payoff objectives

s0 s1 s2 s3 s4 s5 −2 2 3 1 −2 1

Energy objective Given initial credit v0 ∈ N, PosEnergyG(v0) = {π ∈ Plays(G) | ∀ n ≥ 0 : v0 + EL(π(n)) ∈ N} Mean-payoff objective Given threshold v ∈ Q, MeanPayoffG(v) = {π ∈ Plays(G) | MP(π) ≥ v}

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 5 / 42

slide-12
SLIDE 12

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Energy and mean-payoff objectives

s0 s1 s2 s3 s4 s5 −2 2 3 1 −2 1

λ1(s3) = s4

⊲ λ1 wins for MeanPayoffG( −1

4 )

⊲ λ1 loses for PosEnergyG(v0), for any arbitrary high initial credit

λ1(s3) = s5

⊲ λ1 wins for MeanPayoffG( 1

2)

⊲ λ1 wins for PosEnergyG(v0), with v0 = 2

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 5 / 42

slide-13
SLIDE 13

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Decision problems

Unknown initial credit problem: ∃? v0 ∈ N, λ1 ∈ Λ1 s.t. λ1 wins for PosEnergyG(v0) Mean-payoff threshold problem: Given v ∈ Q, ∃? λ1 ∈ Λ1 s.t. λ1 wins for MeanPayoffG(v) MPG threshold v problem equivalent to EG−v unknown initial credit problem [BFL+08].

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 6 / 42

slide-14
SLIDE 14

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Complexity of EGs and MPGs

EGs MPGs Memory to win memoryless memoryless [CdAHS03, BFL+08] [EM79, LL69] Decision problem NP ∩ coNP NP ∩ coNP

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 7 / 42

slide-15
SLIDE 15

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

1 Classical energy and mean-payoff games 2 Extensions to multi-dimensions and parity 3 Strategy synthesis 4 Randomization as a substitute to finite-memory 5 Conclusion and ongoing work

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 8 / 42

slide-16
SLIDE 16

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Multi-dimensional weights

s0 s1 s2 s3 s4 s5 −2 2 3 1 −2 1

G = (S1, S2, sinit, E, w) w : E → Z

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 9 / 42

slide-17
SLIDE 17

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Multi-dimensional weights

s0 s1 s2 s3 s4 s5 (2, 1) (1, −2) (0, −2) (−3, 3) (0, 1) (1, −1) (0, 0) (1, 0)

G = (S1, S2, sinit, E, k, w) w : E → Zk ⊲ multiple quantitative aspects ⊲ natural extensions of energy and mean-payoff objectives and associated decision problems

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 9 / 42

slide-18
SLIDE 18

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

MEGs & MMPGs

Finite memory suffice for MEGs [CDHR10].

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 10 / 42

slide-19
SLIDE 19

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

MEGs & MMPGs

Finite memory suffice for MEGs [CDHR10]. However, infinite memory is needed for MMPGs, even with

  • nly one player! [CDHR10]

s0 s1 (2, 0) (0, 2) (0, 0) (0, 0)

⊲ To obtain MP(π) = (1, 1), P1 has to visit s0 and s1 for longer and longer intervals before jumping from one to the other. ⊲ Any finite-memory strategy induces an ultimately periodic play s.t. MP(π) = (x, y), x + y < 2. ⊲ With lim sup as MP the gap is huge : (2, 2).

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 10 / 42

slide-20
SLIDE 20

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

MEGs & MMPGs

If players are restricted to finite memory [CDHR10],

⊲ MEGs and MMPGs are still determined and they are log-space equivalent, ⊲ the unknown initial credit and the mean-payoff threshold problems are coNP-complete, ⊲ no clue on memory bounds for P1 (for P2, we know it is memoryless).

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 11 / 42

slide-21
SLIDE 21

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

MEGs & MMPGs

If players are restricted to finite memory [CDHR10],

⊲ MEGs and MMPGs are still determined and they are log-space equivalent, ⊲ the unknown initial credit and the mean-payoff threshold problems are coNP-complete, ⊲ no clue on memory bounds for P1 (for P2, we know it is memoryless).

Other interesting results on decision problems on MEGs are proved in [FJLS11]. Surprisingly, given a fixed initial vector, the problem becomes EXPSPACE-hard.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 11 / 42

slide-22
SLIDE 22

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Parity

s0 s1 s2 s3 s4 s5 −2 2 3 1 −2 1

G = (S1, S2, sinit, E, w)

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 12 / 42

slide-23
SLIDE 23

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Parity

s0 p = 1 s1 p = 1 s2 p = 3 s3 p = 2 s4 p = 0 s5 p = 2 −2 2 3 1 −2 1

Gp =

  • S1, S2, sinit, E, w, p
  • p : S → N

Par(π) = min {p(s) | s ∈ Inf(π)} ⊲ ParityGp = {π ∈ Plays(Gp) | Par(π) mod 2 = 0} ⊲ canonical way to express ω-regular

  • bjectives

⊲ achieve the energy or mean-payoff

  • bjective while satisfying the parity

condition

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 12 / 42

slide-24
SLIDE 24

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Parity

s0 p = 1 s1 p = 1 s2 p = 3 s3 p = 2 s4 p = 0 s5 p = 2 −2 2 3 1 −2 1

To win the energy parity objective, P1 must

⊲ visit s4 infinitely often, ⊲ alternate with visits of s5 to fund future visits of s4.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 12 / 42

slide-25
SLIDE 25

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Parity

s0 p = 1 s1 p = 1 s2 p = 3 s3 p = 2 s4 p = 0 s5 p = 2 −2 2 3 1 −2 1

To win the energy parity objective, P1 must

⊲ visit s4 infinitely often, ⊲ alternate with visits of s5 to fund future visits of s4.

To achieve optimality for the mean-payoff parity objective, P1 should wait longer and longer between visits of s4.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 12 / 42

slide-26
SLIDE 26

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

EPGs & MPPGs

Exponential memory suffice for EPGs and deciding the winner is in NP ∩ coNP [CD10].

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 13 / 42

slide-27
SLIDE 27

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

EPGs & MPPGs

Exponential memory suffice for EPGs and deciding the winner is in NP ∩ coNP [CD10]. Infinite memory is needed for MPPGs and deciding the winner is in NP ∩ coNP [CHJ05, BMOU11].

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 13 / 42

slide-28
SLIDE 28

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

EPGs & MPPGs

Exponential memory suffice for EPGs and deciding the winner is in NP ∩ coNP [CD10]. Infinite memory is needed for MPPGs and deciding the winner is in NP ∩ coNP [CHJ05, BMOU11]. Finite-memory ε-strategies for MPPGs always exist [BCHJ09]. P1 has a winning strategy for the MPPG G, p, w iff P1 has a winning strategy for the EPG G, p, w + ε, with ε =

1 |S|+1

[CD10].

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 13 / 42

slide-29
SLIDE 29

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Restriction to finite memory

Infinite memory:

⊲ needed for MMPGs & MPPGs, ⊲ practical implementation is unrealistic.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 14 / 42

slide-30
SLIDE 30

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Restriction to finite memory

Infinite memory:

⊲ needed for MMPGs & MPPGs, ⊲ practical implementation is unrealistic.

Finite memory:

⊲ preserves game determinacy, ⊲ provides equivalence between energy and mean-payoff settings, ⊲ the way to go for strategy synthesis.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 14 / 42

slide-31
SLIDE 31

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Restriction to finite memory

Infinite memory:

⊲ needed for MMPGs & MPPGs, ⊲ practical implementation is unrealistic.

Finite memory:

⊲ preserves game determinacy, ⊲ provides equivalence between energy and mean-payoff settings, ⊲ the way to go for strategy synthesis.

Our goals:

⊲ bounds on memory, ⊲ strategy synthesis algorithm, ⊲ encoding of memory as randomness.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 14 / 42

slide-32
SLIDE 32

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

1 Classical energy and mean-payoff games 2 Extensions to multi-dimensions and parity 3 Strategy synthesis 4 Randomization as a substitute to finite-memory 5 Conclusion and ongoing work

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 15 / 42

slide-33
SLIDE 33

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Obtained results

MEGs MMPGs

  • ptimal

finite-memory optimal

  • ptimal

Memory exp. exp. infinite [CDHR10] Synthesis EXPTIME EXPTIME /

By [CDHR10], we only have to consider MEGs. Recall that the unknown initial credit decision problem is coNP-complete.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 16 / 42

slide-34
SLIDE 34

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Upper memory bound: SCTs

s0 s1 s2 s3 s4 s5 (−1, 1) (0, 2) (0, 1) (0, 0) (1, −1) (−2, 1) (−2, 1) (0, −1) (2, 0)

A winning strategy λ1 for initial credit v0 = (2, 0) is

⊲ λ1(∗s1s3) = s4, ⊲ λ1(∗s2s3) = s5, ⊲ λ1(∗s5s3) = s5.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 17 / 42

slide-35
SLIDE 35

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Upper memory bound: SCTs

s0 s1 s2 s3 s4 s5 (−1, 1) (0, 2) (0, 1) (0, 0) (1, −1) (−2, 1) (−2, 1) (0, −1) (2, 0)

A winning strategy λ1 for initial credit v0 = (2, 0) is

⊲ λ1(∗s1s3) = s4, ⊲ λ1(∗s2s3) = s5, ⊲ λ1(∗s5s3) = s5.

Lemma: To win, P1 must be able to enforce positive cycles.

⊲ Self-covering paths on VASS [Rac78, RY86]. ⊲ Self-covering trees (SCTs) on reachability games over VASS [BJK10].

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 17 / 42

slide-36
SLIDE 36

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Upper memory bound: SCTs

s0 s1 s2 s3 s4 s5 (−1, 1) (0, 2) (0, 1) (0, 0) (1, −1) (−2, 1) (−2, 1) (0, −1) (2, 0)

s0, (0, 0) s1, (−1, 1) s2, (0, 2) s3, (−1, 2) s3, (0, 2) s4, (0, 1) s5, (−2, 3) s0, (0, 0) s3, (0, 3)

Pebble moves ⇒ strategy.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 17 / 42

slide-37
SLIDE 37

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Upper memory bound: SCTs

T = (Q, R) is a SCT for s0, Θ : Q → S × Zk is a labeling function. Root labeled s0, (0, . . . , 0). Non-leaf nodes have

⊲ unique child if P1, ⊲ all possible children if P2.

Leafs have energy ancestors: ancestors with lower label.

s0, (0, 0) s1, (−1, 1) s2, (0, 2) s3, (−1, 2) s3, (0, 2) s4, (0, 1) s5, (−2, 3) s0, (0, 0) s3, (0, 3)

Pebble moves ⇒ strategy.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 17 / 42

slide-38
SLIDE 38

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Upper memory bound: SCTs for VASS games

Theorem (application of [BJK10]): On a VASS game with weights in {−1, 0, 1}k, if state s is winning for P1, there is a SCT for s whose depth is at most l = 2(d−1)·|S| · (|S| + 1)c·k2, with c a constant independent of the considered VASS game and d its branching degree. If there exists a winning strategy, there exists a “compact” one.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 18 / 42

slide-39
SLIDE 39

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Upper memory bound: SCTs for MEGs

exp.

w : E → {−1, 0, 1}k l = 2(d−1)·|S| · (|S| + 1)c·k2 Depth bound from [BJK10].

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 19 / 42

slide-40
SLIDE 40

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Upper memory bound: SCTs for MEGs

2-exp.

w : E → {−1, 0, 1}k l = 2(d−1)·|S| · (|S| + 1)c·k2 ⇓ w : E → Zk, W max absolute weight, V bits to encode W l = 2(d−1)·W ·|S| · (W · |S| + 1)c·k2 = 2(d−1)·2V ·|S| · (W · |S| + 1)c·k2 Naive approach: blow-up by W in the size of the state space.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 19 / 42

slide-41
SLIDE 41

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Upper memory bound: SCTs for MEGs

2-exp. 3-exp.

w : E → {−1, 0, 1}k l = 2(d−1)·|S| · (|S| + 1)c·k2 ⇓ w : E → Zk, W max absolute weight, V bits to encode W l = 2(d−1)·W ·|S| · (W · |S| + 1)c·k2 = 2(d−1)·2V ·|S| · (W · |S| + 1)c·k2 ⇓ Width bounded by L = dl Naive approach: width increases exponentially with depth.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 19 / 42

slide-42
SLIDE 42

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Upper memory bound: SCTs for MEGs

2-exp. 3-exp.

w : E → {−1, 0, 1}k l = 2(d−1)·|S| · (|S| + 1)c·k2 ⇓ w : E → Zk, W max absolute weight, V bits to encode W l = 2(d−1)·W ·|S| · (W · |S| + 1)c·k2 = 2(d−1)·2V ·|S| · (W · |S| + 1)c·k2 ⇓ Width bounded by L = dl Naive approach: overall, 3-exp. memory ≤ L · l.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 19 / 42

slide-43
SLIDE 43

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Upper memory bound: SCTs for MEGs

1-exp. 2-exp.

w : E → {−1, 0, 1}k l = 2(d−1)·|S| · (|S| + 1)c·k2 ⇓ w : E → Zk, W max absolute weight, l = 2(d−1)·|S| · (W · |S| + 1)c·k2 ⇓ Width bounded by L = dl Refined approach: no blow-up in exponent as branching is preserved.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 19 / 42

slide-44
SLIDE 44

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Upper memory bound: SCTs for MEGs

1-exp. 1-exp.

w : E → {−1, 0, 1}k l = 2(d−1)·|S| · (|S| + 1)c·k2 ⇓ w : E → Zk, W max absolute weight, l = 2(d−1)·|S| · (W · |S| + 1)c·k2 ⇓ Width bounded by L = 2 · l · W + k − 1 k − 1

  • Refined approach: merge equivalent nodes, width is bounded by

number of incomparable labels (see next slide).

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 19 / 42

slide-45
SLIDE 45

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Upper memory bound: SCTs for MEGs

1-exp. 1-exp.

w : E → {−1, 0, 1}k l = 2(d−1)·|S| · (|S| + 1)c·k2 ⇓ w : E → Zk, W max absolute weight, l = 2(d−1)·|S| · (W · |S| + 1)c·k2 ⇓ Width bounded by L = 2 · l · W + k − 1 k − 1

  • Refined approach: overall, single exp. memory ≤ L · l.
  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 19 / 42

slide-46
SLIDE 46

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Upper memory bound: merging nodes in SCTs

Key idea to reduce width to single exp.

⊲ P1 only cares about the energy level. ⊲ If he can win with energy v, he can win with energy ≥ v.

s0 s1 s2 s3 s4 s5 (−1, 1) (0, 2) (0, 1) (0, 0) (1, −1) (−2, 1) (0, −1) (2, 0)

s0, (0, 0) s1, (−1, 1) s2, (0, 2) s3, (−1, 2) s3, (0, 2) s4, (0, 1) s5, (−2, 3) s0, (0, 0) s3, (0, 3)

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 20 / 42

slide-47
SLIDE 47

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Upper memory bound

Theorem: The size of memory needed for a finite-memory winning strategy in an energy game G = (S1, S2, sinit, E, k, w) is upper bounded by an exponential memSize(|S|, k, d, W ) = l · |S| · 2 · l · W + k − 1 k − 1

  • ,

with l = 2(d−1)·|S| · (W · |S| + 1)c·k2, d the branching degree of the game, W the largest weight on any edge and c a constant independant of the game. Note that given l, it is easy to see that the needed initial credit is bounded by l · W .

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 21 / 42

slide-48
SLIDE 48

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Lower memory bound

Theorem: There exists a family of games (G(K))K≥1, = (S1, S2, sinit, E, k = 2 · K, w) such that for any initial credit, P1 needs exponential memory to win.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 22 / 42

slide-49
SLIDE 49

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Lower memory bound

s1 s1,L s1,R sK sK,L sK,R t1 t1,L t1,R tK tK,L tK,R

∀ 1 ≤ i ≤ K, w((◦, si)) = w((◦, ti)) = (0, . . . , 0), w((si, si,L)) = −w((si, si,R)) = w((ti, ti,L)) = −w((ti, ti,R)), ∀ 1 ≤ j ≤ k, w((si, si,L))(j) =      = 1 if j = 2 · i − 1 = −1 if j = 2 · i = 0 otherwise .

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 23 / 42

slide-50
SLIDE 50

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Lower memory bound

s1 s1,L s1,R sK sK,L sK,R t1 t1,L t1,R tK tK,L tK,R

If P1 plays according to a Moore machine with less than 2K states, he takes the same decision in some state tx for the two highlighted prefixes (let x = K w.n.l.o.g.). ⇒ P2 can alternate and enforce decrease by 1 every two visits ⇒ P1 loses for any v0 ∈ Nk.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 23 / 42

slide-51
SLIDE 51

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Symbolic strategy synthesis algorithm

Theorem: Let G = (S1, S2, sinit, E, k, w) be a multi energy game. If Player 1 has a winning strategy in G, a Moore machine whose size is at most exponential in G can be constructed in time bounded by an exponential in G.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 24 / 42

slide-52
SLIDE 52

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Symbolic strategy synthesis algorithm

Theorem: Let G = (S1, S2, sinit, E, k, w) be a multi energy game. If Player 1 has a winning strategy in G, a Moore machine whose size is at most exponential in G can be constructed in time bounded by an exponential in G. Idea: greatest fixed point of a CpreC operator. ⊲ Exponential bound on the size of manipulated sets (∼ width). ⊲ Exponential bound on the number of iterations if a winning strategy exists (∼ depth).

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 24 / 42

slide-53
SLIDE 53

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Symbolic strategy synthesis algorithm: Cpre

C = l · W ∈ N, U(C) = (S1 ∪ S2) × [0..C]k, U(C) = 2U(C), the powerset of U(C), CpreC : U(C) → U(C), CpreC(V ) =

{(s1, e1) ∈ U(C) | s1 ∈ S1 ∧ ∃(s1, s) ∈ E, ∃(s, e2) ∈ V : e2 ≤ e1 + w(s1, s)} ∪ {(s2, e2) ∈ U(C) | s2 ∈ S2 ∧ ∀(s2, s) ∈ E, ∃(s, e1) ∈ V : e1 ≤ e2 + w(s2, s)}

⊲ Intuitively, compute for each state the sets of winning initial credits, represented by minimal elements of upper closed sets.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 25 / 42

slide-54
SLIDE 54

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Symbolic strategy synthesis algorithm: Cpre

C C

s s′ s′′

P1 can win for energy levels in the upper closed sets.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 26 / 42

slide-55
SLIDE 55

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Symbolic strategy synthesis algorithm: Cpre

C C

s s′ s′′

P1 can win for energy levels in the upper closed sets.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 26 / 42

slide-56
SLIDE 56

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Symbolic strategy synthesis algorithm: Cpre

C C

s s′ s′′

P1 can win for energy levels in the upper closed sets.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 26 / 42

slide-57
SLIDE 57

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Symbolic strategy synthesis algorithm: Cpre

C C

s s′ s′′

P1 can win for energy levels in the upper closed sets.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 26 / 42

slide-58
SLIDE 58

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Symbolic strategy synthesis algorithm: Cpre

C C

s s′ s′′

P1 can win for energy levels in the upper closed sets.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 26 / 42

slide-59
SLIDE 59

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Symbolic strategy synthesis algorithm: Cpre

Lemma: Let G = (S1, S2, sinit, E, k, w) be a multi energy game, in which all absolute values of weights are bounded by W , if Player 1 has a winning strategy in G and T = (Q, R) is a self-covering tree for G of depth l, then (sinit, C) ∈ Cpre∗

C for C = W · l.

Lemma: Let G = (S1, S2, sinit, E, k, w) be a multi energy game, let C ∈ N, if there exists c ∈ N such that (sinit, c) ∈ Cpre∗

C then

Player 1 has a winning strategy in G for initial credit c and a memory used by Player 1 can be bounded by |Min(CpreC)|.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 27 / 42

slide-60
SLIDE 60

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Symbolic strategy synthesis algorithm: Cpre

Lemma: Let G = (S1, S2, sinit, E, k, w) be a multi energy game, in which all absolute values of weights are bounded by W , if Player 1 has a winning strategy in G and T = (Q, R) is a self-covering tree for G of depth l, then (sinit, C) ∈ Cpre∗

C for C = W · l.

Lemma: Let G = (S1, S2, sinit, E, k, w) be a multi energy game, let C ∈ N, if there exists c ∈ N such that (sinit, c) ∈ Cpre∗

C then

Player 1 has a winning strategy in G for initial credit c and a memory used by Player 1 can be bounded by |Min(CpreC)|. Incremental approach can be used, by increasing the value of C inch by inch. Efficient implementation seems within reach.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 27 / 42

slide-61
SLIDE 61

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Corollary for MMPGs and summary

Corollary (thanks to [CDHR10]): Exponential memory is both sufficient and, in general, necessary for finite-memory winning on

  • MMPGs. Finite-memory strategy synthesis is in EXPTIME.

MEGs MMPGs

  • ptimal

finite-memory optimal

  • ptimal

Memory exp. exp. infinite [CDHR10] Synthesis EXPTIME EXPTIME /

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 28 / 42

slide-62
SLIDE 62

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

1 Classical energy and mean-payoff games 2 Extensions to multi-dimensions and parity 3 Strategy synthesis 4 Randomization as a substitute to finite-memory 5 Conclusion and ongoing work

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 29 / 42

slide-63
SLIDE 63

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Obtained results

Question: when and how can P1 trade his pure finite-memory strategy for an equally powerful randomized memoryless one ?

MEGs EPGs MMPGs MPBGs MPPGs 1-player × × √ √ √ (conj.) 2-player × × × √ √ (conj.)

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 30 / 42

slide-64
SLIDE 64

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Obtained results

Question: when and how can P1 trade his pure finite-memory strategy for an equally powerful randomized memoryless one ?

MEGs EPGs MMPGs MPBGs MPPGs 1-player × × √ √ √ √ √ (conj.) 2-player × × × √ √ √ √ (conj.)

⇒ Mean-payoff B¨ uchi games.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 30 / 42

slide-65
SLIDE 65

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Probabilistic semantics

B¨ uchi: sure almost-sure.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 31 / 42

slide-66
SLIDE 66

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Probabilistic semantics

B¨ uchi: sure almost-sure. Mean-payoff:

⊲ α-satisfaction. Given α ∈ [0, 1], ∀ λ2 ∈ Λ2, Pλ1,λ2

sinit

(MP≥v) ≥ α. ⊲ β-expectation. Given β ∈ Qk, ∀ λ2 ∈ Λ2, Eλ1,λ2

sinit

(MP) ≥ β. ⊲ 1-satisfaction of MP≥v ⇒ v-expectation for MP.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 31 / 42

slide-67
SLIDE 67

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Probabilistic semantics

B¨ uchi: sure almost-sure. Mean-payoff:

⊲ α-satisfaction. Given α ∈ [0, 1], ∀ λ2 ∈ Λ2, Pλ1,λ2

sinit

(MP≥v) ≥ α. ⊲ β-expectation. Given β ∈ Qk, ∀ λ2 ∈ Λ2, Eλ1,λ2

sinit

(MP) ≥ β. ⊲ 1-satisfaction of MP≥v ⇒ v-expectation for MP.

⇒ Almost-sure semantics.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 31 / 42

slide-68
SLIDE 68

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Mean-payoff B¨ uchi games

  • Remark. MPBGs require infinite memory for optimality.

s0 s1 −1 ⊲ P1 has to delay his visits of s1 for longer and longer intervals.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 32 / 42

slide-69
SLIDE 69

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Mean-payoff B¨ uchi games

  • Remark. MPBGs require infinite memory for optimality.

s0 s1 −1 ⊲ P1 has to delay his visits of s1 for longer and longer intervals. Theorem: In MPBGs, ε-optimality can be achieved using randomized memoryless strategies, both for satisfaction and expectation semantics.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 32 / 42

slide-70
SLIDE 70

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

MPBGs: sketch of proof

s0 s1 −1

1 Let G = (S1, S2, sinit, E, w, F), with F the set of B¨

uchi states. Let n = |S|. Let Win be the set of winning states for the MPB objective with threshold 0 (w.n.l.o.g.). For all s ∈ Win, P1 has two uniform memoryless strategies λgfe

1

and λ♦F

1

s.t.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 33 / 42

slide-71
SLIDE 71

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

MPBGs: sketch of proof

s0 s1 −1

1 Let G = (S1, S2, sinit, E, w, F), with F the set of B¨

uchi states. Let n = |S|. Let Win be the set of winning states for the MPB objective with threshold 0 (w.n.l.o.g.). For all s ∈ Win, P1 has two uniform memoryless strategies λgfe

1

and λ♦F

1

s.t.

λgfe

1

ensures that any cycle of its outcome have MP ≥ 0 [CD10],

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 33 / 42

slide-72
SLIDE 72

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

MPBGs: sketch of proof

s0 s1 −1 −1

1 Let G = (S1, S2, sinit, E, w, F), with F the set of B¨

uchi states. Let n = |S|. Let Win be the set of winning states for the MPB objective with threshold 0 (w.n.l.o.g.). For all s ∈ Win, P1 has two uniform memoryless strategies λgfe

1

and λ♦F

1

s.t.

λgfe

1

ensures that any cycle of its outcome have MP ≥ 0 [CD10], λ♦F

1

ensures reaching F in at most n steps, while staying in Win.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 33 / 42

slide-73
SLIDE 73

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

MPBGs: sketch of proof

2 For ε > 0, we build a pure finite-memory λpf 1 s.t.

(a) it plays λgfe

1

for 2 · W · n ε − n steps, then (b) it plays λ♦F

1

for n steps, then again (a).

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 34 / 42

slide-74
SLIDE 74

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

MPBGs: sketch of proof

2 For ε > 0, we build a pure finite-memory λpf 1 s.t.

(a) it plays λgfe

1

for 2 · W · n ε − n steps, then (b) it plays λ♦F

1

for n steps, then again (a).

This ensures that

⊲ F is visited infinitely often, ⊲ the total cost of phases (a) + (b) is bounded by −2 · W · n, and thus the mean-payoff is at least −ε.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 34 / 42

slide-75
SLIDE 75

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

MPBGs: sketch of proof

3 Based on λpf 1 , we build a randomized memoryless strategy λrm 1

s.t. in each state,

(a) it plays as λgfe

1

with probability at least 1 − ε 2 · W · n, (b) it plays as λ♦F

1

with the remaining probability.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 35 / 42

slide-76
SLIDE 76

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

MPBGs: sketch of proof

3 Based on λpf 1 , we build a randomized memoryless strategy λrm 1

s.t. in each state,

(a) it plays as λgfe

1

with probability at least 1 − ε 2 · W · n, (b) it plays as λ♦F

1

with the remaining probability.

B¨ uchi

⊲ Probability of playing as λ♦F

1

for n steps in a row and ensuring visit of F strictly positive at all times. ⊲ Thus λrm

1

almost-sure winning for the B¨ uchi objective.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 35 / 42

slide-77
SLIDE 77

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

MPBGs: sketch of proof

Mean-payoff ⊲ Long-term frequencies of transitions within a given state maintained. ⊲ P2 may use the same strategy on the graph induced by λpf

1

and the MDP induced by λrm

1

to achieve the same overall transition probabilities. ⊲ Achieving plays π with MP(π) < −ε with strictly positive probability on the MDP would induce that P2 can enforce such a play on the graph and lead to contradiction. ⊲ Thus λrm

1

almost-sure winning for the MP objective with threshold −ε.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 36 / 42

slide-78
SLIDE 78

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Summary

MEGs EPGs MMPGs MPBGs MPPGs 1-player × × √ √ √ (conj.) 2-player × × × √ √ (conj.)

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 37 / 42

slide-79
SLIDE 79

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

1 Classical energy and mean-payoff games 2 Extensions to multi-dimensions and parity 3 Strategy synthesis 4 Randomization as a substitute to finite-memory 5 Conclusion and ongoing work

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 38 / 42

slide-80
SLIDE 80

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Conclusion

Quantitative objectives Restriction to finite-memory (practical interest) Exponential memory bounds EXPTIME synthesis Randomness instead of memory

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 39 / 42

slide-81
SLIDE 81

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Results Overview

Strategy synthesis

MEGs MMPGs

  • ptimal

finite-memory optimal

  • ptimal

Memory exp. exp. infinite [CDHR10] Synthesis EXPTIME EXPTIME /

Randomness as a substitute for finite-memory

MEGs EPGs MMPGs MPBGs MPPGs 1-player × × √ √ √ (conj.) 2-player × × × √ √ (conj.)

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 40 / 42

slide-82
SLIDE 82

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Ongoing and future work

Extend results on MEGs/MMPGs to MEPGs/MMPPGs. Consider alternative, more natural definition of MP-like

  • bjective, with good synthesis properties.
  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 41 / 42

slide-83
SLIDE 83

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

  • Thanks. Questions ?
  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 42 / 42

slide-84
SLIDE 84

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Roderick Bloem, Krishnendu Chatterjee, Thomas A. Henzinger, and Barbara Jobstmann. Better quality in synthesis through quantitative objectives. In Proc. of CAV, volume 5643 of LNCS, pages 140–156. Springer, 2009. Patricia Bouyer, Ulrich Fahrenberg, Kim Guldstrand Larsen, Nicolas Markey, and Jir´ ı Srba. Infinite runs in weighted timed automata with energy constraints. In Proc. of FORMATS, volume 5215 of LNCS, pages 33–47. Springer, 2008. Tom´ as Br´ azdil, Petr Jancar, and Anton´ ın Kucera. Reachability games on extended vector addition systems with states. In Proc. of ICALP, volume 6199 of LNCS, pages 478–489. Springer, 2010.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 42 / 42

slide-85
SLIDE 85

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Patricia Bouyer, Nicolas Markey, J¨

  • rg Olschewski, and Michael

Ummels. Measuring permissiveness in parity games: Mean-payoff parity games revisited. In Proc. of ATVA, volume 6996 of LNCS, pages 135–149. Springer, 2011. Krishnendu Chatterjee and Laurent Doyen. Energy parity games. In Proc. of ICALP, volume 6199 of LNCS, pages 599–610. Springer, 2010. Arindam Chakrabarti, Luca de Alfaro, Thomas A. Henzinger, and Mari¨ elle Stoelinga. Resource interfaces. In Proc. of EMSOFT, volume 2855 of LNCS, pages 117–133. Springer, 2003.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 42 / 42

slide-86
SLIDE 86

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

Krishnendu Chatterjee, Laurent Doyen, Thomas A. Henzinger, and Jean-Fran¸ cois Raskin. Generalized mean-payoff and energy games. In Proc. of FSTTCS, volume 8 of LIPIcs, pages 505–516. Schloss Dagstuhl - Leibniz-Zentrum fuer Informatik, 2010. Krishnendu Chatterjee, Thomas A. Henzinger, and Marcin Jurdzinski. Mean-payoff parity games. In Proc. of LICS, pages 178–187. IEEE Computer Society, 2005.

  • A. Ehrenfeucht and J. Mycielski.

Positional strategies for mean payoff games. International Journal of Game Theory, 8(2):109–113, 1979. Uli Fahrenberg, Line Juhl, Kim G. Larsen, and Jir´ ı Srba. Energy games in multiweighted automata.

  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 42 / 42

slide-87
SLIDE 87

EGs & MPGs Multi-dim. & parity Synthesis Randomization Conclusion

In Proc. of ICTAC, volume 6916 of LNCS, pages 95–115. Springer, 2011. T.M. Liggett and S.A. Lippman. Short notes: Stochastic games with perfect information and time average payoff. Siam Review, 11(4):604–607, 1969. Charles Rackoff. The covering and boundedness problems for vector addition systems.

  • Theor. Comput. Sci., 6:223–231, 1978.

Louis E. Rosier and Hsu-Chun Yen. A multiparameter analysis of the boundedness problem for vector addition systems.

  • J. Comput. Syst. Sci., 32(1):105–135, 1986.
  • Strat. Synth. for Quant. Obj.

Chatterjee, Randour, Raskin 42 / 42