STV: Model Checking for Strategies under Imperfect Information - - PowerPoint PPT Presentation

stv model checking for strategies under imperfect
SMART_READER_LITE
LIVE PREVIEW

STV: Model Checking for Strategies under Imperfect Information - - PowerPoint PPT Presentation

STV: Model Checking for Strategies under Imperfect Information Damian Kurpiewski Institute of Computer Science Polish Academy of Sciences (joint work with Wojtek Jamroga and Micha Knapik) LAMAS, 09/05/2020 Outline 2 / 1 Outline 3 / 1


slide-1
SLIDE 1

STV: Model Checking for Strategies under Imperfect Information

Damian Kurpiewski

Institute of Computer Science Polish Academy of Sciences (joint work with Wojtek Jamroga and Michał Knapik)

LAMAS, 09/05/2020

slide-2
SLIDE 2

Outline

2 / 1

slide-3
SLIDE 3

Outline

3 / 1

slide-4
SLIDE 4

ATL: What Agents Can Achieve

  • ATL: Alternating-time Temporal Logic [Alur et al. 1997-2002]
  • Temporal logic meets game theory
  • Main idea: cooperation modalities
  • A

Φ: coalition A has a collective strategy to enforce Φ ❀ Φ can include temporal operators: X (next), F (sometime in the future), G (always in the future), U (strong until)

4 / 1

slide-5
SLIDE 5

ATL with incomplete information

  • Imperfect information (q ∼a q′)

5 / 1

slide-6
SLIDE 6

ATL with incomplete information

  • Imperfect information (q ∼a q′)
  • Imperfect recall - agent memory coded within state of the model

5 / 1

slide-7
SLIDE 7

ATL with incomplete information

  • Imperfect information (q ∼a q′)
  • Imperfect recall - agent memory coded within state of the model
  • Uniform strategies - specify same choices for indistinguishable states:

q ∼a q′ = ⇒ sa(q) = sa(q′)

5 / 1

slide-8
SLIDE 8

ATL with incomplete information

  • Imperfect information (q ∼a q′)
  • Imperfect recall - agent memory coded within state of the model
  • Uniform strategies - specify same choices for indistinguishable states:

q ∼a q′ = ⇒ sa(q) = sa(q′)

  • Fixpoint equivalences do not hold anymore

5 / 1

slide-9
SLIDE 9

ATL with incomplete information

  • Imperfect information (q ∼a q′)
  • Imperfect recall - agent memory coded within state of the model
  • Uniform strategies - specify same choices for indistinguishable states:

q ∼a q′ = ⇒ sa(q) = sa(q′)

  • Fixpoint equivalences do not hold anymore
  • Model checking ATLir is ∆p

2-complete

5 / 1

slide-10
SLIDE 10

Example - Simple Model of Voting and Coercion

q0 q1

votei,1

q2

votei,2

q3

votei,1

q4

votei,1

q5

votei,2

q6

votei,2

q7

finishi votei,1

q8

finishi votei,1 puni

q9

finishi votei,1

q10

finishi votei,1 puni

q11

finishi votei,2 puni

q12

finishi votei,2

q13

finishi votei,2 puni

q14

finishi votei,2

(vote1, −) (vote2, −) ( g i v e , − ) (ng, −) ( g i v e , − ) (ng, −) ( − , n p ) (−, pun) ( − , n p ) (−, pun) ( − , n p ) (−, pun) ( − , n p ) (−, pun)

c c c c

(wait, −) (wait, −) (wait, −)

6 / 1

slide-11
SLIDE 11

Example Formulae

coercer F(¬pun1 ∨ vote1,1): “Coercer can coerce Voter to vote for first candidate”

7 / 1

slide-12
SLIDE 12

Example Formulae

coercer F(¬pun1 ∨ vote1,1): “Coercer can coerce Voter to vote for first candidate” FALSE

7 / 1

slide-13
SLIDE 13

Example Formulae

coercer F(¬pun1 ∨ vote1,1): “Coercer can coerce Voter to vote for first candidate” FALSE

voter1 G(¬pun1 ∧ ¬vote1,1): “Voter can avoid punishment without voting for first candidate”

7 / 1

slide-14
SLIDE 14

Example Formulae

coercer F(¬pun1 ∨ vote1,1): “Coercer can coerce Voter to vote for first candidate” FALSE

voter1 G(¬pun1 ∧ ¬vote1,1): “Voter can avoid punishment without voting for first candidate” TRUE

7 / 1

slide-15
SLIDE 15

Outline

8 / 1

slide-16
SLIDE 16

Approximate Verification of Strategic Ability M | =ir ϕ : DIFFICULT!

9 / 1

slide-17
SLIDE 17

Approximate Verification of Strategic Ability M | =ir ϕ : DIFFICULT!

M | =ir ϕ M | = LB(ϕ) M | = UB(ϕ)

slide-18
SLIDE 18

Approximate Verification of Strategic Ability M | =ir ϕ : DIFFICULT!

M | =ir ϕ M | = LB(ϕ) M | = UB(ϕ) perfect information

slide-19
SLIDE 19

Approximate Verification of Strategic Ability M | =ir ϕ : DIFFICULT!

M | =ir ϕ M | = LB(ϕ) M | = UB(ϕ) perfect information

  • ur contribution
slide-20
SLIDE 20

Approximate Verification of Strategic Ability M | =ir ϕ : DIFFICULT!

M | =ir ϕ M | = LB(ϕ) M | = UB(ϕ) perfect information

  • ur contribution

9 / 1

slide-21
SLIDE 21

Domino DFS

q0 start q2 q1 q4 p q3 p q5 ¬p (A, U) (B, ⋆) (A, V) (A, V) (A, U) (B, ⋆) (B, ⋆) (A, U) (A, V) 1

10 / 1

slide-22
SLIDE 22

Domino DFS

q0 start q2 q1 q4 p q3 p q5 ¬p (A, U) (B, ⋆) (A, V) (A, V) (A, U) (B, ⋆) (B, ⋆) (A, U) (A, V) 1

11 / 1

slide-23
SLIDE 23

Domino DFS

q0 start q2 q1 q4 p q3 p q5 ¬p (A, U) (B, ⋆) (A, V) (A, V) (A, U) (B, ⋆) (B, ⋆) (A, U) (A, V) 1

12 / 1

slide-24
SLIDE 24

Domino DFS

q0 start q2 q1 q4 p q3 p q5 ¬p (A, U) (B, ⋆) (A, V) (A, V) (A, U) (B, ⋆) (B, ⋆) (A, U) (A, V) 1

13 / 1

slide-25
SLIDE 25

Domino DFS

q0 start q2 q1 q4 p q3 p q5 ¬p (A, U) (B, ⋆) (A, V) (A, V) (A, U) (B, ⋆) (B, ⋆) (A, U) (A, V) 1

14 / 1

slide-26
SLIDE 26

Outline

15 / 1

slide-27
SLIDE 27

Implemented models

  • Bridge scenario
  • Castles
  • TianJi
  • Drones
  • Simple Voting

16 / 1

slide-28
SLIDE 28

Bridge scenario

17 / 1

slide-29
SLIDE 29

Bridge scenario

  • Typical bridge play scenario, modified by two variables: n, k
  • Each player holds k cards in hand
  • Deck consists of 4n cards in total
  • We consider only endplay
  • Random deal
  • Four players - S, W, N, E
  • Declarer (S) handles his own cards and the ones of the dummy (N)
  • Players remember already played cards
  • Everyone see dummy cards
  • NoTrump contract

18 / 1

slide-30
SLIDE 30

DEMO

19 / 1

slide-31
SLIDE 31

Experimental results - Bridge scenario

Conf. DominoDFS MCMAS Approx.

  • Approx. opt.

(1, 1) 0.0006 0.12 0.0008 < 0.0001 (2, 2) 0.01 8712∗ 0.01 < 0.0001 (3, 3) 0.8 timeout 0.8 0.06 (4, 4) 160 timeout 384 5.5 (5, 5)∗ 1373 timeout 8951 39 (5, 5) memout timeout memout 138 (6, 6)∗ memout timeout memout 4524

20 / 1

slide-32
SLIDE 32

Experimental results - Castles

Conf. DominoDFS MCMAS SMC (1, 1, 1) 0.3 65 63 (2, 1, 1) 1.5 12898 184 (3, 1, 1) 25 timeout 6731 (2, 2, 1) 25 timeout 4923 (2, 2, 2) 160 timeout timeout (3, 2, 2) 2688 timeout timeout (3, 3, 2) timeout timeout timeout

21 / 1

slide-33
SLIDE 33

THANK YOU

22 / 1