temporal logics for multi agent systems
play

Temporal Logics for Multi-Agent Systems Tom Henzinger IST Austria - PowerPoint PPT Presentation

Temporal Logics for Multi-Agent Systems Tom Henzinger IST Austria Joint work with Rajeev Alur, Guy Avni, Krish Chatterjee, Luca de Alfaro, Orna Kupferman, and Nir Piterman. Shielded Control Plant Shield (discrete- Black-box Controller


  1. Temporal Logics for Multi-Agent Systems Tom Henzinger IST Austria Joint work with Rajeev Alur, Guy Avni, Krish Chatterjee, Luca de Alfaro, Orna Kupferman, and Nir Piterman.

  2. Shielded Control Plant Shield (discrete- Black-box Controller event) (e.g. data-driven, learned) Shield can ensure safety and fairness (temporal-logic specification), performance (quantitative spec), and/or incremental regimes.

  3. Multiple Agents (e.g. plant, controller, shield; robotics) A 1 : A 2 : bool x := 0 bool y := 0 loop loop choice choice | x := 0 | y := x | x := x+1 mod 2 | y := x+1 mod 2 end choice end choice end loop end loop Φ 1 : Φ 2 :  ( x ¸ y )  (y = 0)

  4. State Space as Graph 8  (x ¸ y) X 00 01 9  (x ¸ y)  10 11

  5. State Space as Graph 8  (x ¸ y) X 00 01 9  (x ¸ y)  10 11 hh A 1 ii  (x ¸ y) hh A 2 ii  (y = 0)

  6. State Space as Game 00 00 8  (x ¸ y) 00 X 9  (x ¸ y)  10 01 10 10 01 01 hh A 1 ii  (x ¸ y) X hh A 2 ii  (y = 0)  11 11 11

  7. State Space as Game 00 If A 2 keeps y = 0,  then A 1 can keep x ¸ y. 00 00 10 01 10 10 01 01 11 11 11

  8. Reactive Synthesis Agent Synthesis (a.k.a. discrete-event control) Given: agent A, specification Φ , and environment E Find: refinement A’ of A so that A’||E satisfies Φ Solution: A’ = winning strategy in game A against E for objective Φ

  9. Reactive Synthesis Agent Synthesis (a.k.a. discrete-event control) Given: agent A, specification Φ , and environment E Find: refinement A’ of A so that A’||E satisfies Φ Solution: A’ = winning strategy in game A against E for objective Φ Multi-Agent Synthesis (e.g. shielded or distributed control) Given: -two agents A 1 and A 2 -specifications Φ 1 and Φ 2 for A 1 and A 2 Find: refinements A’ 1 and A’ 2 of A 1 and A 2 so that A’ 1 ||A’ 2 ||S satisfies Φ 1 Æ Φ 2 for every fair scheduler S

  10. Mutual Exclusion while( true ) { while( true ) { flag[1] := true; turn := 2; flag[2] := true; turn :=1; choice choice | while( flag[1] ) nop; | while( flag[1] ) nop; | while( flag[2] ) nop; | while( flag[2] ) nop; | while( turn=1 ) nop; | while( turn=1 ) nop; | while( turn=2 ) nop; | while( turn=2 ) nop; | while( flag[1] & turn=2 ) nop; | while( flag[1] & turn=2 ) nop; | while( flag[1] & turn=1 ) nop; | while( flag[1] & turn=1 ) nop; | while( flag[2] & turn=1 ) nop; | while( flag[2] & turn=1 ) nop; | while( flag[2] & turn=2 ) nop; | while( flag[2] & turn=2 ) nop; end choice; end choice; CritSec; flag[1] := false; CritSec; flag[2] := false; nonCritSec; nonCritSec; } }

  11. Multi-Agent Synthesis Formulation 1 Do there exist refinements A’ 1 and A’ 2 so that [A’ 1 || A’ 2 || S] µ ( Φ 1 Æ Φ 2 ) for every fair scheduler S ? Solution: game A 1 ||A 2 against S for objective Φ 1 Æ Φ 2 Too weak (solution has A 1 and A 2 cooperate, e.g. alternate).

  12. Multi-Agent Synthesis Formulation 2 Do there exist refinements A’ 1 and A’ 2 so that 1. [A’ 1 || A 2 || S] µ Φ 1 2. [A 1 || A’ 2 || S] µ Φ 2 for every fair scheduler S ? Solution: two games A 1 against A 2 ||S for objective Φ 1 , and A 2 against A 1 ||S for objective Φ 2 Too strong (answer is NO, e.g. because agent may stay in CritSec).

  13. Multi-Agent Synthesis Formulation 3 Do there exist refinements A’ 1 and A’ 2 so that 1. [A’ 1 || A 2 || S] µ ( Φ 2 ) Φ 1 ) 2. [A 1 || A’ 2 || S] µ ( Φ 1 ) Φ 2 ) 3. [A’ 1 || A’ 2 || S] µ ( Φ 1 Æ Φ 2 ) for every fair scheduler S ?

  14. Mutual Exclusion while( true ) { while( true ) { flag[2] := true; turn := 1; flag[1] := true; turn := 2; while( flag[1] & turn=2 ) nop; while( flag[2] & turn=1 ) nop; CritSec; flag[2] := false; CritSec; flag[1] := false; nonCritSec; nonCritSec; } } Solution is exactly Peterson’s mutual-exclusion protocol.

  15. Games on Labeled Graphs system states nodes = observations node labels = state transitions edges = transition costs edge labels = agents players =

  16. Labeled Graph q1 a 3 1 q3 q2 b a 0 1 q5 q4 c a 1-agent system without uncertainty.

  17. Markov Decision Process q1 a 3 1 q3 q2 b a 0.4 0.6 q5 q4 c a 1-agent system with uncertainty.

  18. Labeled Graph x(q1) = q3 q1 x(q1,q3) = {q4: 0.4; q5: 0.6} a 3 1 q3 q2 b a 0 1 q5 q4 c a State q 2 Q } c (x)@q1 = 0.4 Strategy x: Q * ! D(Q) avg (x)@q1 = 0.8 x@q: probability space on Q !

  19. Markov Decision Process q1 x(q1) = q3 a 3 1 q3 q2 b a 0.4 0.6 q5 q4 c a State q 2 Q } c (x)@q1 = 0.4 Strategy x: Q * ! D(Q) avg (x)@q1 = 1 x@q: probability space on Q !

  20. Turn-based Game q1 a 3 1 q3 q2 b a 0 1 q5 q4 c a Asynchronous 2-agent system without uncertainty.

  21. Stochastic Game q1 a 3 1 q3 q2 b a 0.4 0.6 q5 a c q4 1 0 q7 q6 b c Asynchronous 2-agent system with uncertainty.

  22. Turn-based Game q1 x(q1) = q3 y(q1,q3) = {q4: 0.4; q5: 0.6} a 3 1 q3 q2 b a 0 1 q5 q4 c a State q 2 Q } c (x,y)@q1 = 0.4 Strategies x,y: Q * ! D(Q) avg (x,y)@q1 = 0.8 (x,y)@q: probability space on Q !

  23. Stochastic Game q1 x(q1) = q3 y(q1,q3,q4) = {q6: 0.4; q7: 0.6} a 3 1 q3 q2 b a 0.4 0.6 q5 a c q4 1 0 q7 q6 b c State q 2 Q Strategies x,y: Q * ! D(Q) } c (x,y)@q1 = 0.4 avg (x,y)@q1 = 0.92 (x,y)@q: probability space on Q !

  24. Concurrent Game Player Left moves: q1 Player Right moves: a {1,2} {1,2} 2,2 1,1 2,1 1,2 q3 q4 q5 q2 a b b c Synchronous 2-agent system without uncertainty.

  25. Concurrent Stochastic Game q1 a q3 q4 q5 q2 a b b c Player Column moves: 1 2 q1: {1,2} q2: 0.3 q2: 0.1 q3: 0.2 q3: 0.1 1 q4: 0.5 q4: 0.5 Player Row moves: q5: q5: 0.3 Matrix game {1,2} at each node. q2: q2: 1.0 q3: 0.2 q3: 2 q4: 0.1 q4: q5: 0.7 q5: Synchronous 2-agent system with uncertainty.

  26. Concurrent Game Player Left moves: q1 Player Right moves: a {1,2} {1,2} 2,2 1,1 2,1 1,2 q3 q4 q5 q2 a b b c x(q1) = 2 State q 2 Q y(q1) = {1: 0.4; 2: 0.6} Strategies x,y: Q * ! D(Moves) } c (x,y)@q1 = 0.6 (x,y)@q: probability space on Q !

  27. Concurrent Stochastic Game q1 a q3 q4 q5 q2 a b b c Player Column moves: 1 2 q1: {1,2} q2: 0.3 q2: 0.1 q3: 0.2 q3: 0.1 1 q4: 0.5 q4: 0.5 Player Row moves: q5: q5: 0.3 {1,2} q2: q2: 1.0 q3: 0.2 q3: 2 q4: 0.1 q4: q5: 0.7 q5: x(q1) = 2 State q 2 Q Strategies x,y: Q * ! y(q1) = {1: 0.4; 2: 0.6} D(Moves) (x,y)@q: probability space on Q ! } c (x,y)@q1 = 0.28

  28. Timed Games, Hybrid Games, etc.

  29. Strategy Logic 1. first-order quantification over sorted strategies 2. linear temporal formulas over observation sequences 3. interpreted over states q ² (9 x) (8 y) Á iff there exists a player-1 strategy x such that for all player-2 strategies y Á (x,y)@q = 1

  30. Alternating-Time Temporal Logic 1. path quantifiers over sets of players 2. linear temporal formulas over observation sequences 3. interpreted over states q ² hh Tii Á iff if the game starts from state q the players in set T can ensure that the LTL formula Á holds with probability 1

  31. Alternating-Time Temporal Logic 1. path quantifiers over sets of players 2. linear temporal formulas over observation sequences 3. interpreted over states q ² hh Tii Á iff if the game starts from state q the players in set T can ensure that the LTL formula Á holds with probability 1 hh;ii Á = 8 Á hh Uii Á = 9 Á where U is the set of all players [[T]] Á = : hh U\Tii : Á “the players in U\T cannot prevent Á”

  32. ATL * µ SL hh Tii Á = (9 x 1 ,…,x m 2 ¦ T ) (8 y 1 ,…,y n 2 ¦ U\T ) Á

  33. ATL * ( SL Player 1 can ensure Á 1 if player 2 ensures Á 2 : (9 x)(8 y) ( ((8 x’) Á 2 (x’,y)) ) Á 1 (x,y) )

  34. ATL * ( SL Player 1 can ensure Á 1 if player 2 ensures Á 2 : (9 x)(8 y) ( ((8 x’) Á 2 (x’,y)) ) Á 1 (x,y) ) The strategy x dominates all strategies w.r.t. objective Á: (8 x’)(8 y) ( Á(x’,y) ) Á(x,y) )

  35. ATL * ( SL Player 1 can ensure Á 1 if player 2 ensures Á 2 : (9 x)(8 y) ( ((8 x’) Á 2 (x’,y)) ) Á 1 (x,y) ) The strategy x dominates all strategies w.r.t. objective Á: (8 x’)(8 y) ( Á(x’,y) ) Á(x,y) ) The strategy profile (x,y) is a secure Nash equilibrium: (9 x)(9 y) ( (Á 1 ÆÁ 2 ) (x,y) Æ(8 y’) (Á 2 ) Á 1 ) (x,y’) Æ(8 x’) (Á 1 ) Á 2 ) (x’,y) )

  36. ATL ATL is the fragment of ATL * in which every temporal operator is preceded by a path quantifier: hh Tii ° a single-shot game hh Tii } b reachability game Tii ฀ c hh safety game

  37. ATL ATL is the fragment of ATL * in which every temporal operator is preceded by a path quantifier: hh Tii ° a single-shot game hh Tii } b reachability game Tii ฀ c hh safety game Not in ATL: hh T ii ฀ } c Buchi game hh T ii Á ! -regular (parity) game

  38. Pure Winning L,R R,L Player 1: Player 2: miss hit {moveL,moveR} {throwL,throwR} L,L R,R hh P2ii pure } hit X hh P2ii } hit  Player 2 needs randomness to win.

  39. Limit Winning W,W R,T hit Player 1: Player 2: wait {Wait,Run} {Wait,Throw} home R,W W,T hh P1ii } home X hh P1ii limit } home  Player 1 can win with probability arbitrarily close to 1.

  40. Quantitative ATL hh P1ii Á = (9 x) (8 y) ( Á(x,y) = 1 ) hh P1ii limit Á = ( sup x inf y Á(x,y) ) = 1

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend