PDL as a Multi-Agent Strategy Logic Jan van Eijck CWI & ILLC, - - PowerPoint PPT Presentation
PDL as a Multi-Agent Strategy Logic Jan van Eijck CWI & ILLC, - - PowerPoint PPT Presentation
PDL as a Multi-Agent Strategy Logic Jan van Eijck CWI & ILLC, Amsterdam September 17, 2012 Abstract We propose a new perspective on PDL as a multi-agent strategic logic (MASL). This logic for strategic reasoning has group strategies as
Overview
- The Pebble Puzzle
- Reasoning About Programs
- Reasoning About Actions
- Strategic Games: the Prisoner’s Dilemma
- Group Strategies in Games
- Key Notions: Best Response, Nash Equilibrium
- Voting as a Multi-Agent Game
- MASL: Language and Expressiveness
- Soundness and Completeness
- Connections, Further Work
The Pebble Puzzle An urn cointains 70 pebbles; 35 of them are white and 35 are black. There is a pile of black pebbles available outside the urn. Pebble Algorithm
- While there are still enough pebbles in the urn:
– pick two pebbles; if they have the same colour, put back a black pebble
- therwise, put back the white pebble.
In every step of the algorithm one pebble gets removed. After 69 steps, there is one pebble left. What is its colour?
module Pebbles where data Color = W | B deriving (Eq,Show) drawPebble :: [Color] -> [Color] drawPebble [] = [] drawPebble [x] = [x] drawPebble (W:W:xs) = drawPebble (B:xs) drawPebble (B:B:xs) = drawPebble (B:xs) drawPebble (W:B:xs) = drawPebble (W:xs) drawPebble (B:W:xs) = drawPebble (W:xs) numberW :: [Color] -> Int numberW = length . (filter (\x -> x == W)) parityW :: [Color] -> Int parityW xs = mod (numberW xs) 2 prop_invariant = \xs -> parityW xs == parityW (drawPebble xs)
Sir Tony Hoare
Formal Specification With Hoare Triples In general a triple initial state – statement – final state {P} S {Q} has the following operational meaning: If execution of S in a state that satisfies P terminates, then the termination state is guaranteed to satisfy Q. Such triples {P} S {Q} are called Hoare triples after Tony Hoare. The predicate for the initial state is called the precondition, and the predicate for the final state is called the postcondition.
assignment {ϕv
a} v := a {ϕ}
skip {ϕ} SKIP {ϕ} sequence {ϕ} C1 {ψ} {ψ} C2 {χ} {ϕ} C1; C2 {χ} conditional choice {ϕ ∧ B} C1 {ψ} {ϕ ∧ ¬B} C2 {ψ} {ϕ} if B then C1 else C2 {ψ} guarded iteration {ϕ ∧ B} C {ϕ} {ϕ} while B do C {ϕ ∧ ¬B} precondition strengthening N | = ϕ′ → ϕ {ϕ} C {ψ} {ϕ′} C {ψ} postcondition weakening {ϕ} C {ψ} N | = ψ → ψ′ {ϕ} C {ψ′}
Vaughan Pratt
Hoare Logic as a Fragment of Dynamic Logic Hoare logic is a fragment of a more general system of (propositional) dynamic logic. The language of propositional dynamic logic was defined by Pratt in [13, 14] as a generic language for reasoning about computation. Ax- iomatisations were given independently by Segerberg [16], Fisher/Ladner [8], and Parikh [10]. These axiomatisations make the connection be- tween propositional dynamic logic and modal logic very clear.
PDL Language Let p range over the set of basic propositions P, and let a range over a set of basic actions A. Then the formulae ϕ and programs α of propositional dynamic logic are given by: ϕ ::= ⊤ | p | ¬ϕ | ϕ1 ∨ ϕ2 | αϕ α ::= a |?ϕ | α1; α2 | α1 ∪ α2 | α∗ Abbreviation: [α]ϕ abbreviates¬α¬ϕ.
Expressing Hoare Triples in PDL Floyd-Hoare correctness assertions are expressible in PDL, as fol-
- lows. If ϕ, ψ are PDL formulae and α is a PDL program, then
{ϕ} α {ψ} translates into ϕ → [α]ψ. Clearly, {ϕ} α {ψ} holds in a state in a model iff ϕ → [α]ψ is true in that state in that model.
PDL Axiomatisation Aioms are all propositional tautologies, plus the following axioms (we give box ([α])versions here, but every axiom has an equivalent diamond (α) version): (K) ⊢ [α](ϕ → ψ) → ([α]ϕ → [α]ψ) (test) ⊢ [?ϕ1]ϕ2 ↔ (ϕ1 → ϕ2) (sequence) ⊢ [α1; α2]ϕ ↔ [α1][α2]ϕ (choice) ⊢ [α1 ∪ α2]ϕ ↔ [α1]ϕ ∧ [α2]ϕ (mix) ⊢ [α∗]ϕ ↔ ϕ ∧ [α][α∗]ϕ (induction) ⊢ (ϕ ∧ [α∗](ϕ → [α]ϕ)) → [α∗]ϕ and the following rules of inference: (modus ponens) From ⊢ ϕ1 and ⊢ ϕ1 → ϕ2, infer ⊢ ϕ2. (modal generalisation) From ⊢ ϕ, infer ⊢ [α]ϕ.
The Loop Invariance Rule In the presence of the other axioms, the induction axiom is equivalent to the loop invariance rule: ϕ → [α]ϕ ϕ → [α∗]ϕ
Deriving Hoare Rules in PDL The Floyd-Hoare inference rules can now be derived in PDL. As an example we derive the rule for guarded iteration: {ϕ ∧ ψ} α {ψ} {ψ} WHILE ϕ DO α {¬ϕ ∧ ψ} Let the premise {ϕ ∧ ψ} α {ψ} be given, i.e. assume (1). ⊢ (ϕ ∧ ψ) → [α]ψ. (1) We wish to derive the conclusion ⊢ {ψ} WHILE ϕ DO α {¬ϕ ∧ ψ}, i.e. we wish to derive (2). ⊢ ψ → [(?ϕ; α)∗; ?¬ϕ](¬ϕ ∧ ψ). (2)
From (1) by means of propositional reasoning: ⊢ ψ → (ϕ → [α]ψ). From this, by means of the test and sequence axioms: ⊢ ψ → [?ϕ; α]ψ. Applying the loop invariance rule gives: ⊢ ψ → [(?ϕ; α)∗]ψ. Since ψ is propositionally equivalent with ¬ϕ → (¬ϕ ∧ ψ), we get from this by propositional reasoning: ⊢ ψ → [(?ϕ; α)∗](¬ϕ → (¬ϕ ∧ ψ)). The test axiom and the sequencing axiom yield the desired result (2).
Strategic Games: The Prisoner’s Dilemma cooperate defect cooperate c, c c, d defect d, c d, d With output function o : {c, d}2 → {x, y, z, u}2: cooperate defect cooperate x, x y, z defect z, y u, u Fixing the preferences of the players: z > x > u > y. With numerical utilities: cooperate defect cooperate 2, 2 0, 3 defect 3, 0 1, 1
Group Strategies in PD Game are the Strategy Profiles cc cd dc dd cc cd cd cc dc dd dd dc dc cc dd cd
Key Notions: Best Response Let (s′
i, s−i) be the strategy profile that is like s for all players except
i, but has si replaced by s′
- i. A strategy si is a best response in s if
∀s′
i ∈ Si ui(s) ≥ ui(s′ i, s−i).
Example in PD game. Let s = (d, c). The first player defects, the second player cooperates. Is d a best response for player 1 in (d, c)? Yes, because (d, c) gives payoff 3 for player 1, while the alternative (c, c) only gives payoff 2. So player 1 cannot do better than play d.
John Nash
Key Notions: Pure Nash Equilibrium A strategy profile s is a (pure) Nash equilibrium if each si is a best response in s: ∀i ∈ N ∀s′
i ∈ Si ui(s) ≥ ui(s′ i, s−i).
A game G is Nash if G has a (pure) Nash equilibrium. (d, d) is a Nash equilibrium for the PD game, so the PD game is Nash.
Charles Dodgson, also known as Lewis Carroll
Voting as a Multi-Agent Game Voting can be seen as a form of multi-agent decision making, with the voters as agents [7]. Voting is the process of selecting an item or a set of items from a finite set A of alternatives, on the basis of the stated preferences of a set of voters. We assume that the preferences of a voter are represented by a ballot: a linear ordering of A. Let ord(A) be the set of all ballots on A. If there are three alternatives a, b, c, and a voter prefers a over b and b
- ver c, then her ballot is abc.
Example
- Assume there are three voters {1, 2, 3}.
- Assume there are three alternatives {a, b, c}.
- Then profiles are vectors of ballots.
- Example profile where the first voter has ballot abc, the second
voter has ballot abc, the third voter has ballot bca, and so on: (abc, abc, bca).
Voting Rules A voting rule V for set of alternatives A is a function from A-profiles to P+(A) (the set of non-empty subsets of A). If V (P) = B, then the members of B are called the winners of P under V . A voting rule is resolute if V (P) is a singleton for any profile P. Example voting rule: voting by absolute majority. Selects an alternative with more than 50 % of the votes as winner, and returns the whole set of alternatives otherwise. (abc, abc, bca). Absolute majority selects a as winner, for a has two votes, b has one.
Strategizing in Voting: Gibbard-Satterthwaite Ballot Profile Vector of ballots. Resolute Voting Rule Function V from ballot profiles to alternatives. P ∼i P′ P and P′ differ at most in the ballot for i. Strategy-Proofness V is strategy-proof if P ∼i P′ implies that, from the perspective of P, V (P) is at least as good for i as V (P′). (Weak) Non-Imposition V has at least three possible outcomes. Dictatorship V is a dictatorship if there is some voter k such that V maps any profile to the top-ranking alternative in the k-ballot. GS Theorem Any resolute voting rule that is strategy-proof and weakly non-imposed is a dictatorship.
Group Actions in Voting Games Voters: 1, 2, 3. Alternatives: a, b, c. Voting rule: absolute majority. 1 is the row player, 2 the column player, and 3 the table player
a: a b c a a a a b a b a, b, c c a a, b, c c b: a b c a a b a, b, c b b b b c a, b, c b c c: a b c a a a, b, c c b a, b, c b c c c c c
Payoffs
- Assume ballot abc.
- Then a > b > c, and plausibly a > {a, b, c} > c.
- Assume {a, b, c} and b give the same payoff.
- Fix the payoff function for voters i of type abc as
ui(a) = 2, ui(b) = ui({a, b, c}) = 1, ui(c) = 0.
- If we do similarly for the other voter types, then this fixes the
strategic game for voting according to the majority rule over the set of alternatives {a, b, c}.
Suppose 1 has ballot abc, 2 has bca, and has 3 cab This fixes the strategic game: a: a b c a
(2, 0, 1) (2, 0, 1) (2, 0, 1)
b
(2, 0, 1) (1, 2, 0) (1, 1, 1)
c
(2, 0, 1) (1, 1, 1) (0, 1, 2)
b: a b c a
(2, 0, 1) (1, 2, 0) (1, 1, 1)
b
(1, 2, 0) (1, 2, 0) (1, 2, 0)
c
(1, 1, 1) (1, 2, 0) (0, 1, 2)
c: a b c a
(2, 0, 1) (1, 1, 1) (0, 1, 2)
b
(1, 1, 1) (1, 2, 0) (0, 1, 2)
c
(0, 1, 2) (0, 1, 2) (0, 1, 2)
Nash Equilibrium If they all cast their vote according to their true ballot, then 1 votes a, 2 votes b and 3 votes c. Then the outcome is a tie, {a, b, c}, with payoff (1, 1, 1). This is a Nash equilibrium: the vote cast by each player is a best response in the strategy profile.
Changing the Game Now let’s change the voting rule slightly, by switching to majority voting with tie breaking, where abc as the tie breaking order. This changes the majority rule into a resolute voting rule. It also changes the strategic game . . .
New Strategic Game a: a b c a
(2, 0, 1) (2, 0, 1) (2, 0, 1)
b
(2, 0, 1) (1, 2, 0) (2, 0, 1)
c
(2, 0, 1) (2, 0, 1) (0, 1, 2)
b: a b c a
(2, 0, 1) (1, 2, 0) (2, 0, 1)
b
(1, 2, 0) (1, 2, 0) (1, 2, 0)
c
(2, 0, 1) (1, 2, 0) (0, 1, 2)
c: a b c a
(2, 0, 1) (2, 0, 1) (0, 1, 2)
b
(2, 0, 1) (1, 2, 0) (0, 1, 2)
c
(0, 1, 2) (0, 1, 2) (0, 1, 2)
Coalition Formation
- If the players all vote according to their true preference, the out-
come is a because of the tie breaking, with payoff given by (2, 0, 1).
- But this is no longer a Nash equilibrium, for player 2 can improve
his payoff from 0 to 1 by casting vote c, which causes the outcome to change into c, with payoff (0, 1, 2).
- The strategy triple (a, c, c) is a Nash equilibrium.
- The voting rule seems to favour voter 1 with ballot abc, because
the tie breaking rule uses this order for tie breaking.
- Still, the voter with this ballot ends up losing the game, because
the other two players have an incentive to form a coalition against player 1.
A Language for MASL
- Basic actions: the full strategy profiles.
- Assumption for convenience: the choice of strategies is the same
for each player (as in voting).
- State of the game: a strategy vector where each player has deter-
mined her strategy.
- Individual strategies emerge as unions of group strategies.
Example: strategy c for the first player in the PD game cc cd dc dd c,? ? c,? ? c,? ? c,? ? c,? ? c,? ? c,? ?
Strategy Terms
- Strategy terms of MASL are:
t ::= a | ? ? | ! !
- Here a ranges over the set of all player strategies A.
- “?
?” denotes an individual strategy for an adversary player.
- “!
!” denotes the current strategy of a player.
- Random terms serve to model what adversaries do, and current
terms serve to model what happens when players stick to a previ-
- us choice.
MASL Actions and Formulas c ::= (t1 . . . , tn) The MASL strategy vectors occur as atoms and as modalities in MASL formulas. Assume that p ranges over a set of game outcome values, that is: assume an outcome function o : AN → P, and let p range over P. The MASL language is built in the usual PDL manner by mutual recursion of action expressions and formulas: ϕ ::= ⊤ | c | p | ¬ϕ | ϕ1 ∧ ϕ2 | [γ]ϕ γ ::= c | ?ϕ | γ1; γ2 | γ1 ∪ γ2 | γ∗
Interpretation Let s ∈ An. Then s is a strategy profile, with individual strategies taken from A. Let i ∈ N. Then [ [·] ]A,s,i is a function that maps each c to a subset of A, and [ [·] ]A,s is a function that maps each group strategy to a set of strategy profiles ⊆ An, as follows: [ [a] ]A,s,i = {a} [ [? ?] ]A,s,i = A [ [! !] ]A,s,i = {s[i]} [ [(t1 . . . , tn)] ]A,s = [ [t1] ]A,s,1 × · · · × [ [tn] ]A,s,n
Examples
- Let A = {a, b, c}, and let n = 3.
- Then a strategic change by the first player to b, while both other
players stick to their vote is expressed as (b, ! !, ! !). In a game state (a, b, b) this is interpreted as {((a, b, b), (b, b, b))}.
- A strategic change by the first player to b, given that the second
player sticks to her vote, while the third player may or may not change, is expressed by (b, ! !, ? ?). In the context of a strategy pro- file s = (a, b, c), this is interpreted as follows: [ [(b, ! !, ? ?)] ]A,s = {b} × {b} × {a, b, c}.
- (?
?, c, c) represents the group strategy where players 2 and 3 both play c. This is a strategy for the coalition of 2 and 3 against 1.
- The formula that expresses that the coalition of 1 and 2 can force
- utcome c by both voting c is (abbreviating the singleton outcome
{c} as c): [(? ?, c, c)]c.
- The strategy (?
?, ? ?, c) is different from (! !, ! !, c), for the latter ex- presses the individual strategy for player 3 of playing c, in a con- text where the two other players do not change their strategy.
- If we represent a strategy for player i as a relation, then we have
to take into account that the individual choice of i does need in- formation about how the others move to determine the outcome. The relation for the individual choice a of player i is given by [ [(? ?, · · · , ? ?, a, ? ?, · · · , ? ?)] ]A,s = A × · · · A × {a} × A × · · · A. This is computed from all choices that the other players could make (all strategies for the other players).
- Compare this with
[ [(! !, · · · , ! !, a, ! !, · · · , ! !)] ]A,s = {s[1]} × · · · {s[i − 1]} × {a} × {s[i + 1]} × · · · {s[n]}. This is the action where player i switches to a, while all other players stick to their strategies.
Interpretation of (c, ! !) in the PD Game cc cd dc dd c,! ! c,! ! c,! ! c,! !
Truth Definition (let M = (N, AN, o), and s ∈ AN) M, s | = ⊤ always M, s | = S iff s ∈ [ [S] ]s M, s | = p iff s ∈ o−1(p) M, s | = ¬ϕ iff M, s | = ϕ M, s | = ϕ1 ∧ ϕ2 iff M, s | = ϕ1 and M, s | = ϕ2 M, s | = [γ]ϕ iff for all t with (s, t) ∈ [ [γ] ]M : M, t | = ϕ [ [S] ]M = {(s, t) | t ∈ [ [S] ]A,s} [ [?ϕ] ]M = {(s, s) | M, s | = ϕ} [ [γ1; γ2] ]M = [ [γ1] ]M ◦ [ [γ2] ]M [ [γ1 ∪ γ2] ]M = [ [γ1] ]M ∪ [ [γ2] ]M [ [γ∗] ]M = ([ [γ] ]M)∗.
Abbreviations
- Let (ia, !
!) abbreviate the strategy vector (! !, · · · , ! !, a, ! !, · · · , ! !), with a in i-th position, and ! ! everywhere else.
- Let [(i, !
!)]ϕ abbreviate
a∈A[(ia, !
!)]ϕ.
- Then [(i, !
!)]ϕ expresses that all strategies to which player i can switch from the current strategy profile result in a strategy profile where ϕ holds (provided that the other players keep their strate- gies fixed).
Abbreviations ctd
- Let (ia, ?
?) abbreviate the strategy vector (? ?, · · · , ? ?, a, ? ?, · · · , ? ?), with a in i-th position, and ? ? everywhere else.
- Let [(i, ?
?)]ϕ abbreviate
a∈A[(ia, ?
?)]ϕ.
- Then [(i, ?
?)]ϕ expresses that all strategies for i guarantee ϕ, no matter what the other players do.
- Let (?
?) abbreviate (? ?, · · · , ? ?) (the strategy vector that everywhere has ? ?).
- Then (?
?)ϕ expresses that in some game state ϕ holds.
Representing Payoffs
- To represent payoffs, assume that basic propositions are payoff
vectors u, and that the payoff values are in a finite set U (the set
- f all utilities that can be assigned in the game).
- Define ui ≥ v as
w∈U,w≥v u[i] = w
- Then ui ≥ v expresses that player i gets at least v.
- Define ui > v as
w∈U,w>v u[i] = w.
- Then ui > v expresses that player i gets more than v.
Weak Dominance Intuitively, and i-strategy a is weakly dominant if a is at least as good for i against any moves the other players can make as any alternative b for a. In our logic:
- v∈U
- b∈A−{a}
[(ib, ? ?)](ui ≥ v → (ia, ! !)ui ≥ v).
Nash Equilibrium The following formula expresses that the current strategy profile is a Nash equilibrium:
- i∈N
- v∈U
(ui ≥ v ∧ [(i, ! !)]¬ui > v). The following formula expresses that the game is Nash: (? ?)
- i∈N
- v∈U
(ui ≥ v ∧ [(i, ! !)]¬ui > v).
Majority Voting Let Ma be the set of all full strategy vectors where a majority of the players play a. Then
- x∈A
- c∈Mx
[c]x expresses that the game is a voting game with majority rule.
Resoluteness Assume that the proposition a expresses that a is among the winners given the current profile. A voting rule is resolute if there is always exactly one winner. Viewing voting according to a voting rule as a game, the following formula expresses that the game is resolute: [(? ?)]
- a∈A
(a ∧
- b∈A−{a}
¬b).
Strategy-Proofness A voting rule is strategy proof if it holds for any profile S and for any player (voter) i that changing his vote (action) does not give an
- utcome that is better (according to the preferences of i in S) that the
- utcome in S. This is expressed by the following formula:
[(? ?)]
- i∈N
- v∈U
(ui ≥ v ∧ ¬(i, ! !)ui > v).
Non-Imposedness A voting rule is (weakly) non-imposed if at least three outcomes are possible. Viewing voting as a game, we can use the following formula to ex- press this:
- a∈A
- b∈A−{a}
- c∈A−{a,b}
((? ?)a ∧ (? ?)b ∧ (? ?)c).
Dictatorship In a multi-agent game setting, a dictator is a player who can always get what he wants, where getting what you want is getting a payoff that is at least as good as anything any other player can achieve. Here is the formula for that, using the abbreviation [(i, ? ?)]ϕ: [(i, ? ?)]
- v∈U
(ui ≥ v ∧
- j∈N−{i}
¬uj > v).
Gibbard-Satterthwaite in MASL The classic Gibbard-Satterthwaite theorem [9, 15] states that all rea- sonable voting rules allow strategizing, or put otherwise, that no rea- sonable voting rule is strategy-proof. Resoluteness, strategy-proofness, non-imposedness and dictatorship are the four properties in terms of which the Gibbard-Satterthwaite theorem is formulated. Statement of the Gibbard-Satterthwaite theorem in our logic: RES ∧ SP ∧ NI ⊢ DICT. To do: prove this in MASL.
Meta-Strategies: Tit-for-Tat Tit-for-Tat as a meta-strategy for the PD game [3] is the instruction to copy one’s opponents last choice, thereby giving immediate, and rancour-free, reward and punishment. Here is a picture of the Tit-for- Tat meta-strategy for player 2, with the states indicating the outcomes
- f the last play of the game:
cc cd dc dd ? ?, c ? ?, d ? ?, c ? ?, c ? ?, d ? ?, d This works because we may think of the current state of the game as the result of the last play of PD, remembered in the state.
Tit-For-Tat as a MASL Action Expression The following MASL action expression describes this meta-strategy for player 2. (?(c, ? ?); (? ?, c) ∪ ?(d, ? ?); (? ?, d))∗ What this says is: if the last action by the opponent was a c, then reward, otherwise (the last action by the opponent was a d) punish. To turn this into a meta-strategy for player 1, just swap all pairs: (?(? ?, c); (c, ? ?) ∪ ?(? ?, d); (d, ? ?))∗ Note that tit-for-tat for the PD game boils down to the same thing as the copycat meta-strategy.
Calculus for MASL To axiomatize this logic, we can use the well-known proof system for PDL [16, 10], with appropriate axioms for the strategy vectors added. Call a strategy vector c = (c1 . . . , cn) determinate if for no i ∈ N ci = ? ?. Vector axioms are:
- 1. Effectivity:
[c]c.
- 2. Seriality:
c⊤.
- 3. Functionality:
cϕ → [c]ϕ for all determinate strategy vectors S.
- 4. Adversary power:
Let c have ? ? in position i, and let ci
a be the result of replacing ?
? in position i in c by a. Then: [c]ϕ ↔
- a∈A
[ci
a]ϕ.
Note that this uses the assumption that the set A of available ac- tions is finite.
- 5. Determinate current choice:
Let c have ! ! in position i, and let ci
a be the result of replacing !
! at position i in S by a. Then: (ia, ! !) → (c ↔ ci
a).
Soundness and Completeness These axioms are sound for the intended interpretation. Completeness can be shown by the usual canonical model construc- tion for PDL (see [4]): Theorem 1 The calculus for MASL is complete. MASL has the same complexity for model checking and satisfiability as PDL: Model checking for PDL and MASL is PTIME-complete. Sat solving for PDL and MASL is EXPTIME-complete.
Connections
- Coalition Logic [12]
- Game Logic [11]
- Strategy Logic [5]
- Alternating-time Temporal Logic [2]
- Kn with intersection [1]
- . . .
Further Work Extend MASL to epistemic MASL. Language for EMASL: ϕ ::= ⊤ | c | p | ¬ϕ | ϕ1 ∧ ϕ2 | [γ]ϕ | [α]ϕ γ ::= c | ?ϕ | γ1; γ2 | γ1 ∪ γ2 | γ∗ α ::= i | iˇ |?ϕ | α1; α2 | α1 ∪ α2 | α∗ The interpretations of the i operators (the atoms of α actions) can be arbitrary. Define Kiϕ as [(i ∪ iˇ)∗]ϕ, and you have a reflexive, symmetric and transitive knowledge operator (see [6]).
References [1] T. Agotnes and N. Alechina. Reasoning about joint action and coalitional ability in Kn with intersection. In J. Leite, P. Torroni,
- T. Agotnes, G. Boella, and L. van der Torre, editors, Computa-
tional Logic in Multi-Agent Systems – 12th International Work- shop (CLIMA XII), number 6814 in Lecture Notes in Computer Science, pages 139–156. Springer, 2011. [2] R. Alur, T. Henzinger, and O. Kupferman. Alternating-time tem- poral logic. Journal of the ACM, 49:672–713, 2002. [3] R. Axelrod. The Evolution of Cooperation. Basic Books, New York, 1984. [4] P. Blackburn, M. de Rijke, and Y. Venema. Modal Logic. Cam- bridge Tracts in Theoretical Computer Science. Cambridge Uni- versity Press, 2001.
[5] K. Chatterjee, T. A. Henzinger, and N. Piterman. Strategy logic.
- Inf. Comput., 208(6):677–693, June 2010.
[6] J. v. Eijck and Y. Wang. Propositional Dynamic Logic as a logic
- f belief revision.
In W. Hodges and R. de Queiros, editors, Proceedings of Wollic’08, number 5110 in Lecture Notes in Ar- tificial Intelligence, pages 136–148. Springer, 2008. [7] R. Farquharson. Theory of Voting. Blackwell, 1969. [8] M. Fischer and R. Ladner. Propositional dynamic logic of regular programs. Journal of Computer and System Sciences, 18(2):194–211, 1979. [9] A. Gibbard. Manipulation of voting schemes: A general result. Econometrica, 41:587–601, 1973. [10] R. Parikh. The completeness of propositional dynamic logic.
In Mathematical Foundations of Computer Science 1978, pages 403–415. Springer, 1978. [11] R. Parikh. Propositional game logic. In IEEE Symposium on Foundations of Computer Science, pages 195–200, 1983. [12] M. Pauly. Logic for Social Software. PhD thesis, ILLC, Ams- terdam, 2001. [13] V. Pratt. Semantical considerations on Floyd–Hoare logic. Pro- ceedings 17th IEEE Symposium on Foundations of Computer Science, pages 109–121, 1976. [14] V. Pratt. Application of modal logic to programming. Studia Logica, 39:257–274, 1980. [15] M. Satterthwaite. Strategy-proofness and Arrow’s conditions: Existence and correspondence theorems for voting procedures
and social welfare functions. Journal of Economic Theory, 10:187–217, 1975. [16] K. Segerberg. A completeness theorem in the modal logic of
- programs. In T. Traczyck, editor, Universal Algebra and Appli-