Models of Language Evolution Replicator dynamic & signaling - - PowerPoint PPT Presentation

models of language evolution
SMART_READER_LITE
LIVE PREVIEW

Models of Language Evolution Replicator dynamic & signaling - - PowerPoint PPT Presentation

Models of Language Evolution Replicator dynamic & signaling Michael Franke Class survey results Topics for today 1 replicator dynamic 2 evolutionary dynamics of signaling games Replicator dynamic (discrete) Replicator dynamic (continuous)


slide-1
SLIDE 1

Models of Language Evolution

Replicator dynamic & signaling Michael Franke

slide-2
SLIDE 2

Class survey results

slide-3
SLIDE 3
slide-4
SLIDE 4
slide-5
SLIDE 5
slide-6
SLIDE 6

Topics for today

1 replicator dynamic 2 evolutionary dynamics of signaling games

slide-7
SLIDE 7

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

7 / 36

slide-8
SLIDE 8

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Levels of analysis in EGT

are attractors of the dynamics static solutions? is macro-dynamics the mean field of the micro-dynamics?

Static Solutions

Nash equilibrium evolutionary stability . . .

Marco-Dynamics (Population-Level)

replicator dynamic best response dynamic . . .

Micro-Dynamics (Agent-Based)

imitate the best conditional imitation reinforcement learning . . . 8 / 36

slide-9
SLIDE 9

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Replicator dynamic

  • arguably, the most general formalization of fitness-proportional growth
  • mathematically convenient:
  • discrete version numeric simulation
  • continuous version mathematical proof
  • uniform formalism, multiply interpretable
  • clear connection with stability & equilibrium notions

9 / 36

slide-10
SLIDE 10

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Utility in Mean-Field Populations

Recap

  • let ni be the number of agents playing action ai
  • let n be the size of the population
  • population aggregate is a probability vector

p where: pi = ni n

  • if population is huge, the average payoff of ai is:

U(ai, p) = ∑

j

pj × U(ai, aj)

  • U(ai,

p) is the fitness of ai (given the population state)

  • U(

p) = ∑i pi × U(ai, p) is the average fitness in the population

10 / 36

slide-11
SLIDE 11

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Discrete-time replicator dynamics: Derivation

Assumptions Average offspring of an individual playing ai is a positive scaling function F of i’s fitness U(ai, p): F(x) = kx with k > 0.

  • n′

i is the number of individuals playing ai at the next discrete time step

  • n′

i = ni F(U(ai,

p)) p′

i =

n′

i

∑j n′

j

= ni F(U(ai, p)) ∑j nj F(U(aj, p)) = ni k U(ai, p) ∑j nj k U(aj, p) = ni U(ai, p) ∑j nj U(aj, p) = n pi U(ai, p) ∑j n pj U(aj, p) = pi U(ai, p) ∑j pj U(aj, p) = pi U(ai, p) U( p)

11 / 36

slide-12
SLIDE 12

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Discrete time replicator dynamic

p′

i = pi

U(ai, p) ∑j pj U(aj, p) = pi U(ai, p) U( p) = proportion of i × fitness of i average fitness If pi = 0, frequency pi of players of type ai . . . . . . increases when i’s fitness is higher than average; . . . decreases when i’s fitness is lower than average; . . . stays constant when i’s fitness is exactly average. If pi = 0, then p′

i = 0.

12 / 36

slide-13
SLIDE 13

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Time series

Coordination: U =

  • 1

1

  • 13 / 36
slide-14
SLIDE 14

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Time series

Prisoner’s Dilemma: U =

  • 2

3 1

  • 14 / 36
slide-15
SLIDE 15

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Time series

Hawks & Doves: U =

  • 1

7 2 3

  • 15 / 36
slide-16
SLIDE 16

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Time series

Coordination: U =

  • 1

2

  • 16 / 36
slide-17
SLIDE 17

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Source Code for Plots

17 / 36

slide-18
SLIDE 18

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Analyzing the replicator dynamics

U =

  • 1

2

  • p′

1 =

p2

1

2(1 − p1)2 + p2

1

0.2 0.4 0.6 0.8 1 0.2 0.4 0.6 0.8 1 p1 p′

1

0.5 1 0.5 1 p1 p2 1 p1

18 / 36

slide-19
SLIDE 19

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

19 / 36

slide-20
SLIDE 20

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Continuous time replicator dynamics

Derivation

˙ pi = dpi(t) dt = lim

δt→0

pi(t + δt) − pi(t) δt

  • [def. of derivative]

= p′

i − pi

[discrete time RD gives limit step] = pi U(ai, p) U( p) − pi [def. of discrete time RD] = pi U(ai, p) − U( p) U( p) [“payoff-adjusted RD”] ”=” pi [U(ai, p) − U( p)] [drop constant denominator]

20 / 36

slide-21
SLIDE 21

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Continuous time replicator dynamic

˙ pi = pi [U(ai, p) − U( p)] = proportion of i × [fitness of i − average fitness] If pi = 0, frequency pi of players of type ai . . . . . . increases when i’s fitness is higher than average; . . . decreases when i’s fitness is lower than average; . . . stays constant when i’s fitness is exactly average. If pi = 0, then ˙ pi = 0.

21 / 36

slide-22
SLIDE 22

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Conditional imitation

Assumptions

  • “mean field population”: huge and homogeneous
  • every agent plays fixed strategy for long periods of time
  • occasionally i considers to adopt j’s strategy
  • switching probability is proportional to how much better j’s strategy is than i’s

Revision protocol A revision protocol gives the average propensity (non-normalized probability) of agent i switching to agent j’s strategy: ρ

p ij = pj

  • U(aj,

p) − U(ai, p)

  • +

= proportion of j × fitness difference between j and i (if j is fitter)

22 / 36

(e.g. Helbing, 1996; Schlag, 1998)

slide-23
SLIDE 23

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Derivation of the replicator dynamic

˙ pi = flow into i − flow out of i = ∑

j

pjρ

p ji − ∑ j

piρ

p ij

= ∑

j

pjpi

  • U(ai,

p) − U(aj, p)

  • + − ∑

j

pipj

  • U(aj,

p) − U(ai, p)

  • +

= pi ∑

j

pj

  • U(ai,

p) − U(aj, p)

  • + −
  • U(aj,

p) − U(ai, p)

  • +
  • = pi ∑

j

pj

  • U(ai,

p) − U(aj, p)

  • = pi

j

pj U(ai, p) − ∑

j

pj U(aj, p)

  • = pi [U(ai,

p) − U( p)]

23 / 36

slide-24
SLIDE 24

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Rest points, dynamic stability & attraction

  • A rest point is a state

p with ˙ pi = 0 for all i.

  • a rest point

p is (weakly / Lyapunov) stable iff:

  • all nearby points stay

nearby

  • for all open neighborhoods U of

p there is a neighborhood O ⊆ U of p such that any point in O never migrates out of U

  • a rest point

p is attractive iff:

  • all nearby points

converge to it

  • there is an open neighborhood U of

p such that all points in U converge to p

  • basin of attraction of an attractive rest point:
  • biggest U with the above property
  • a rest point

p is asymptotically stable (aka. an attractor) iff:

  • all nearby points

converge to it (on a path that stays close)

  • it is stable and attractive

24 / 36

slide-25
SLIDE 25

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Replicator dynamic, equilibrium & evolutionary stability

Equilibrium

1 nes ⊆ rest points 2 snes ⊆ attractors 3 if an interior orbit converges to

p, then p is a ne

4 if a rest point is stable, then it is a ne

Evolutionary stability

1 esss ⊆ attractors 2 nsss ⊆ Lyapunov stable 3 all interior esss are global attractors,

i.e., attract all interior points Special case: “potential games” (U = UT)

1 esss = attractors 2 every interior orbit converges (to a

ne)

25 / 36

(e.g. Hofbauer and Sigmund, 1998)

slide-26
SLIDE 26

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling 26 / 36

slide-27
SLIDE 27

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

27 / 36

slide-28
SLIDE 28

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Early simulation evidence

In 2-2-2 Lewis games (with equiprobable states), all simulation runs of the (discrete, symmetric) RD converged to signaling systems.

28 / 36

(Skyrms, 1996)

slide-29
SLIDE 29

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Postitive result

2-2-2 Lewis game, equiprobable

In a 2-2-2 Lewis game with equiprobable states, the set of initial population states that do not converge to a signaling system under the replicator dynamics has Lebesgue measure zero. Lebesgue measure zero: has an extension that does not stretch across all dimensions.

29 / 36

(Huttegger, 2007)

slide-30
SLIDE 30

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Negative result

2-2-2 Lewis game, non-equiprobable

In a 2-2-2 Lewis game with non-equiprobable states, the set of initial population states that do not converge to a signaling system under the replicator dynamics has positive Lebesgue measure. Reason: there are now mixed nsss; these must be attractors, because they are Lyapunov stable (generally) and interior points must converge to a nes (partnership games).

30 / 36

(Huttegger, 2007)

slide-31
SLIDE 31

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Negative result

n-n-n Lewis games, equiprobable

In a n-n-n Lewis game with equiprobable states, the set of initial population states that do not converge to a signaling system under the replicator dynamics has positive Lebesgue measure. Reason: there are now mixed neutrally stable strategies (nsss, so-called partial pooling equilibria). Basin of attraction: ca. 5% of simulation runs in the symmetric RD converge to partial pooling equilibria.

31 / 36

(Huttegger, 2007; Pawlowitsch, 2008)

slide-32
SLIDE 32

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Party pooper: partial pooling

ma mb mc t1 t2 t3 a1 a2 a3 p 1-p q 1-q

32 / 36

(Pawlowitsch, 2008)

slide-33
SLIDE 33

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Positive result

3-3-3 Lewis game, equiprobable

In a 3-3-3 Lewis game with equiprobable states, the set of initial population states that do not converge to a signaling system under the replicator-mutator dynamics (with uniform small mutation rates) seems to have Lebesgue measure zero. Replicator-mutator dynamics: replicator dynamics with mutation.

33 / 36

(Huttegger et al., 2010)

slide-34
SLIDE 34

Replicator dynamic (discrete) Replicator dynamic (continuous) Dynamics of signaling

Upshot

While the evolution of perfect information transfer is not an evolutionary certainty (even in idealized models), at least partial information transfer seems almost guaranteed by success-conditioned selection of communicative strategies.

34 / 36

slide-35
SLIDE 35

Homework

read this paper:

  • Simon Kirby et al. (2014). “Iterated Learning and the Evolution of Language”. In:

Current Opinion in Neurobiology 28, pp. 108–114

slide-36
SLIDE 36

References

Helbing, Dirk (1996). “A Stochastic Behavioral Model and a ‘Microscopic’ Foundation of Evolutionary Game Theory”. In: Theory and Decision 40.2, pp. 149–179. Hofbauer, Josef and Karl Sigmund (1998). Evolutionary Games and Population Dynamics. Cambridge, Massachusetts: Cambridge University Press. Huttegger, Simon M. (2007). “Evolution and the Explanation of Meaning”. In: Philosophy

  • f Science 74, pp. 1–27.

Huttegger, Simon M. et al. (2010). “Evolutionary Dynamics of Lewis Signaling Games: Signaling Systems vs. Partial Pooling”. In: Synthese 172.1, pp. 177–191. Kirby, Simon et al. (2014). “Iterated Learning and the Evolution of Language”. In: Current Opinion in Neurobiology 28, pp. 108–114. Pawlowitsch, Christina (2008). “Why Evolution does not Always Lead to an Optimal Signaling System”. In: Games and Economic Behavior 63.1, pp. 203–226. Schlag, Karl H. (1998). “Why Imitate, and If So, How?” In: Journal of Economic Theory 78.1,

  • pp. 130–156.

Skyrms, Brian (1996). Evolution of the Social Contract. Cambridge University Press.