Multi-agent learning
The repli ato r dynamiGerard Vreeswijk, Intelligent Software Systems, Computer Science Department, Faculty of Sciences, Utrecht University, The Netherlands.
Wednesday 10th June, 2020
Multi-agent learning The repliato r dynami Gerard Vreeswijk , - - PowerPoint PPT Presentation
Multi-agent learning The repliato r dynami Gerard Vreeswijk , Intelligent Software Systems, Computer Science Department, Faculty of Sciences, Utrecht University, The Netherlands. Wednesday 10 th June, 2020 disrete repliato r disrete
Gerard Vreeswijk, Intelligent Software Systems, Computer Science Department, Faculty of Sciences, Utrecht University, The Netherlands.
Wednesday 10th June, 2020
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 2
p ropAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 2
■ Symmetric games, symmetric
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 2
■ Symmetric games, symmetric
■ Evolutionary game theory:
p ropAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 2
■ Symmetric games, symmetric
■ Evolutionary game theory:
p rop■ The replicator dynamic.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 2
■ Symmetric games, symmetric
■ Evolutionary game theory:
p rop■ The replicator dynamic.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 2
■ Symmetric games, symmetric
■ Evolutionary game theory:
p rop■ The replicator dynamic.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 2
■ Symmetric games, symmetric
■ Evolutionary game theory:
p rop■ The replicator dynamic.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 2
■ Symmetric games, symmetric
■ Evolutionary game theory:
p rop■ The replicator dynamic.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 2
■ Symmetric games, symmetric
■ Evolutionary game theory:
p rop■ The replicator dynamic.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 2
■ Symmetric games, symmetric
■ Evolutionary game theory:
p rop■ The replicator dynamic.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 2
■ Symmetric games, symmetric
■ Evolutionary game theory:
p rop■ The replicator dynamic.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 2
■ Symmetric games, symmetric
■ Evolutionary game theory:
p rop■ The replicator dynamic.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 3
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 4
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 4
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 4
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 4
■ As a bi-matrix.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 4
■ As a bi-matrix. ■ As a partially filled bi-matrix.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 4
■ As a bi-matrix. ■ As a partially filled bi-matrix. ■ As a plain matrix.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 5
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 6
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 6
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 6
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 6
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 6
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 7
symmetri equilib riumAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 7
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 7
■ !! Symmetric equilibria can be identified with strategies !!
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 7
■ !! Symmetric equilibria can be identified with strategies !! ■ (Theorem.) Every symmetric game has at least one symmetric
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 7
■ !! Symmetric equilibria can be identified with strategies !! ■ (Theorem.) Every symmetric game has at least one symmetric
■ (Fact.) Symmetric games can have a-symmetric equilibria.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 7
■ !! Symmetric equilibria can be identified with strategies !! ■ (Theorem.) Every symmetric game has at least one symmetric
■ (Fact.) Symmetric games can have a-symmetric equilibria.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 7
■ !! Symmetric equilibria can be identified with strategies !! ■ (Theorem.) Every symmetric game has at least one symmetric
■ (Fact.) Symmetric games can have a-symmetric equilibria.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 8
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 9
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 10
p ropAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 10
■ There are n, say 5, species.
p ropAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 10
■ There are n, say 5, species. An encounter
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 10
■ There are n, say 5, species. An encounter
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 10
■ There are n, say 5, species. An encounter
■ The population consists of a very large
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 10
■ There are n, say 5, species. An encounter
■ The population consists of a very large
■ We are
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 10
■ There are n, say 5, species. An encounter
■ The population consists of a very large
■ We are
■ The
tness ofj=1 pjAij
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 10
■ There are n, say 5, species. An encounter
■ The population consists of a very large
■ We are
■ The
tness ofj=1 pjAij
■ The
average tness isi=1 pi fi
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 11
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 12
repli ato r dynami sAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 12
■ Defined for a single species by Taylor and Jonker (1978), and named
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 12
■ Defined for a single species by Taylor and Jonker (1978), and named
■ The replicator equation is the first game dynamics studied in
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 12
■ Defined for a single species by Taylor and Jonker (1978), and named
■ The replicator equation is the first game dynamics studied in
Taylor P.D., Jonker L. “Evolutionarily stable strategies and game dynamics” in: Math. Biosci. 1978;40(1), pp. 145-156.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 12
■ Defined for a single species by Taylor and Jonker (1978), and named
■ The replicator equation is the first game dynamics studied in
Taylor P.D., Jonker L. “Evolutionarily stable strategies and game dynamics” in: Math. Biosci. 1978;40(1), pp. 145-156. Schuster P., Sigmund K. “Replicator dynamics” in: J. Theor. Biol. 1983, 100(3), pp. 533-538.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 13
repli ato r equation relative s o re matrix p ropAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 13
■ The
repli ato r equation models how n different specifies grow (orAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 13
■ The
repli ato r equation models how n different specifies grow (or■ It is assumed that if an individual of species i interacts with an
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 13
■ The
repli ato r equation models how n different specifies grow (or■ It is assumed that if an individual of species i interacts with an
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 13
■ The
repli ato r equation models how n different specifies grow (or■ It is assumed that if an individual of species i interacts with an
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 13
■ The
repli ato r equation models how n different specifies grow (or■ It is assumed that if an individual of species i interacts with an
■ The number of individuals of species i is denoted by qi, or qi(t).
p ropAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 13
■ The
repli ato r equation models how n different specifies grow (or■ It is assumed that if an individual of species i interacts with an
■ The number of individuals of species i is denoted by qi, or qi(t). ■
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 13
■ The
repli ato r equation models how n different specifies grow (or■ It is assumed that if an individual of species i interacts with an
■ The number of individuals of species i is denoted by qi, or qi(t). ■
■ So pi ∝ qi and p1 + · · · + pn = 1.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 14
tness tness ve to rAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 14
■ The
tness of an individual is itsAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 14
■ The
tness of an individual is its■ Example. Suppose
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 14
■ The
tness of an individual is its■ Example. Suppose
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 14
■ The
tness of an individual is its■ Example. Suppose
■ Average fitness:
3
j=1
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 14
■ The
tness of an individual is its■ Example. Suppose
■ Average fitness:
3
j=1
■ Fitness of species 1:
3
j=1
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 14
■ The
tness of an individual is its■ Example. Suppose
■ Average fitness:
3
j=1
■ Fitness of species 1:
3
j=1
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 14
■ The
tness of an individual is its■ Example. Suppose
■ Average fitness:
3
j=1
■ Fitness of species 1:
3
j=1
■ Species 2 and 3 have
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 15
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 16
i(t) = dpi(t)/dt.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 16
i(t) = dpi(t)/dt.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 16
i(t) = dpi(t)/dt.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 16
i(t) = dpi(t)/dt.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 16
i(t) = dpi(t)/dt.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 16
i(t) = dpi(t)/dt.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 16
i(t) = dpi(t)/dt.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 16
i(t) = dpi(t)/dt.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 16
i(t) = dpi(t)/dt.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 16
i(t) = dpi(t)/dt.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 16
i(t) = dpi(t)/dt.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 17
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 18
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 19
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 20
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 20
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 20
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 20
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 20
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 20
■
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 20
■
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 20
■
■ A rest point p is called
(Ly apunov) stable if for every neighborhood UAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 20
■
■ A rest point p is called
(Ly apunov) stable if for every neighborhood UAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 20
■
■ A rest point p is called
(Ly apunov) stable if for every neighborhood U■ A rest point p is called
asymptoti ally stable if p has a neighborhood UAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 20
■
■ A rest point p is called
(Ly apunov) stable if for every neighborhood U■ A rest point p is called
asymptoti ally stable if p has a neighborhood UAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 21
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 21
■ Nash equilibrium ⇔
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 21
■ Nash equilibrium ⇔
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 21
■ Nash equilibrium ⇔
■ Nash equilibrium ⇒ rest
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 21
■ Nash equilibrium ⇔
■ Nash equilibrium ⇒ rest
■ Fully mixed rest point ⇒
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 21
■ Nash equilibrium ⇔
■ Nash equilibrium ⇒ rest
■ Fully mixed rest point ⇒
■ Strict Nash equilibrium ⇒
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 21
■ Nash equilibrium ⇔
■ Nash equilibrium ⇒ rest
■ Fully mixed rest point ⇒
■ Strict Nash equilibrium ⇒
■ Limit point in the interior of
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 21
■ Nash equilibrium ⇔
■ Nash equilibrium ⇒ rest
■ Fully mixed rest point ⇒
■ Strict Nash equilibrium ⇒
■ Limit point in the interior of
■ Asymptotically stable in the
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 22
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 23
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 24
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 25
dis rete step equation birth and death rate absolute gro wthAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 25
■ The
dis rete step equation is given byAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 25
■ The
dis rete step equation is given by■ To prevent negative proportions and sudden extermination, it is
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 25
■ The
dis rete step equation is given by■ To prevent negative proportions and sudden extermination, it is
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 25
■ The
dis rete step equation is given by■ To prevent negative proportions and sudden extermination, it is
■ The
absolute gro wth of species i isAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 26
dis rete repli ato r equation intrinsi birth and death ratio tness dep endent birth ratio dis rete step equationAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 26
■ The
dis rete repli ato r equation:Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 26
■ The
dis rete repli ato r equation:■ The idea of the discrete replicator equation: between two generations,
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 26
■ The
dis rete repli ato r equation:■ The idea of the discrete replicator equation: between two generations,
■ The DRE follows from the
dis rete step equation:Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 27
j=1 qj(t + 1) =
j=1 qj(t)[1 + β + fj(t)]
1 q(t)qi(t)[1 + β + fi(t)] 1 q(t) ∑n j=1 qj(t)[1 + β + fj(t)]
j=1 pj(t)[1 + β + fj(t)]
j=1 pj(t) + β ∑n j=1 pj(t) + ∑n j=1 pj(t) fj(t)]
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 28
dis rete repli ato r equationAuthor: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 28
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 28
■ If pi was 0 it remains 0.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 28
■ If pi was 0 it remains 0. ■ If pi was positive it remains positive.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 29
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 30
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 31
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 32
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 32
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 32
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 32
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 32
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 33
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 34
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 34
■ Idea: reduce time steps t = 1 to smaller time steps t = δ, where
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 34
■ Idea: reduce time steps t = 1 to smaller time steps t = δ, where
■ The idea is the following:
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 34
■ Idea: reduce time steps t = 1 to smaller time steps t = δ, where
■ The idea is the following:
■ What if δ = 0?
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 34
■ Idea: reduce time steps t = 1 to smaller time steps t = δ, where
■ The idea is the following:
■ What if δ = 0? What if δ = 1?
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 35
δ→0
δ→0 qi(β + fi(t))
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 36
j=1 qj(t + δ)
j=1
j=1
j=1 pj(t)[1 + δ(β + fj(t))]
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 36
j=1 qj(t + δ)
j=1
j=1
j=1 pj(t)[1 + δ(β + fj(t))]
■ What if δ = 0?
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 36
j=1 qj(t + δ)
j=1
j=1
j=1 pj(t)[1 + δ(β + fj(t))]
■ What if δ = 0? What if δ = 1?
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 37
1+δ(β+ ¯ f(t)) − pi(t)
1+δ((β+ fi(t)) 1+δ(β+ ¯ f(t)) − 1
δ→0
δ→0 pi(t)
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 38
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 39
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 39
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 39
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 39
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 39
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 39
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 39
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 39
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 39
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 39
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 39
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 39
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 40
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 41
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 42
SN ESS NSS GSS ASS LSS LR NE FP * i * * SN = strict Nash, ESS - evol’y stable strategy, GSS = glob’y stable state, ASS = asymp’y stable state, NSS = neutrally stable strategy, LR = limit of replicator, LSS = Lyapunov stable state, FP = fixed point, * = only if fully mixed, i = isolated NE. Dotted: indirect implication. Blue: game theory; olive: evolutionary game theory; green: the replicator dynamic.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 43
■ SN ⇒ ESS: cf. slides evolutionary games and, e.g., Th 7.7.12 of Sh&LB. ■ ESS ⇒ NSS: cf. slides evolutionary games and, e.g., Game Theory Evolving (2nd ed.) by H. Gintis. ■ ESS ⇒ NE: cf. slides evolutionary games and, e.g., Sh&LB Th 7.7.11. ■ ESS ⇒∗ GSS: cf., e.g., Th. 12.7 Gintis. ■ ESS ⇒ ASS: cf., e.g., Th. 7.7.13 Sh&LB,
■ NSS ⇒ LSS: cf. Sec. 3.5 Weibull. ■ GSS ⇒ ASS: by definition of the two concepts. ■ ASS ⇒ LSS: by definition of the two concepts. ■ ASS ⇒ LR: by definition of the two concepts. ■ ASS ⇒i NE: Th 7.7.8 Sh&LB,
■ LSS ⇒ NE: Th 7.7.6 Sh&LB, 7.2.1(c) Hofbauer & Sigmund. ■ LR ⇒∗ NE: Th. 7.2.1(b) H&S. ■ NE ⇒ FP: Th. 7.2.1(a) H&S, Th 7.7.5 Sh&LB, Th. 12.6 Gintis. ■ LR ⇒ FP: Ch. 6 Weibull.
Author: Gerard Vreeswijk. Slides last modified on June 10th, 2020 at 14:01 Multi-agent learning: The replicator dynamic, slide 44
Weibull, J. W. (1997). Evolutionary game theory. MIT press. Hofbauer, J., & Sigmund, K. (1998). Evolutionary games and population dynamics. Cambridge university press. Gintis, H. (2001). Game theory evolving: A problem-centered introduction to modeling strategic interaction (2nd ed.). Princeton: Princeton University Press. Shoham, Y., & Leyton-Brown, K. (2008). Multiagent systems: Algorithmic, game-theoretic, and logical foundations. Cambridge University Press. Vreeswijk, G.A.W. (2011). Evolutionary game theory. (Slides)