An introduction to mean field games and their applications Pr. - - PowerPoint PPT Presentation

an introduction to mean field games and their applications
SMART_READER_LITE
LIVE PREVIEW

An introduction to mean field games and their applications Pr. - - PowerPoint PPT Presentation

An introduction to mean field games and their applications Pr. Olivier Gu eant (Universit e Paris 1 Panth eon-Sorbonne) Mathematical Coffees Huawei/FSMP joint seminar September 2018 Table of contents 1. Introduction 2. Static


slide-1
SLIDE 1

An introduction to mean field games and their applications

  • Pr. Olivier Gu´

eant (Universit´ e Paris 1 Panth´ eon-Sorbonne) Mathematical Coffees – Huawei/FSMP joint seminar September 2018

slide-2
SLIDE 2

Table of contents

  • 1. Introduction
  • 2. Static mean field games
  • 3. MFG in continuous time (with continuous state space)
  • 4. Numerics and examples
  • 5. Special for Huawei: MFG on graphs
  • 6. Conclusion

1

slide-3
SLIDE 3

Introduction

slide-4
SLIDE 4

The speaker

  • PhD thesis on mean field games under the supervision of

Pierre-Louis Lions.

  • Academic positions at Paris 7, then ENSAE, and now Paris 1.
  • Main research field: optimal control and applications (incl.

mean field games, stochastic optimal control in finance, reinforcement learning, etc.).

  • Start-up (MFG Labs) with Lasry and Lions (created in 2009 –

acquired in 2013 by Havas).

2

slide-5
SLIDE 5

Mean field games – In the beginning were...

Pierre-Louis Lions and Jean-Michel Lasry, who introduced mean field games (MFG) in 2006. → Similar ideas arose in electrical engineering (Caines, Huang, Malham´ e, 2006)

3

slide-6
SLIDE 6

Introduction - Mean field games

Game theory

  • The study of strategic interactions.
  • Central concept of Nash equilibrium.
  • In MFG: the number of players is large.

4

slide-7
SLIDE 7

Introduction - Mean field games

Game theory

  • The study of strategic interactions.
  • Central concept of Nash equilibrium.
  • In MFG: the number of players is large.

Mean field

  • Approximation as in physics, here to model strategic

interactions, not interactions between particles.

  • Philosophical difference: freedom... however, as Spinoza said:

This is that human freedom, which all boast that they possess, and which consists solely in the fact, that men are conscious of their own desire, but are ignorant of the causes whereby that desire has been determined.

  • Difference for the maths: humans anticipate, particles do not!

Main consequence: the equations are not simply forward in time.

4

slide-8
SLIDE 8

Numerous applications

5

slide-9
SLIDE 9

Numerous applications

Economics

  • Economic growth and

inequality.

  • Oil extraction.
  • Mining industries.
  • Labor market.
  • etc.

5

slide-10
SLIDE 10

Numerous applications

Economics

  • Economic growth and

inequality.

  • Oil extraction.
  • Mining industries.
  • Labor market.
  • etc.

Population dynamics

  • Waves in stadiums (ola).
  • Structure of cities.
  • Traffic jam and other forms
  • f congestion.
  • etc.

5

slide-11
SLIDE 11

Numerous applications

Economics

  • Economic growth and

inequality.

  • Oil extraction.
  • Mining industries.
  • Labor market.
  • etc.

Population dynamics

  • Waves in stadiums (ola).
  • Structure of cities.
  • Traffic jam and other forms
  • f congestion.
  • etc.

Finance

  • Competition between asset

managers.

  • Optimal execution of several

brokers.

  • etc.

5

slide-12
SLIDE 12

Many forms but two main characteristics

6

slide-13
SLIDE 13

Many forms but two main characteristics

Different forms of mean field games

  • Static games / games in discrete time / differential games

(continuous time).

  • Discrete / continuous state space.

6

slide-14
SLIDE 14

Many forms but two main characteristics

Different forms of mean field games

  • Static games / games in discrete time / differential games

(continuous time).

  • Discrete / continuous state space.

Two main characteristics

  • Continuum of anonymous players.
  • All players maximize the same objective function (possible to

generalize to several populations of players).

6

slide-15
SLIDE 15

Many forms but two main characteristics

Different forms of mean field games

  • Static games / games in discrete time / differential games

(continuous time).

  • Discrete / continuous state space.

Two main characteristics

  • Continuum of anonymous players.
  • All players maximize the same objective function (possible to

generalize to several populations of players). A fixed-point equilibrium approach

  • Each infinitesimal player takes the distribution of players as

given.

  • The distribution of players proceeds from all individual choices.

6

slide-16
SLIDE 16

A large and worldwide community (not exhaustive and slightly

  • utdated)
  • Coll`

ege de France: P.-L. Lions + J.-M. Lasry

  • Dauphine: P. Cardaliaguet, J. Salomon, G. Turinici
  • Paris-Diderot: Y. Achdou + Ph.D. students
  • Nice: F. Delarue
  • Italy (Roma + Padua): I. Capuzzo-Dolcetta, F. Camilli, M.

Bardi

  • Princeton: R. Carmona (+ Ph.D. students), B. Moll (econ)
  • Columbia: Daniel Lacker
  • Chicago: R. Lucas (econ)
  • McGill: P. Caines + collaborators around the world
  • KAUST: D. Gomes (+ Ph.D. students), P. Markowich
  • Hong Kong + Dallas: A. Bensoussan

7

slide-17
SLIDE 17

Some references

8

slide-18
SLIDE 18

Some references

Initial papers

  • J.-M. Lasry and P.-L. Lions. Jeux `

a champ moyen i. le cas

  • stationnaire. C. R. Acad. Sci. Paris, 343(9), 2006.
  • J.-M. Lasry and P.-L. Lions. Jeux `

a champ moyen ii. horizon fini et contrˆ

  • le optimal. C. R. Acad. Sci. Paris, 343(10), 2006.
  • J.-M. Lasry and P.-L. Lions. Mean field games. Japanese

Journal of Mathematics, 2(1), Mar. 2007.

8

slide-19
SLIDE 19

Some references

Initial papers

  • J.-M. Lasry and P.-L. Lions. Jeux `

a champ moyen i. le cas

  • stationnaire. C. R. Acad. Sci. Paris, 343(9), 2006.
  • J.-M. Lasry and P.-L. Lions. Jeux `

a champ moyen ii. horizon fini et contrˆ

  • le optimal. C. R. Acad. Sci. Paris, 343(10), 2006.
  • J.-M. Lasry and P.-L. Lions. Mean field games. Japanese

Journal of Mathematics, 2(1), Mar. 2007. Courses and notes

  • 5 years of PLL’s lectures about MFG available on the website
  • f the Coll`

ege de France (in French).

  • Notes by P. Cardaliaguet

8

slide-20
SLIDE 20

Some references (cont’d)

Some applications: O. Gu´ eant, J.-M. Lasry and P.-L. Lions. Mean field games and applications, in Paris-Princeton Lectures on Mathematical Finance, 2010

9

slide-21
SLIDE 21

Static mean field games

slide-22
SLIDE 22

Reminder about game theory

  • Game theory studies strategic interactions.
  • N players. Strategies (x1, x2, . . . , xN) ∈ E N (E compact set).
  • Player i has utility (or score) ui(xi, x−i).

10

slide-23
SLIDE 23

Reminder about game theory

  • Game theory studies strategic interactions.
  • N players. Strategies (x1, x2, . . . , xN) ∈ E N (E compact set).
  • Player i has utility (or score) ui(xi, x−i).
  • Key notion: Nash equilibrium

Nash equilibrium (x∗

1, . . . , x∗ N) is a Nash equilibrium ⇐

⇒ for any player i, x∗

i is the

best strategy when others play x∗

−i.

i.e.: ∀i, x∗

i

maximizes xi → ui(xi, x∗

−i). 10

slide-24
SLIDE 24

N → +∞

Mean field hypotheses

  • Players have the same objective function ui = u.
  • Players are anonymous: ∀xi,

x−i → u(xi, x−i) is a symmetrical function. u(xi, x−i) = u  xi, 1 N − 1

  • j=i

δxj   = u

  • xi, mN−1(x−i)
  • .

11

slide-25
SLIDE 25

N → +∞

Mean field hypotheses

  • Players have the same objective function ui = u.
  • Players are anonymous: ∀xi,

x−i → u(xi, x−i) is a symmetrical function. u(xi, x−i) = u  xi, 1 N − 1

  • j=i

δxj   = u

  • xi, mN−1(x−i)
  • .

Static MFGs A static MFG is given by a function U : (x, m) ∈ E × P(E) → U(x, m), where m stands for the distribution of the players’ strategies. Remark: P(E) is the (compact) set of probability measures on E.

11

slide-26
SLIDE 26

Nash-MFG equilibrium

What is a Nash equilibrium when N tends to +∞?

  • A Nash equilibrium with N players is a tuple (x∗

1, x∗ 2, . . . , x∗ N).

When N → +∞, an equilibrium is a probability measure m.

12

slide-27
SLIDE 27

Nash-MFG equilibrium

What is a Nash equilibrium when N tends to +∞?

  • A Nash equilibrium with N players is a tuple (x∗

1, x∗ 2, . . . , x∗ N).

When N → +∞, an equilibrium is a probability measure m. Definition: Nash-MFG m is a Nash-MFG equilibrium ⇐ ⇒ The support of m is included in the argmax of x → U(x, m) ⇐ ⇒ For any probability measure f ∈ P(E) on the set of strategies E,

  • E

U(x, m)m ≥

  • E

U(x, m)f

12

slide-28
SLIDE 28

Nash-MFG equilibrium

What is a Nash equilibrium when N tends to +∞?

  • A Nash equilibrium with N players is a tuple (x∗

1, x∗ 2, . . . , x∗ N).

When N → +∞, an equilibrium is a probability measure m. Definition: Nash-MFG m is a Nash-MFG equilibrium ⇐ ⇒ The support of m is included in the argmax of x → U(x, m) ⇐ ⇒ For any probability measure f ∈ P(E) on the set of strategies E,

  • E

U(x, m)m ≥

  • E

U(x, m)f This definition shows that m solves a (rather uncommon) fixed-point problem.

12

slide-29
SLIDE 29

Underlying mathematical result

Theorem

  • Let us assume that U is continuous.
  • Let us consider a sequence ((xN

1 , . . . , xN N ))N where

∀N, (xN

1 , . . . , xN N ) is a Nash equilibrium of the N-player game

corresponding to U|E×E N/SN. Then, up to a subsequence, ∃m ∈ P(E) such that:

  • 1. mN(xN

1 , . . . , xN N ) weakly converges towards m.

  • 2. m is a Nash-MFG equilibrium.

13

slide-30
SLIDE 30

Underlying mathematical result

Theorem

  • Let us assume that U is continuous.
  • Let us consider a sequence ((xN

1 , . . . , xN N ))N where

∀N, (xN

1 , . . . , xN N ) is a Nash equilibrium of the N-player game

corresponding to U|E×E N/SN. Then, up to a subsequence, ∃m ∈ P(E) such that:

  • 1. mN(xN

1 , . . . , xN N ) weakly converges towards m.

  • 2. m is a Nash-MFG equilibrium.

Remark: this results can also be adapted to prove an existence result (by using mixed strategies).

13

slide-31
SLIDE 31

What about uniqueness?

Uniqueness If U is decreasing in the sense that ∀m1 = m2,

  • (U(x, m1) − U(x, m2))(m1 − m2) < 0

then, an equilibrium is unique.

14

slide-32
SLIDE 32

What about uniqueness?

Uniqueness If U is decreasing in the sense that ∀m1 = m2,

  • (U(x, m1) − U(x, m2))(m1 − m2) < 0

then, an equilibrium is unique. This type of monotonicity result is ubiquitous in the MFG literature.

14

slide-33
SLIDE 33

MFG and planning

Variational characterization (planner’s problem) If there exists a function m → F(m) on P(E) such that DF = U, then any maximum of F is a Nash-MFG equilibrium.

15

slide-34
SLIDE 34

MFG and planning

Variational characterization (planner’s problem) If there exists a function m → F(m) on P(E) such that DF = U, then any maximum of F is a Nash-MFG equilibrium. Remark 1: Sometimes, there exists a global problem whose solution corresponds to a MFG equilibrium.

15

slide-35
SLIDE 35

MFG and planning

Variational characterization (planner’s problem) If there exists a function m → F(m) on P(E) such that DF = U, then any maximum of F is a Nash-MFG equilibrium. Remark 1: Sometimes, there exists a global problem whose solution corresponds to a MFG equilibrium. Remark 2: Uniqueness is related to the strict concavity of F, hence the monotonicity assumption on U.

15

slide-36
SLIDE 36

MFG and planning

Variational characterization (planner’s problem) If there exists a function m → F(m) on P(E) such that DF = U, then any maximum of F is a Nash-MFG equilibrium. Remark 1: Sometimes, there exists a global problem whose solution corresponds to a MFG equilibrium. Remark 2: Uniqueness is related to the strict concavity of F, hence the monotonicity assumption on U. Put your towel on the beach

  • Objective function: U(x, m) = −x2 − γm(x).

15

slide-37
SLIDE 37

MFG and planning

Variational characterization (planner’s problem) If there exists a function m → F(m) on P(E) such that DF = U, then any maximum of F is a Nash-MFG equilibrium. Remark 1: Sometimes, there exists a global problem whose solution corresponds to a MFG equilibrium. Remark 2: Uniqueness is related to the strict concavity of F, hence the monotonicity assumption on U. Put your towel on the beach

  • Objective function: U(x, m) = −x2 − γm(x).
  • Global problem: F(m) =
  • −x2m(x)dx − γ

2m(x)2dx. 15

slide-38
SLIDE 38

MFG and planning

Variational characterization (planner’s problem) If there exists a function m → F(m) on P(E) such that DF = U, then any maximum of F is a Nash-MFG equilibrium. Remark 1: Sometimes, there exists a global problem whose solution corresponds to a MFG equilibrium. Remark 2: Uniqueness is related to the strict concavity of F, hence the monotonicity assumption on U. Put your towel on the beach

  • Objective function: U(x, m) = −x2 − γm(x).
  • Global problem: F(m) =
  • −x2m(x)dx − γ

2m(x)2dx.

  • Unique equilibrium, of the form m(x) = 1

γ (λ − x2)+ 15

slide-39
SLIDE 39

MFG in continuous time (with continuous state space)

slide-40
SLIDE 40

Differential games

Static games are interesting but MFGs are really powerful in continuous time (differential games): The real power of MFGs in continuous time

  • Differential/stochastic calculus.
  • Ordinary and partial differential equations.
  • Numerical methods.

16

slide-41
SLIDE 41

Differential games

Static games are interesting but MFGs are really powerful in continuous time (differential games): The real power of MFGs in continuous time

  • Differential/stochastic calculus.
  • Ordinary and partial differential equations.
  • Numerical methods.

Also, very general results have been obtained with probabilistic methods (see Carmona, Delarue).

16

slide-42
SLIDE 42

Reminder of (stochastic) optimal control

Agent’s dynamics dXt = αtdt + σdWt, X0 = x

17

slide-43
SLIDE 43

Reminder of (stochastic) optimal control

Agent’s dynamics dXt = αtdt + σdWt, X0 = x Objective function sup

(αs)s≥0

E T (f (Xs) − L(αs)) ds + g(XT)

  • 17
slide-44
SLIDE 44

Reminder of (stochastic) optimal control

Agent’s dynamics dXt = αtdt + σdWt, X0 = x Objective function sup

(αs)s≥0

E T (f (Xs) − L(αs)) ds + g(XT)

  • Remarks:
  • f and L can also include a time dependency (e.g. discount

rate).

  • Stationary (infinite horizon)/Ergodic problems can also be

considered.

17

slide-45
SLIDE 45

Reminder of (stochastic) optimal control

Main tool: value function The best “score” an agent can expect when he is in x at time t: u(t, x) = sup

(αs)s≥t

E T

t

(f (Xs) − L(αs)) ds + g(XT)|Xt = x

  • 18
slide-46
SLIDE 46

Reminder of (stochastic) optimal control

Main tool: value function The best “score” an agent can expect when he is in x at time t: u(t, x) = sup

(αs)s≥t

E T

t

(f (Xs) − L(αs)) ds + g(XT)|Xt = x

  • PDE

u “solves” the Hamilton-Jacobi(-Bellman) equation: ∂tu + σ2 2 ∆u + H(∇u) = −f (x), u(T, x) = g(x), where H(p) = supα α · p − L(α).

18

slide-47
SLIDE 47

Reminder of (stochastic) optimal control

Main tool: value function The best “score” an agent can expect when he is in x at time t: u(t, x) = sup

(αs)s≥t

E T

t

(f (Xs) − L(αs)) ds + g(XT)|Xt = x

  • PDE

u “solves” the Hamilton-Jacobi(-Bellman) equation: ∂tu + σ2 2 ∆u + H(∇u) = −f (x), u(T, x) = g(x), where H(p) = supα α · p − L(α). Optimal control The optimal control is α∗(t, x) = ∇H(∇u(t, x)).

18

slide-48
SLIDE 48

From optimal control problems to mean field games

  • Continuum of players.

19

slide-49
SLIDE 49

From optimal control problems to mean field games

  • Continuum of players.
  • Each player has a position X i that evolves according to:

dX i

t = αi tdt + σdW i t ,

X i

0 = xi

Remark: only independent idiosyncratic risks (common noise has also been studied but it is more complicated).

19

slide-50
SLIDE 50

From optimal control problems to mean field games

  • Continuum of players.
  • Each player has a position X i that evolves according to:

dX i

t = αi tdt + σdW i t ,

X i

0 = xi

Remark: only independent idiosyncratic risks (common noise has also been studied but it is more complicated).

  • Each player optimizes:

max

(αi

s)s≥0

E T

  • f (X i

s , m(s, ·)) − L(αi s, m(s, ·))

  • ds +g(X i

T, m(T, ·))

  • 19
slide-51
SLIDE 51

From optimal control problems to mean field games

  • Continuum of players.
  • Each player has a position X i that evolves according to:

dX i

t = αi tdt + σdW i t ,

X i

0 = xi

Remark: only independent idiosyncratic risks (common noise has also been studied but it is more complicated).

  • Each player optimizes:

max

(αi

s)s≥0

E T

  • f (X i

s , m(s, ·)) − L(αi s, m(s, ·))

  • ds +g(X i

T, m(T, ·))

  • The Nash-equilibrium t ∈ [0, T] → m(t, ·) must be consistent

with the decisions of the agents.

19

slide-52
SLIDE 52

Examples

Repulsion

  • f (x, m) = −m(t, x) − δx2 and g = 0.

→ Willingness to be close to 0 but far from other players.

  • Quadratic cost: L(α) = α2

2 . 20

slide-53
SLIDE 53

Examples

Repulsion

  • f (x, m) = −m(t, x) − δx2 and g = 0.

→ Willingness to be close to 0 but far from other players.

  • Quadratic cost: L(α) = α2

2 .

Congestion Cost of the form L(α, m(t, x)) = α2

2 (1 + m(t, x)). 20

slide-54
SLIDE 54

Partial differential equations

  • u value function of the control problem (with given m).
  • m distribution of the players

21

slide-55
SLIDE 55

Partial differential equations

  • u value function of the control problem (with given m).
  • m distribution of the players

MFG PDEs (HJB) ∂tu + σ2

2 ∆u + H(∇u, m) = −f (x, m)

(K) ∂tm + ∇ · (m∇pH(∇u, m)) = σ2

2 ∆m

where H(p, m) = supα α · p − L(α, m). u(T, x) = g(x), m(0, x) = m0(x)

21

slide-56
SLIDE 56

Partial differential equations

  • u value function of the control problem (with given m).
  • m distribution of the players

MFG PDEs (HJB) ∂tu + σ2

2 ∆u + H(∇u, m) = −f (x, m)

(K) ∂tm + ∇ · (m∇pH(∇u, m)) = σ2

2 ∆m

where H(p, m) = supα α · p − L(α, m). u(T, x) = g(x), m(0, x) = m0(x) The optimal control is α∗(t, x) = ∇pH(∇u(t, x), m(t, ·)).

21

slide-57
SLIDE 57

Remarks and variants

Forward/Backward The system of PDEs is a forward/backward problem:

  • The HJB equation is backward in time (terminal condition)

because agents anticipate the future.

  • The transport equation is forward in time because it

corresponds to the dynamics of the agents.

22

slide-58
SLIDE 58

Remarks and variants

Forward/Backward The system of PDEs is a forward/backward problem:

  • The HJB equation is backward in time (terminal condition)

because agents anticipate the future.

  • The transport equation is forward in time because it

corresponds to the dynamics of the agents. Other frameworks

  • Stationary setting (infinite horizon)
  • Ergodic setting

22

slide-59
SLIDE 59

Remarks and variants

Forward/Backward The system of PDEs is a forward/backward problem:

  • The HJB equation is backward in time (terminal condition)

because agents anticipate the future.

  • The transport equation is forward in time because it

corresponds to the dynamics of the agents. Other frameworks

  • Stationary setting (infinite horizon)
  • Ergodic setting

Related problem Same equations with initial and final conditions on m and no terminal condition on u: the problem is then that of finding the right terminal payoff g so that agents go from m0 to mT.

22

slide-60
SLIDE 60

Some results

23

slide-61
SLIDE 61

Some results

Existence A wide variety of PDE results, depending on f , L, g and σ.

23

slide-62
SLIDE 62

Some results

Existence A wide variety of PDE results, depending on f , L, g and σ. Uniqueness If the cost function L does not depend on m and if f is decreasing in the sense: ∀m1 = m2,

  • (f (x, m1) − f (x, m2))(m1 − m2) < 0

then a solution of the PDEs system is unique.

23

slide-63
SLIDE 63

Some results

Existence A wide variety of PDE results, depending on f , L, g and σ. Uniqueness If the cost function L does not depend on m and if f is decreasing in the sense: ∀m1 = m2,

  • (f (x, m1) − f (x, m2))(m1 − m2) < 0

then a solution of the PDEs system is unique. Remarks:

23

slide-64
SLIDE 64

Some results

Existence A wide variety of PDE results, depending on f , L, g and σ. Uniqueness If the cost function L does not depend on m and if f is decreasing in the sense: ∀m1 = m2,

  • (f (x, m1) − f (x, m2))(m1 − m2) < 0

then a solution of the PDEs system is unique. Remarks:

  • Same criterion as above.

23

slide-65
SLIDE 65

Some results

Existence A wide variety of PDE results, depending on f , L, g and σ. Uniqueness If the cost function L does not depend on m and if f is decreasing in the sense: ∀m1 = m2,

  • (f (x, m1) − f (x, m2))(m1 − m2) < 0

then a solution of the PDEs system is unique. Remarks:

  • Same criterion as above.
  • For more general cost functions L (e.g. congestion), there is a

more general criterion (see Lions, or see the result in graphs).

23

slide-66
SLIDE 66

MFG with quadratic cost/Hamiltonian

MFG equations with quadratic cost function L(α) = α2

2 on the

domain [0, T] × Ω, Ω standing for (0, 1)d: (HJB) ∂tu + σ2 2 ∆u + 1 2|∇u|2 = −f (x, m) (K) ∂tm + ∇ · (m∇u) = σ2 2 ∆m

24

slide-67
SLIDE 67

MFG with quadratic cost/Hamiltonian

MFG equations with quadratic cost function L(α) = α2

2 on the

domain [0, T] × Ω, Ω standing for (0, 1)d: (HJB) ∂tu + σ2 2 ∆u + 1 2|∇u|2 = −f (x, m) (K) ∂tm + ∇ · (m∇u) = σ2 2 ∆m Examples of conditions

  • Boundary conditions: ∂u

∂n = ∂m ∂n = 0 on (0, T) × ∂Ω

  • Terminal condition: u(T, x) = g(x).
  • Initial condition: m(0, x) = m0(x) ≥ 0.

24

slide-68
SLIDE 68

MFG with quadratic cost/Hamiltonian

MFG equations with quadratic cost function L(α) = α2

2 on the

domain [0, T] × Ω, Ω standing for (0, 1)d: (HJB) ∂tu + σ2 2 ∆u + 1 2|∇u|2 = −f (x, m) (K) ∂tm + ∇ · (m∇u) = σ2 2 ∆m Examples of conditions

  • Boundary conditions: ∂u

∂n = ∂m ∂n = 0 on (0, T) × ∂Ω

  • Terminal condition: u(T, x) = g(x).
  • Initial condition: m(0, x) = m0(x) ≥ 0.

The optimal control is α∗(t, x) = ∇u(t, x).

24

slide-69
SLIDE 69

Change of variables

Theorem: u = σ2 log(φ), m = φψ Let’s consider a smooth solution (φ, ψ) (with φ > 0) of: ∂tφ + σ2 2 ∆φ = − 1 σ2 f (x, φψ)φ (Eφ) ∂tψ − σ2 2 ∆ψ = 1 σ2 f (x, φψ)ψ (Eψ)

  • Boundary conditions: ∂φ

∂n = ∂ψ ∂n = 0 on (0, T) × ∂Ω

  • Terminal condition: φ(T, ·) = exp
  • uT (·)

σ2

  • .
  • Initial condition: ψ(0, ·) = m0(·)

φ(0,·)

Then (u, m) = (σ2 log(φ), φψ) is a solution of (MFG).

25

slide-70
SLIDE 70

Change of variables

Theorem: u = σ2 log(φ), m = φψ Let’s consider a smooth solution (φ, ψ) (with φ > 0) of: ∂tφ + σ2 2 ∆φ = − 1 σ2 f (x, φψ)φ (Eφ) ∂tψ − σ2 2 ∆ψ = 1 σ2 f (x, φψ)ψ (Eψ)

  • Boundary conditions: ∂φ

∂n = ∂ψ ∂n = 0 on (0, T) × ∂Ω

  • Terminal condition: φ(T, ·) = exp
  • uT (·)

σ2

  • .
  • Initial condition: ψ(0, ·) = m0(·)

φ(0,·)

Then (u, m) = (σ2 log(φ), φψ) is a solution of (MFG). Nice existence results exist on this system (see some of my papers).

25

slide-71
SLIDE 71

Numerics and examples

slide-72
SLIDE 72

Numerical methods

  • Variational formulation: when a global maximization problem

exists, gradient-descent/ascent can be used (see Lachapelle, Salomon, Turinici)

  • Finite difference method (Achdou and Capuzzo-Dolcetta)
  • Specific methods in the quadratic cost case (see Gu´

eant).

26

slide-73
SLIDE 73

Examples with population dynamics

Toy problem in the quadratic case

  • f (x, ξ) = −16(x − 1/2)2 − 0.1 max(0, min(5, ξ)), i.e. agents

want to live near x = 1

2 but they do not want to live together.

  • T = 0.5
  • g = 0
  • σ = 1
  • m0(x) =

µ(x) 1

0 µ(x′)dx′ , where

µ(x) = 1 + 0.2 cos

  • π
  • 2x − 3

2 2 .

27

slide-74
SLIDE 74

Toy problem in the quadratic case

The functions φ and ψ.

28

slide-75
SLIDE 75

Toy problem in the quadratic case

The dynamics of the distribution m.

29

slide-76
SLIDE 76

Examples with population dynamics (videos provided by Y. Achdou)

Going out of a movie theater (1)

  • We consider a movie theatre with 6 rows, and 2 doors in the

front to exit.

  • Neumann conditions on walls.
  • Homogenous Dirichlet conditions at the doors.
  • Running penalty while staying in the room.
  • Congestion effects.

30

slide-77
SLIDE 77

Examples with population dynamics (videos provided by Y. Achdou)

Going out of a movie theater (2)

  • The same movie theatre with 6 rows, and 2 doors in the front

to exit.

  • One door only will be open at a pre-defined time, but nobody

knows which one.

31

slide-78
SLIDE 78

Numerous economic applications

Many models in economics and finance – for instance:

  • Interaction between economic growth and inequalities (where

Pareto distributions play a central role). → See Gu´ eant, Lasry, Lions (Paris-Princeton lectures). → Similar ideas developed by Lucas and Moll.

32

slide-79
SLIDE 79

Numerous economic applications

Many models in economics and finance – for instance:

  • Interaction between economic growth and inequalities (where

Pareto distributions play a central role). → See Gu´ eant, Lasry, Lions (Paris-Princeton lectures). → Similar ideas developed by Lucas and Moll.

  • Competition between asset managers.

→ Gu´ eant (Risk and decision analysis, 2013)

32

slide-80
SLIDE 80

Numerous economic applications

Many models in economics and finance – for instance:

  • Interaction between economic growth and inequalities (where

Pareto distributions play a central role). → See Gu´ eant, Lasry, Lions (Paris-Princeton lectures). → Similar ideas developed by Lucas and Moll.

  • Competition between asset managers.

→ Gu´ eant (Risk and decision analysis, 2013)

  • Oil extraction (`

a la Hotelling) with noise. → See Gu´ eant, Lasry, Lions (Paris-Princeton lectures).

32

slide-81
SLIDE 81

Numerous economic applications

Many models in economics and finance – for instance:

  • Interaction between economic growth and inequalities (where

Pareto distributions play a central role). → See Gu´ eant, Lasry, Lions (Paris-Princeton lectures). → Similar ideas developed by Lucas and Moll.

  • Competition between asset managers.

→ Gu´ eant (Risk and decision analysis, 2013)

  • Oil extraction (`

a la Hotelling) with noise. → See Gu´ eant, Lasry, Lions (Paris-Princeton lectures).

  • A long-term model for the mining industries.

→ Joint work of Achdou, Giraud, Lasry, Lions, and Scheinkman.

32

slide-82
SLIDE 82

Special for Huawei: MFG on graphs

slide-83
SLIDE 83

Framework

MFGs are often written on continuous state spaces, but what about discrete structures?

33

slide-84
SLIDE 84

Framework

MFGs are often written on continuous state spaces, but what about discrete structures? Notations for graph

33

slide-85
SLIDE 85

Framework

MFGs are often written on continuous state spaces, but what about discrete structures? Notations for graph

  • Graph G. Nodes indexed by integers from 1 to N.

33

slide-86
SLIDE 86

Framework

MFGs are often written on continuous state spaces, but what about discrete structures? Notations for graph

  • Graph G. Nodes indexed by integers from 1 to N.
  • ∀i ∈ N = {1, . . . , N}:
  • V(i) ⊂ N \ {i} the set of nodes j for which a directed edge

exists from i to j (cardinal: di).

  • V−1(i) ⊂ N \ {i} the set of nodes j for which a directed edge

exists from j to i.

33

slide-87
SLIDE 87

Framework (continued)

Players, strategies, and costs

34

slide-88
SLIDE 88

Framework (continued)

Players, strategies, and costs

  • Each player’s position: Markov chain (Xt)t with values in G.

34

slide-89
SLIDE 89

Framework (continued)

Players, strategies, and costs

  • Each player’s position: Markov chain (Xt)t with values in G.
  • Instantaneous transition probabilities at time t:

λt(i, ·) : V(i) → R+ (∀i ∈ N)

34

slide-90
SLIDE 90

Framework (continued)

Players, strategies, and costs

  • Each player’s position: Markov chain (Xt)t with values in G.
  • Instantaneous transition probabilities at time t:

λt(i, ·) : V(i) → R+ (∀i ∈ N)

  • Instantaneous cost L(i, (λi,j)j∈V(i)) to set the value of λ(i, j)

to λi,j.

34

slide-91
SLIDE 91

Hypotheses

Hypotheses on L

  • Super-linearity hypothesis:

∀i ∈ N, lim

λ∈R

di + ,|λ|→+∞

L(i, λ) |λ| = +∞

  • Convexity hypothesis:

∀i ∈ N, λ ∈ Rdi

+ → L(i, λ) is strictly convex. 35

slide-92
SLIDE 92

Hypotheses

Hypotheses on L

  • Super-linearity hypothesis:

∀i ∈ N, lim

λ∈R

di + ,|λ|→+∞

L(i, λ) |λ| = +∞

  • Convexity hypothesis:

∀i ∈ N, λ ∈ Rdi

+ → L(i, λ) is strictly convex.

Also, we define: ∀i ∈ N, p ∈ Rdi → H(i, p) = sup

λ∈R

di +

λ · p − L(i, λ).

35

slide-93
SLIDE 93

Mean field game - control problem

Control problem

36

slide-94
SLIDE 94

Mean field game - control problem

Control problem

  • Admissible Markovian controls:

A =

  • (λt(i, j))t∈[0,T],i∈N,j∈V(i) |t → λt(i, j) ∈ L∞(0, T)
  • 36
slide-95
SLIDE 95

Mean field game - control problem

Control problem

  • Admissible Markovian controls:

A =

  • (λt(i, j))t∈[0,T],i∈N,j∈V(i) |t → λt(i, j) ∈ L∞(0, T)
  • For λ ∈ A and a given function m : [0, T] → PN we define

the payoff function: Jm : [0, T] × N × A → R by: Jm(t, i, λ) = E T

t

(−L(Xs, λs(Xs, ·)) + f (Xs, m(s))) ds +g (XT, m(T))

  • for (Xs)s∈[t,T] a Markov chain on G, starting from i at time t,

with instantaneous transition probabilities given by (λs)s∈[t,T].

36

slide-96
SLIDE 96

Mean field game - control problem

Control problem

  • Admissible Markovian controls:

A =

  • (λt(i, j))t∈[0,T],i∈N,j∈V(i) |t → λt(i, j) ∈ L∞(0, T)
  • For λ ∈ A and a given function m : [0, T] → PN we define

the payoff function: Jm : [0, T] × N × A → R by: Jm(t, i, λ) = E T

t

(−L(Xs, λs(Xs, ·)) + f (Xs, m(s))) ds +g (XT, m(T))

  • for (Xs)s∈[t,T] a Markov chain on G, starting from i at time t,

with instantaneous transition probabilities given by (λs)s∈[t,T].

36

slide-97
SLIDE 97

Nash equilibrium

Nash-MFG equilibrium A differentiable function m : t ∈ [0, T] → (m(t, i))i ∈ PN is said to be a Nash-MFG equilibrium, if there exists an admissible control λ ∈ A such that: ∀˜ λ ∈ A, ∀i ∈ N, Jm(0, i, λ) ≥ Jm(0, i, ˜ λ) and ∀i ∈ N, d dt m(t, i) =

  • j∈V−1(i)

λt(j, i)m(t, j) −

  • j∈V(i)

λt(i, j)m(t, i) In that case, λ is called an optimal control.

37

slide-98
SLIDE 98

The G-MFG equations

Definition (The G-MFG equations) The G-MFG equations consist in a system of 2N equations, the unknown being t ∈ [0, T] → (u(t), m(t)): ∀i ∈ N, d dt u(t, i)+H

  • i, (u(t, j) − u(t, i))j∈V(i)
  • +f (i, m(t)) = 0,

∀i, d dt m(t, i) =

  • j∈V−1(i)

∂H(j, ·) ∂pi

  • (u(t, k) − u(t, j))k∈V(j)
  • m(t, j)

  • j∈V(i)

∂H(i, ·) ∂pj

  • (u(t, k) − u(t, i))k∈V(i)
  • m(t, i)

with u(T, i) = g(i, m(T)) and m(0) = m0 ∈ PN given.

38

slide-99
SLIDE 99

The G-MFG equations

Proposition (The G-MFG equations as a sufficient condition) Let m0 ∈ PN and let us consider a C 1 solution (u(t), m(t)) of the G-MFG equations with (m(0, 1), . . . , m(0, N)) = m0.

39

slide-100
SLIDE 100

The G-MFG equations

Proposition (The G-MFG equations as a sufficient condition) Let m0 ∈ PN and let us consider a C 1 solution (u(t), m(t)) of the G-MFG equations with (m(0, 1), . . . , m(0, N)) = m0. Then:

  • t → m(t) = (m(t, 1), . . . , m(t, N)) is a Nash-MFG equilibrium
  • The relations λt(i, j) = ∂H(i,·)

∂pj

  • (u(t, k) − u(t, i))k∈V(i)
  • define an optimal control.

39

slide-101
SLIDE 101

Existence of a solution

40

slide-102
SLIDE 102

Existence of a solution

Proposition (Existence of a solution to the G-MFG equations) Let m0 ∈ PN. Under the assumptions made above, there exists a C 1 solution (u, m) of the G-MFG equations such that m(0) = m0.

40

slide-103
SLIDE 103

Existence of a solution

Proposition (Existence of a solution to the G-MFG equations) Let m0 ∈ PN. Under the assumptions made above, there exists a C 1 solution (u, m) of the G-MFG equations such that m(0) = m0. Sketch of proof (Fixed point):

  • Comparison principle leads a priori bounds on u

sup

i∈N

u(·, i)∞ ≤ sup

i∈N

g(i, ·)∞ +

  • sup

i∈N

f (i, ·)∞ + sup

i∈N

|H(i, 0)|

  • T.
  • ⇒ bounds on dm

dt .

  • Ascoli + Schauder to conclude.

40

slide-104
SLIDE 104

Uniqueness of smooth solutions

Proposition (Uniqueness for the solution of the G-MFG equations)

41

slide-105
SLIDE 105

Uniqueness of smooth solutions

Proposition (Uniqueness for the solution of the G-MFG equations) Assume that f and g are such that: ∀(m, µ) ∈ PN ×PN,

N

  • i=1

(f (i, m)−f (i, µ))(mi −µi) ≥ 0 = ⇒ m = µ and ∀(m, µ) ∈ PN×PN,

N

  • i=1

(g(i, m)−g(i, µ))(mi−µi) ≥ 0 = ⇒ m = µ

41

slide-106
SLIDE 106

Uniqueness of smooth solutions

Proposition (Uniqueness for the solution of the G-MFG equations) Assume that f and g are such that: ∀(m, µ) ∈ PN ×PN,

N

  • i=1

(f (i, m)−f (i, µ))(mi −µi) ≥ 0 = ⇒ m = µ and ∀(m, µ) ∈ PN×PN,

N

  • i=1

(g(i, m)−g(i, µ))(mi−µi) ≥ 0 = ⇒ m = µ Then, if ( u, m) and (˜ u, ˜ m) are two C 1 solutions of the G-MFG equations, we have m = ˜ m and u = ˜ u.

41

slide-107
SLIDE 107

The G-Master equations

Definition (The G-Master equations) The G-Master equations consist in N equations, the unknown being (t, m) ∈ [0, T] × PN → (U1(t, m), . . . , UN(t, m)). ∀i ∈ N, ∂Ui ∂t (t, m) + H

  • i, (Uj(t, m) − Ui(t, m))j∈V(i)
  • +

N

  • l=1

∂Ui ∂ml (t, m)  

  • j∈V−1(l)

∂H(j, ·) ∂pl

  • (Uk(t, m) − Uj(t, m))k∈V(j)
  • mj

  • j∈V(l)

∂H(l, ·) ∂pj

  • (Uk(t, m) − Ul(t, m))k∈V(l)
  • ml

  + f (i, m) = 0 with Ui(T, m) = g(i, m).

42

slide-108
SLIDE 108

The G-Master equations

Proposition (From G-Master equations to G-MFG equations)

43

slide-109
SLIDE 109

The G-Master equations

Proposition (From G-Master equations to G-MFG equations) If (t, m) ∈ [0, T] × PN → (U1(t, m), . . . , UN(t, m)) is a C 1 solution to the G-Master equations.

43

slide-110
SLIDE 110

The G-Master equations

Proposition (From G-Master equations to G-MFG equations) If (t, m) ∈ [0, T] × PN → (U1(t, m), . . . , UN(t, m)) is a C 1 solution to the G-Master equations. If a function m is such that m(0) = m0 ∈ PN and d

dt m(t, i) =

  • j∈V−1(i)

∂H(j, ·) ∂pi

  • (Uk(t, m(t)) − Uj(t, m(t)))k∈V(j)
  • m(t, j)

  • j∈V(i)

∂H(i, ·) ∂pj

  • (Uk(t, m(t)) − Ui(t, m(t)))k∈V(i)
  • m(t, i)

43

slide-111
SLIDE 111

The G-Master equations

Proposition (From G-Master equations to G-MFG equations) If (t, m) ∈ [0, T] × PN → (U1(t, m), . . . , UN(t, m)) is a C 1 solution to the G-Master equations. If a function m is such that m(0) = m0 ∈ PN and d

dt m(t, i) =

  • j∈V−1(i)

∂H(j, ·) ∂pi

  • (Uk(t, m(t)) − Uj(t, m(t)))k∈V(j)
  • m(t, j)

  • j∈V(i)

∂H(i, ·) ∂pj

  • (Uk(t, m(t)) − Ui(t, m(t)))k∈V(i)
  • m(t, i)

Then t ∈ [0, T] → (U1(t, m(t)), . . . , UN(t, m(t)), m(t)) is a solution of the G-MFG equations.

43

slide-112
SLIDE 112

Potential games

Assumptions We suppose that there exist two C 1 functions: F : (m1, . . . , mN) ∈ PN → F(m1, . . . , mN) G : (m1, . . . , mN) ∈ PN → G(m1, . . . , mN) such that ∀i ∈ N: ∂F ∂mi = f (i, ·) ∂G ∂mi = g(i, ·)

44

slide-113
SLIDE 113

Planning problem

We introduce for t ∈ [0, T], mt ∈ PN and a given admissible control (function) λ ∈ A, the payoff function J (t, mt, λ) = T

t

  • F(m(s)) −

N

  • i=1

L(i, (λs(i, j))j∈V(i))m(s, i)

  • ds + G(m(T))

where ∀i ∈ N, m(t, i) = mt

i and ∀i ∈ N, ∀s ∈ [t, T]:

d ds m(s, i) =

  • j∈V−1(i)

λs(j, i)m(s, j) −

  • j∈V(i)

λs(i, j)m(s, i) Optimization problem The (deterministic) optimization problem we consider is, for a given m0 ∈ PN: sup

λ∈A

J (0, m0, λ).

45

slide-114
SLIDE 114

HJ equation

Definition (The G-planning equation) The G-planning equation consists in one PDE in Φ(t, m): ∂Φ ∂t (t, m1, . . . , mN) + H(m1, . . . , mN, ∇Φ) + F(m1, . . . , mN) = 0 with the terminal conditions Φ(T, m) = G(m), where: H(m, p) = sup

(λi,j)i∈N ,j∈V(i) N

  • i=1

 

  • j∈V−1(i)

λj,imj −

  • j∈V(i)

λi,jmi   pi −L(i, (λi,j)j∈V(i))mi

  • =

N

  • i=1

miH

  • i, (pj − pi)j∈V(i)
  • 46
slide-115
SLIDE 115

Solving the planning problem

Proposition Let us consider a C 1 function Φ solution of the G-planning

  • equation. Then, Φ restricted to [0, T] × PN is the value function
  • f the above planning problem, i.e.:

∀(t, mt) ∈ [0, T] × PN, Φ(t, mt) = sup

λ∈A

J (t, mt, λ)

47

slide-116
SLIDE 116

Solving the planning problem

Proposition Moreover, if we define ∀i ∈ N, m(t, i) = mt

i and

∀i ∈ N, ∀s ∈ [t, T] d ds m(s, i) =

  • j∈V−1(i)

λs(j, i)m(s, j) −

  • j∈V(i)

λs(i, j)m(s, i) with λs(i, j) = ∂H(i, ·) ∂pj ∂Φ ∂mk (s, m(s)) − ∂Φ ∂mi (s, m(s))

  • k∈V(i)
  • then λ is an optimal control for the planning problem.

48

slide-117
SLIDE 117

Going back to MFG

Proposition

49

slide-118
SLIDE 118

Going back to MFG

Proposition Let Φ be a C 2 solution of the G-planning equation.

49

slide-119
SLIDE 119

Going back to MFG

Proposition Let Φ be a C 2 solution of the G-planning equation. Define ∀i ∈ N, ∀t ∈ [0, T], ∀m ∈ PN, Ui(t, m) = ∂Φ

∂mi (t, m). 49

slide-120
SLIDE 120

Going back to MFG

Proposition Let Φ be a C 2 solution of the G-planning equation. Define ∀i ∈ N, ∀t ∈ [0, T], ∀m ∈ PN, Ui(t, m) = ∂Φ

∂mi (t, m).

Then, ∇Φ = U = (U1, . . . , UN) verifies the G-Master equations.

49

slide-121
SLIDE 121

Going back to MFG

Proposition Let Φ be a C 2 solution of the G-planning equation. Define ∀i ∈ N, ∀t ∈ [0, T], ∀m ∈ PN, Ui(t, m) = ∂Φ

∂mi (t, m).

Then, ∇Φ = U = (U1, . . . , UN) verifies the G-Master equations. Consequently, if we define ∀i ∈ N, m(0, i) = m0

i for a given

m0 ∈ PN and ∀i ∈ N, ∀s ∈ [t, T]:

d ds m(s, i) = j∈V−1(i) λs(j, i)m(s, j) − j∈V(i) λs(i, j)m(s, i)

with λs(i, j) = ∂H(i,·)

∂pj

  • ∂Φ

∂mk (s, m(s)) − ∂Φ ∂mi (s, m(s))

  • k∈V(i)
  • then m is a Nash-MFG equilibrium and λ is an optimal control for

the initial mean field game (the decentralized problem).

49

slide-122
SLIDE 122

Extending to models with congestion

50

slide-123
SLIDE 123

Extending to models with congestion

We can extend existence and uniqueness of solutions of the G-MFG equations to more general Hamiltonians.

50

slide-124
SLIDE 124

Extending to models with congestion

We can extend existence and uniqueness of solutions of the G-MFG equations to more general Hamiltonians. We are not limited to L(i, (λi,j)j∈V(i), m) = L(i, (λi,j)j∈V(i)) − f (i, m)

50

slide-125
SLIDE 125

Extending to models with congestion

We can extend existence and uniqueness of solutions of the G-MFG equations to more general Hamiltonians. We are not limited to L(i, (λi,j)j∈V(i), m) = L(i, (λi,j)j∈V(i)) − f (i, m) Assumptions

  • Continuity: ∀i ∈ N, L(i, ·, ·) is a continuous function from

Rdi

+ × PN to R

  • Asymptotic super-linearity:

∀i ∈ N, ∀m ∈ PN, limλ∈R

di + ,|λ|→+∞

L(i,λ,m) |λ|

= +∞

50

slide-126
SLIDE 126

Extending to models with congestion

Hamiltonian functions: ∀i ∈ N, p ∈ Rdi, m ∈ PN → H(i, p, m) = sup

λ∈R

di +

λ · p − L(i, λ, m)

51

slide-127
SLIDE 127

Extending to models with congestion

Hamiltonian functions: ∀i ∈ N, p ∈ Rdi, m ∈ PN → H(i, p, m) = sup

λ∈R

di +

λ · p − L(i, λ, m) Hypotheses

  • ∀i ∈ N, H(i, ·, ·) is a continuous function.
  • ∀i ∈ N, ∀m ∈ PN, H(i, ·, m) is a C 1 function with:

∂H ∂p (i, p, m) = argmaxλ∈Rdi

+ λ · p − L(i, λ, m)

  • ∀i ∈ N, ∀j ∈ V(i), ∂H

∂pj (i, ·, ·) is a continuous function. 51

slide-128
SLIDE 128

Extending to models with congestion - Existence

Using the same proof as above: Proposition (Existence) Under the assumptions made above, there exists a C 1 solution (u, m) of the G-MFG equations.

52

slide-129
SLIDE 129

Extending to models with congestion - Uniqueness

Proposition (Uniqueness)

53

slide-130
SLIDE 130

Extending to models with congestion - Uniqueness

Proposition (Uniqueness) Assume that g is such that: ∀(m, µ) ∈ PN×PN,

N

  • i=1

(g(i, m)−g(i, µ))(mi−µi) ≥ 0 = ⇒ m = µ

53

slide-131
SLIDE 131

Extending to models with congestion - Uniqueness

Proposition (Uniqueness) Assume that g is such that: ∀(m, µ) ∈ PN×PN,

N

  • i=1

(g(i, m)−g(i, µ))(mi−µi) ≥ 0 = ⇒ m = µ Assume that the hamiltonian functions can be written as: ∀i ∈ N, ∀p ∈ Rdi, ∀m ∈ PN, H(i, p, m) = Hc(i, p, m) + f (i, m) with ∀i ∈ N, f (i, ·) a continuous function satisfying ∀(m, µ) ∈ PN ×PN,

N

  • i=1

(f (i, m)−f (i, µ))(mi −µi) ≥ 0 = ⇒ m = µ

53

slide-132
SLIDE 132

Extending to models with congestion - Uniqueness

Proposition (Uniqueness (continued)) and ∀i ∈ N, Hc(i, ·, ·) a C 1 function with: ∀j ∈ V(i), ∂Hc

∂pj (i, ·, ·) a C 1 function on Rn × PN 54

slide-133
SLIDE 133

Extending to models with congestion - Uniqueness

Proposition (Uniqueness (continued)) and ∀i ∈ N, Hc(i, ·, ·) a C 1 function with: ∀j ∈ V(i), ∂Hc

∂pj (i, ·, ·) a C 1 function on Rn × PN

Now, let us define A : (q1, . . . , qN, m) ∈ N

i=1 Rdi × PN → (αij(qi, m))i,j ∈ MN

defined by: αij(qi, m) = −∂Hc ∂mj (i, qi, m)

54

slide-134
SLIDE 134

Extending to models with congestion - Uniqueness

Proposition (Uniqueness (continued)) and ∀i ∈ N, Hc(i, ·, ·) a C 1 function with: ∀j ∈ V(i), ∂Hc

∂pj (i, ·, ·) a C 1 function on Rn × PN

Now, let us define A : (q1, . . . , qN, m) ∈ N

i=1 Rdi × PN → (αij(qi, m))i,j ∈ MN

defined by: αij(qi, m) = −∂Hc ∂mj (i, qi, m) Let us also define, ∀i ∈ N, Bi : (qi, m) ∈ Rdi × PN →

  • βi

jk(qi, m)

  • j,k ∈ MN,di defined by:

βi

jk(qi, m) = mi

∂2Hc ∂mj∂qik (i, qi, m)

54

slide-135
SLIDE 135

Extending to models with congestion - Uniqueness

Proposition (Uniqueness (continued))

55

slide-136
SLIDE 136

Extending to models with congestion - Uniqueness

Proposition (Uniqueness (continued)) Let us also define, ∀i ∈ N, C i : (qi, m) ∈ Rdi × PN →

  • γi

jk(qi, m)

  • j,k ∈ Mdi,N defined by:

γi

jk(qi, m) = mi

∂2Hc ∂mk∂qij (i, qi, m)

55

slide-137
SLIDE 137

Extending to models with congestion - Uniqueness

Proposition (Uniqueness (continued)) Let us also define, ∀i ∈ N, C i : (qi, m) ∈ Rdi × PN →

  • γi

jk(qi, m)

  • j,k ∈ Mdi,N defined by:

γi

jk(qi, m) = mi

∂2Hc ∂mk∂qij (i, qi, m) Let us finally define, ∀i ∈ N, Di : (qi, m) ∈ Rdi × PN →

  • δi

jk(qi, m)

  • j,k ∈ Mdi defined by:

δi

jk(qi, m) = mi

∂2Hc ∂qij∂qik (i, qi, m)

55

slide-138
SLIDE 138

Extending to models with congestion - Uniqueness

Proposition (Uniqueness (continued))

56

slide-139
SLIDE 139

Extending to models with congestion - Uniqueness

Proposition (Uniqueness (continued)) Assume that ∀(q1, . . . , qN, m) ∈ N

i=1 Rdi × PN

            A(q1, . . . , qN, m) B1(q1, m) · · · · · · · · · BN(qN, m) C 1(q1, m) D1(q1, m) · · · · · · . . . ... ... . . . . . . . . . ... ... ... . . . . . . . . . ... ... C N(qN, m) · · · · · · DN(qN, m)             ≥ 0

56

slide-140
SLIDE 140

Extending to models with congestion - Uniqueness

Proposition (Uniqueness (continued)) Assume that ∀(q1, . . . , qN, m) ∈ N

i=1 Rdi × PN

            A(q1, . . . , qN, m) B1(q1, m) · · · · · · · · · BN(qN, m) C 1(q1, m) D1(q1, m) · · · · · · . . . ... ... . . . . . . . . . ... ... ... . . . . . . . . . ... ... C N(qN, m) · · · · · · DN(qN, m)             ≥ 0 Then, if ( u, m) and (˜ u, ˜ m) are two C 1 solutions of the G-MFG equations, we have m = ˜ m and u = ˜ u.

56

slide-141
SLIDE 141

Conclusion

slide-142
SLIDE 142

Advantages and limitations of MFGs

57

slide-143
SLIDE 143

Advantages and limitations of MFGs

Advantages

57

slide-144
SLIDE 144

Advantages and limitations of MFGs

Advantages

  • Large class of problems can be modeled.

57

slide-145
SLIDE 145

Advantages and limitations of MFGs

Advantages

  • Large class of problems can be modeled.
  • Benefit from centuries of differential calculus.

57

slide-146
SLIDE 146

Advantages and limitations of MFGs

Advantages

  • Large class of problems can be modeled.
  • Benefit from centuries of differential calculus.
  • Numerical methods to solve PDEs are available.

57

slide-147
SLIDE 147

Advantages and limitations of MFGs

Advantages

  • Large class of problems can be modeled.
  • Benefit from centuries of differential calculus.
  • Numerical methods to solve PDEs are available.
  • Possibility to extend the framework to several continuums of

agents and to add big players.

57

slide-148
SLIDE 148

Advantages and limitations of MFGs

Advantages

  • Large class of problems can be modeled.
  • Benefit from centuries of differential calculus.
  • Numerical methods to solve PDEs are available.
  • Possibility to extend the framework to several continuums of

agents and to add big players.

  • Possibility to have common noise, but equations are more

difficult (Master equation).

57

slide-149
SLIDE 149

Advantages and limitations of MFGs

Advantages

  • Large class of problems can be modeled.
  • Benefit from centuries of differential calculus.
  • Numerical methods to solve PDEs are available.
  • Possibility to extend the framework to several continuums of

agents and to add big players.

  • Possibility to have common noise, but equations are more

difficult (Master equation). Limitations/Drawbacks

57

slide-150
SLIDE 150

Advantages and limitations of MFGs

Advantages

  • Large class of problems can be modeled.
  • Benefit from centuries of differential calculus.
  • Numerical methods to solve PDEs are available.
  • Possibility to extend the framework to several continuums of

agents and to add big players.

  • Possibility to have common noise, but equations are more

difficult (Master equation). Limitations/Drawbacks

  • Only rational/perfect expectations → models are not flexible

enough sometimes.

57

slide-151
SLIDE 151

Advantages and limitations of MFGs

Advantages

  • Large class of problems can be modeled.
  • Benefit from centuries of differential calculus.
  • Numerical methods to solve PDEs are available.
  • Possibility to extend the framework to several continuums of

agents and to add big players.

  • Possibility to have common noise, but equations are more

difficult (Master equation). Limitations/Drawbacks

  • Only rational/perfect expectations → models are not flexible

enough sometimes.

  • Numerical methods on graphs have not been proposed...

maybe Reinforcement Learning.

57

slide-152
SLIDE 152

The End

Thank you. Questions?

58