Learning interacting kernels of mean-field equations of particle - - PowerPoint PPT Presentation

learning interacting kernels of mean field equations of
SMART_READER_LITE
LIVE PREVIEW

Learning interacting kernels of mean-field equations of particle - - PowerPoint PPT Presentation

Learning interacting kernels of mean-field equations of particle systems Fei Lu Department of Mathematics, Johns Hopkins University Joint work with Quanjun Lang Related work with: Mauro Maggioni, Sui Tang, Ming Zhong, Zhongyang Li and Cheng


slide-1
SLIDE 1

Learning interacting kernels of mean-field equations of particle systems

Fei Lu

Department of Mathematics, Johns Hopkins University Joint work with Quanjun Lang Related work with: Mauro Maggioni, Sui Tang, Ming Zhong, Zhongyang Li and Cheng Zhang October 28, 2020 Junior Colloquium, JHU

FL acknowledges supports from JHU, NSF

slide-2
SLIDE 2

An inverse problem Nonparametric Learning Numerical examples

Outline

1

Motivation and problem statement

2

Nonparametric Learning

3

Numerical examples

4

Ongoing work and open problems

2 / 24

slide-3
SLIDE 3

An inverse problem Nonparametric Learning Numerical examples Motivation Previous work

An inverse problem

Consider the mean-field equation ∂tu = ν∆u + ∇ · [u(Kφ ∗ u)], x ∈ Rd, t > 0, where Kφ(x) = ∇(Φ(|x|)) = φ(|x|) x

|x|.

Question: identify φ from data {u(xm, tl)}M,L

m,l=1?

Goal: An algorithm → φ identifiability: function space of learning convergence rate when ∆x = M−1/d → 0

3 / 24

slide-4
SLIDE 4

An inverse problem Nonparametric Learning Numerical examples Motivation Previous work

Motivation

∂tu = ν∆u + ∇ · [u(Kφ ∗ u)]

Interacting particles/agents: d dt X i

t = 1

N

N

  • i′=1

φ(|X j

t − X i t |) X j t − X i t

|X j

t − X i t |

+ √ 2νdBi

t,

i = 1, . . . , N X i

t : the i-th particle’s position; Bi t: Brownian motion

u(x, t) = limN→∞ N

i=1 δ(X i t − x) Propagation of chaos

1st- and 2nd-order models Application in many disciplines:

Statistical physics, quantum mechanics Biology [Keller-Segal1970, Cucker-Smale2000] Social science [Motsch-Tadmor2014] Monte Carlo sampling [Del Moral13] Epidemiology (Agent-based model for COVID19 at Imperial)

4 / 24

slide-5
SLIDE 5

An inverse problem Nonparametric Learning Numerical examples Motivation Previous work

Previous work: finite N

∂tu = ν∆u + ∇ · [u(Kφ ∗ u)] Interacting particles/agents: d dt X i

t = 1

N

N

  • i′=1

φ(|X j

t − X i t |) X j t − X i t

|X j

t − X i t |

+ √ 2νdBi

t,

i = 1, . . . , N Maggioni JHU team: [M., L., Tang, Zhong, Miller, Li, Zhang: PNAS19, SPA20, etc] Data: many trajectories {X (m)

[0,T]}M m=1, ν = 0; ν > 0, finite N

Function space of learning: φ ∈ L2(ρT) with ρT ← |X j

t − X i t |

Nonparametric estimation (Ac = b)

))

Opinion Dynamics Lennard-Jones Prey-Predator

5 / 24

slide-6
SLIDE 6

An inverse problem Nonparametric Learning Numerical examples Motivation Previous work

Previous work: finite N

∂tu = ν∆u + ∇ · [u(Kφ ∗ u)] Interacting particles/agents: d dt X i

t = 1

N

N

  • i′=1

φ(|X j

t − X i t |) X j t − X i t

|X j

t − X i t |

+ √ 2νdBi

t, i = 1, . . . , N

Maggioni JHU team: [M., L., Tang, Zhong, Miller, Li, Zhang: PNAS19, SPA20, etc] Identifiability: a coercivity condition for L2(ρT) Optimal convergence rate: Eµ0[ φT,M,Hn∗ − φtrueL2(ρT )] ≤ C ((log M)/M)

s 2s+1 .

2.5 3 3.5 4 log10(M)

  • 1.6
  • 1.4
  • 1.2
  • 1
  • 0.8
  • 0.6

log10(Rel Err)

Rel Err Slope=-0.34 Optimal decay 12 13 14 15 16 17 18 19 20 21

log2(M)

  • 10
  • 9
  • 8
  • 7
  • 6
  • 5

log2(error) Learning rate errors slope -0.36

  • ptimal decay

Opinion Dynamics Lennard-Jones Prey-Predator

6 / 24

slide-7
SLIDE 7

An inverse problem Nonparametric Learning Numerical examples Motivation Previous work

What if N → ∞? Data: many trajectories {X (m)

[0,T]}M m=1;

Data: density u(x, t) = limN→∞ N

i=1 δ(X i t − x)

{u(xm, tl)}M,L

m,l=1

∂tu = ν∆u + ∇ · [u(Kφ ∗ u)]

7 / 24

slide-8
SLIDE 8

An inverse problem Nonparametric Learning Numerical examples Motivation Previous work

What if N → ∞? Data: many trajectories {X (m)

[0,T]}M m=1;

Data: density u(x, t) = limN→∞ N

i=1 δ(X i t − x)

{u(xm, tl)}M,L

m,l=1

∂tu = ν∆u + ∇ · [u(Kφ ∗ u)] How to estimate φ from data? Minimize E0(ψ) = T

  • Rd
  • ∇.(u(Kψ ∗ u)) − g
  • 2dx dt?

(with g = ∂tu − ν∆u ) Derivatives not available from data.

8 / 24

slide-9
SLIDE 9

An inverse problem Nonparametric Learning Numerical examples A probabilistic error functional Identifiability Convergence rate

Outline

1

Motivation and problem statement

2

Nonparametric learning

◮ A probabilistic error functional ◮ Identifiability: function spaces of learning ◮ Rate of convergence 3

Numerical examples

4

Ongoing work and open problems

9 / 24

slide-10
SLIDE 10

An inverse problem Nonparametric Learning Numerical examples A probabilistic error functional Identifiability Convergence rate

A probabilistic error functional

E(ψ) := 1 T T

  • Rd
  • Kψ ∗ u
  • 2u − 2νu(∇ · Kψ ∗ u) + 2∂tu(Ψ ∗ u)
  • dx dt

= ψ, ψ GT − 2 ψ, φ GT Expectation of the negative log-likelihood of the process

  • dX t = − Kφ ∗ u(X t, t)dt +

√ 2νdBt, L(X t) = u(·, t), Derivative-in-space free! GT is a reproducing kernel for a RKHS

  • φ, ψ

GT := 1 T T

  • Rd (Kφ ∗ u), (Kψ ∗ u)u(x, t)dx dt =
  • R+
  • R+ φ(r)ψ(s) GT (r, s) dr ds

ψ = n

i=1 ciφi ⇒ E(ψ) = c⊤Ac − 2b⊤c with Aij =

φi, φj GT ⇒ Estimator:

  • φn =

n

  • i=1
  • ciφi,
  • c = A−1b

10 / 24

slide-11
SLIDE 11

An inverse problem Nonparametric Learning Numerical examples A probabilistic error functional Identifiability Convergence rate

Discrete data

From data {u(xm, tl)}M,L

m,l=1: Hn = span{φi}n i=1,

  • φn,M,L =

n

  • i=1
  • ci

n,M,Lφi,

with cn,M,L = A−1

n,M,Lbn,M,L.

Inverse problem: well-posed/ identifiable, A−1? Choice of Hn: {φi} and n ? Convergence rate when ∆x = M−1/d → 0? → hypothesis testing and model selection

11 / 24

slide-12
SLIDE 12

An inverse problem Nonparametric Learning Numerical examples A probabilistic error functional Identifiability Convergence rate

Invertibility of A and function space

Recall that H = span{φi}n

i=1,

Aij =

  • φi, φj
  • GT ,

with integral kernel GT → RKHS HGT . if {φi} orthonormal in HGT : A = In if {φi} orthonormal in L2(ρT): minimal eigenvalue of A = cH,T = inf

ψ∈H,ψL2(ρT )=1

ψ, ψ GT > 0 (Coercivity condition)

◮ measure ρT ← |X t − X

′ t| (“pairwise distance”)

12 / 24

slide-13
SLIDE 13

An inverse problem Nonparametric Learning Numerical examples A probabilistic error functional Identifiability Convergence rate

Error bounds

H = L2(ρT) or RKHS HGT . Theorem (Lang-Lu20)

Let H = span{φi}n

i=1 and

φn the projection of φ on H ⊂ H. Assume regularity conditions. Then

  • φn,M,L −

φnH ≤ 2cH,T

−1

Cb√ n + CAn φH

  • (∆x + ∆t),

If if H = L2(ρT): assume coercivity condition on H with cH,T > 0, if H = RKHS, set cH,T = 1 ∆x + ∆t comes from numerical integrator (Riemann sum) Dominating order: n∆x (if ∆t = 0)

13 / 24

slide-14
SLIDE 14

An inverse problem Nonparametric Learning Numerical examples A probabilistic error functional Identifiability Convergence rate

Optimal dimension and rate of convergence

Total error: trade-off

  • φn,M,∞ − φH ≤

φn,M,∞ − φnH

  • inference error

+

  • φn − φH
  • approximation error

Theorem (Lang-Lu20) Assume φn,M,∞ − φnH n(∆x)α and φn − φH n−s. Then, with

  • ptimal dimension n ≈ (∆x)−α/(s+1):
  • φn,M,∞ − φH (∆x)αs/(s+1)

14 / 24

slide-15
SLIDE 15

An inverse problem Nonparametric Learning Numerical examples Smooth kernel Non-smooth kernel Singular kernel

Outline

1

Motivation and problem statement

2

Nonparametric learning

3

Numerical examples

◮ Granular media: smooth kernel φ(r) = 3r 2 ◮ Opinion dynamics: piecewise linear φ ◮ Repulsion-attraction: singular φ 4

Ongoing work and open problems

15 / 24

slide-16
SLIDE 16

An inverse problem Nonparametric Learning Numerical examples Smooth kernel Non-smooth kernel Singular kernel

Numerical example 1: granular media

∂tu = ν∆u + ∇ · [u(Kφ ∗ u)], x ∈ Rd, t > 0, where Kφ(x) = ∇(Φ(|x|)) = φ(|x|) x

|x|. φ(r) = 3r 2

0.5 1 Time t 0.005 0.01 0.015 0.02 0.025 Wasserstein distance Original initial New initial

The solution u(x, t) Estimators of φ Wasserstein W2(u, u)

16 / 24

slide-17
SLIDE 17

An inverse problem Nonparametric Learning Numerical examples Smooth kernel Non-smooth kernel Singular kernel

Numerical example 1: granular media

The solution u(x, t) Estimators of φ

10-1 100 x 10-1 100 Test point Slope = 1.64 Optimal = 2.00 10-1 100 x

  • 6.58 + 10-5
  • 6.58 + 10-4
  • 6.58 + 10-3
  • 6.58 + 10-2
  • 6.58 + 10-1
  • 6.58 + 100

Error functional E Test point Slope = 4.05 Optimal = 4.00

Convergence rate of L2(ρT) error Convergence rate of EM,L almost optimal almost optimal

17 / 24

slide-18
SLIDE 18

An inverse problem Nonparametric Learning Numerical examples Smooth kernel Non-smooth kernel Singular kernel

Numerical Example 2: opinion dynamics

∂tu = ν∆u + ∇ · [u(Kφ ∗ u)], x ∈ Rd, t > 0, where Kφ(x) = ∇(Φ(|x|)) = φ(|x|) x

|x|. φ(r) piecewise linear

0.2 0.4 0.6 0.8 1 Time t 2 4 6 8 Wasserstein distance 10-3 Original initial New initial

The solution u(x, t) Estimators of φ Wasserstein W2(u, u)

18 / 24

slide-19
SLIDE 19

An inverse problem Nonparametric Learning Numerical examples Smooth kernel Non-smooth kernel Singular kernel

Numerical Example 2: opinion dynamics

The solution u(x, t) Estimators of φ

10-1 100 x 1 1.5 2 2.5 3 3.5 4 Test point Slope = 0.74 Optimal = 2.00 10-1 100 x

  • 0.43 + 10-4
  • 0.43 + 10-3
  • 0.43 + 10-2
  • 0.43 + 10-1
  • 0.43 + 100

Error functional E Test point Slope = 3.00 Optimal = 4.00

Convergence rate of L2(ρT) error Convergence rate of EM,L sub-optimal (φ / ∈ W 1,∞) sub-optimal

19 / 24

slide-20
SLIDE 20

An inverse problem Nonparametric Learning Numerical examples Smooth kernel Non-smooth kernel Singular kernel

Numerical example 3: repulsion-attraction

∂tu = ν∆u + ∇ · [u(Kφ ∗ u)], x ∈ Rd, t > 0, where Kφ(x) = ∇(Φ(|x|)) = φ(|x|) x

|x|. φ(r) = r − r −1.5 singular

0.2 0.4 0.6 0.8 1 Time t 1 2 3 4 Wasserstein distance 10-3 Original initial New initial

The solution u(x, t) Estimators of φ Wasserstein W2(u, u)

20 / 24

slide-21
SLIDE 21

An inverse problem Nonparametric Learning Numerical examples Smooth kernel Non-smooth kernel Singular kernel

Numerical example 3: repulsion-attraction

The solution u(x, t) Estimators of φ

10-1 100 x 6 7 8 9 10 11 Test point Slope = 0.23 Optimal = 1.33 10-1 100 x

  • 3.00 + 0.30
  • 3.00 + 0.35
  • 3.00 + 0.40
  • 3.00 + 0.45
  • 3.00 + 0.50

Error functional E Test point Slope = 0.30 Optimal = 2.67

Convergence rate of L2(ρT) error Convergence rate of EM,L low rate: theory does not apply low rate: theory does not apply

21 / 24

slide-22
SLIDE 22

An inverse problem Nonparametric Learning Numerical examples Smooth kernel Non-smooth kernel Singular kernel

Summary and open problems

Problem: Estimate φ of Mean-field equation

∂tu = ν∆u + ∇ · [u(Kφ ∗ u)] from discrete data {u(xm, tl)}M,L

m,l=1.

Solution: Algorithm A probabilistic error functional Estimator by least squares Theory guidance Choice of hypothesis space basis functions & dimension Function space of learning: RKHS v.s. L2(ρT) Optimal learning rate

22 / 24

slide-23
SLIDE 23

An inverse problem Nonparametric Learning Numerical examples Smooth kernel Non-smooth kernel Singular kernel

Open problem and future directions Learning/computation:

◮ 2nd-order systems ◮ High-dimensional state space (Monte Carlo) ◮ non-radial interaction kernel ◮ partial observation of large systems

Coercivity condition on L2(ρT)

◮ GT → strictly positive integral operator?

Singular kernels? Real data applications: learning cell-dynamics

23 / 24

slide-24
SLIDE 24

An inverse problem Nonparametric Learning Numerical examples Smooth kernel Non-smooth kernel Singular kernel

References

Quanjun Lang, F.Lu. Learning interaction kernels in mean-field equations of 1st-order systems of interacting

  • particles. In preparation. 2020
  • F. Lu, M.Maggioni, and S. Tang. Learning interaction kernels in stochastic systems of interacting particles

from multiple trajectories. arXiv2007 Zehong Zhang and F. Lu, Cluster prediction for opinion dynamics from partial observations. arXiv2007

  • Z. Li, F. Lu, S. Tang, C. Zhang, and M. Maggioni. On the identifiability of interaction functions of particle
  • systems. to appear on SPA20
  • F. Lu, M. Maggioni, and S. Tang. Learning interaction kernels in heterogeneous systems of agents from

multiple trajectories. arXiv1912

  • F. Lu, M. Maggioni, S. Tang and M. Zhong. Nonparametric inference of interaction laws in systems of agents

from trajectory data. PNAS, 2019

  • M. Bongini, M. Fornasier, M. Maggioni and M. Hansen. Inferring Interaction Rules From Observations of

Evolutive Systems I: The Variational Approach. M3AS, 27(05), 909-951, 2017

Thank you!

24 / 24