Energy minimization via conic programming hierarchies David de Laat - - PowerPoint PPT Presentation

energy minimization via conic programming hierarchies
SMART_READER_LITE
LIVE PREVIEW

Energy minimization via conic programming hierarchies David de Laat - - PowerPoint PPT Presentation

Energy minimization via conic programming hierarchies David de Laat (TU Delft) IFORS July 14, 2014, Barcelona Energy minimization What is the minimal potential energy E when we distribute N particles in a container V with pair potential w ?


slide-1
SLIDE 1

Energy minimization via conic programming hierarchies

David de Laat (TU Delft)

IFORS July 14, 2014, Barcelona

slide-2
SLIDE 2

Energy minimization

◮ What is the minimal potential energy E when we distribute N

particles in a container V with pair potential w?

slide-3
SLIDE 3

Energy minimization

◮ What is the minimal potential energy E when we distribute N

particles in a container V with pair potential w?

◮ Example: For the Thomson problem we take

V = S2 and w({x, y}) = 1 x − y

slide-4
SLIDE 4

Energy minimization

◮ What is the minimal potential energy E when we distribute N

particles in a container V with pair potential w?

◮ Example: For the Thomson problem we take

V = S2 and w({x, y}) = 1 x − y

◮ Optimization problem:

E = inf

S∈(V

N)

  • P∈(S

2)

w(P)

slide-5
SLIDE 5

Approach

◮ Configurations provide upper bounds on the optimal energy E

slide-6
SLIDE 6

Approach

◮ Configurations provide upper bounds on the optimal energy E ◮ To prove a configuration is good (or optimal) we need good

lower bounds for E

slide-7
SLIDE 7

Approach

◮ Configurations provide upper bounds on the optimal energy E ◮ To prove a configuration is good (or optimal) we need good

lower bounds for E

◮ For this we use infinite dimensional moment hierarchies and

semidefinite programming

slide-8
SLIDE 8

Approach

E

slide-9
SLIDE 9

Approach

E

Difficult minimization problem

slide-10
SLIDE 10

Approach

Et E

Difficult minimization problem

slide-11
SLIDE 11

Approach

Et E

Difficult minimization problem Relaxation to a conic program: Infinite dimensional minimization problem

slide-12
SLIDE 12

Approach

Et E E∗

t

Difficult minimization problem Relaxation to a conic program: Infinite dimensional minimization problem

slide-13
SLIDE 13

Approach

Et E E∗

t

Difficult minimization problem Relaxation to a conic program: Conic dual: Infinite dimensional minimization problem Infinite dimensional maximization problem

slide-14
SLIDE 14

Approach

Et E E∗

t

E∗

t,d

Difficult minimization problem Relaxation to a conic program: Conic dual: Infinite dimensional minimization problem Infinite dimensional maximization problem

slide-15
SLIDE 15

Approach

Et E E∗

t

E∗

t,d

Difficult minimization problem Relaxation to a conic program: Conic dual: Semi-infinite semidefinite program Infinite dimensional minimization problem Infinite dimensional maximization problem

slide-16
SLIDE 16

Finite container

◮ If V = {1, . . . , n} is a finite set, then E is a polynomial

  • ptimization problem:

E = min

  • {i,j}∈(V

2)

w({i, j})xixj : x ∈ {0, 1}n,

  • i∈V

xi = N

slide-17
SLIDE 17

Finite container

◮ If V = {1, . . . , n} is a finite set, then E is a polynomial

  • ptimization problem:

E = min

  • {i,j}∈(V

2)

w({i, j})xixj : x ∈ {0, 1}n,

  • i∈V

xi = N

  • ◮ The Lasserre hierarchy gives a chain E1 ≤ E2 ≤ · · · ≤ En of

lower bounds to the optimal energy E:

slide-18
SLIDE 18

Finite container

◮ If V = {1, . . . , n} is a finite set, then E is a polynomial

  • ptimization problem:

E = min

  • {i,j}∈(V

2)

w({i, j})xixj : x ∈ {0, 1}n,

  • i∈V

xi = N

  • ◮ The Lasserre hierarchy gives a chain E1 ≤ E2 ≤ · · · ≤ En of

lower bounds to the optimal energy E: Et = min

S∈(V

2)

w(S)y(S) : y ∈ R( V

≤2t), y(∅) = 1,

  • y(A ∪ B)
  • A,B∈( V

≤t) 0,

  • x∈V

y(T ∪ {x}) = Ny(T) for T ∈

  • V

≤2t−1

slide-19
SLIDE 19

Infinite container

◮ Assume V is a compact Hausdorff space and w continuous

slide-20
SLIDE 20

Infinite container

◮ Assume V is a compact Hausdorff space and w continuous ◮ V ≤t

  • \ {∅} gets its topology as a quotient of V t
slide-21
SLIDE 21

Infinite container

◮ Assume V is a compact Hausdorff space and w continuous ◮ V ≤t

  • \ {∅} gets its topology as a quotient of V t

◮ Generalization (here s = min{2t, N}):

Et = min

  • λ(w) : λ ∈ M(

V

≤s

  • )≥0, A∗

t λ ∈ M(

V

≤t

  • ×

V

≤t

  • )0,

λ( V

i

  • ) =

N

i

  • for i = 0, . . . , s
slide-22
SLIDE 22

Infinite container

◮ Assume V is a compact Hausdorff space and w continuous ◮ V ≤t

  • \ {∅} gets its topology as a quotient of V t

◮ Generalization (here s = min{2t, N}):

Et = min

  • λ(w) : λ ∈ M(

V

≤s

  • )≥0, A∗

t λ ∈ M(

V

≤t

  • ×

V

≤t

  • )0,

λ( V

i

  • ) =

N

i

  • for i = 0, . . . , s
  • ◮ λ generalizes the moment vector y
slide-23
SLIDE 23

Infinite container

◮ Assume V is a compact Hausdorff space and w continuous ◮ V ≤t

  • \ {∅} gets its topology as a quotient of V t

◮ Generalization (here s = min{2t, N}):

Et = min

  • λ(w) : λ ∈ M(

V

≤s

  • )≥0, A∗

t λ ∈ M(

V

≤t

  • ×

V

≤t

  • )0,

λ( V

i

  • ) =

N

i

  • for i = 0, . . . , s
  • ◮ λ generalizes the moment vector y

◮ M(

V

≤t

  • ×

V

≤t

  • )0 is dual to the cone C(

V

≤t

  • ×

V

≤t

  • )0 of

positive definite kernels

slide-24
SLIDE 24

Infinite container

◮ Assume V is a compact Hausdorff space and w continuous ◮ V ≤t

  • \ {∅} gets its topology as a quotient of V t

◮ Generalization (here s = min{2t, N}):

Et = min

  • λ(w) : λ ∈ M(

V

≤s

  • )≥0, A∗

t λ ∈ M(

V

≤t

  • ×

V

≤t

  • )0,

λ( V

i

  • ) =

N

i

  • for i = 0, . . . , s
  • ◮ λ generalizes the moment vector y

◮ M(

V

≤t

  • ×

V

≤t

  • )0 is dual to the cone C(

V

≤t

  • ×

V

≤t

  • )0 of

positive definite kernels

◮ Relaxation: If S is an N subset of V , then

χS =

  • R∈( S

≤2t)

δR is feasible for Et

slide-25
SLIDE 25

Infinite container

◮ Assume V is a compact Hausdorff space and w continuous ◮ V ≤t

  • \ {∅} gets its topology as a quotient of V t

◮ Generalization (here s = min{2t, N}):

Et = min

  • λ(w) : λ ∈ M(

V

≤s

  • )≥0, A∗

t λ ∈ M(

V

≤t

  • ×

V

≤t

  • )0,

λ( V

i

  • ) =

N

i

  • for i = 0, . . . , s
  • ◮ λ generalizes the moment vector y

◮ M(

V

≤t

  • ×

V

≤t

  • )0 is dual to the cone C(

V

≤t

  • ×

V

≤t

  • )0 of

positive definite kernels

◮ Relaxation: If S is an N subset of V , then

χS =

  • R∈( S

≤2t)

δR is feasible for Et

◮ We have EN = E

slide-26
SLIDE 26

Infinite container

◮ Assume V is a compact Hausdorff space and w continuous ◮ V ≤t

  • \ {∅} gets its topology as a quotient of V t

◮ Generalization (here s = min{2t, N}):

Et = min

  • λ(w) : λ ∈ M(

V

≤s

  • )≥0, A∗

t λ ∈ M(

V

≤t

  • ×

V

≤t

  • )0,

λ( V

i

  • ) =

N

i

  • for i = 0, . . . , s
  • ◮ λ generalizes the moment vector y

◮ M(

V

≤t

  • ×

V

≤t

  • )0 is dual to the cone C(

V

≤t

  • ×

V

≤t

  • )0 of

positive definite kernels

◮ Relaxation: If S is an N subset of V , then

χS =

  • R∈( S

≤2t)

δR is feasible for Et

◮ We have EN = E ◮ Uses techniques from [de Laat-Vallentin 2013]: hierarchy for

packing problems in discrete geometry

slide-27
SLIDE 27

Dual hierarchy

◮ For lower bounds we need feasible solutions of the dual

slide-28
SLIDE 28

Dual hierarchy

◮ For lower bounds we need feasible solutions of the dual ◮ In the dual hierarchy optimization is over scalars ai and

positive definite kernels K ∈ C( V

≤t

  • ×

V

≤t

  • )0:
slide-29
SLIDE 29

Dual hierarchy

◮ For lower bounds we need feasible solutions of the dual ◮ In the dual hierarchy optimization is over scalars ai and

positive definite kernels K ∈ C( V

≤t

  • ×

V

≤t

  • )0:

E∗

t = sup

  • s
  • i=0

N

i

  • ai : a0, . . . , as ∈ R, K ∈ C(

V

≤t

  • ×

V

≤t

  • )0,

ai − AtK ≤ w on V

i

  • for i = 0, . . . , s
slide-30
SLIDE 30

Dual hierarchy

◮ For lower bounds we need feasible solutions of the dual ◮ In the dual hierarchy optimization is over scalars ai and

positive definite kernels K ∈ C( V

≤t

  • ×

V

≤t

  • )0:

E∗

t = sup

  • s
  • i=0

N

i

  • ai : a0, . . . , as ∈ R, K ∈ C(

V

≤t

  • ×

V

≤t

  • )0,

ai − AtK ≤ w on V

i

  • for i = 0, . . . , s
  • ◮ Techniquality: we only put a linear constraint for S ∈

V

i

  • if

the points in S are not too close

slide-31
SLIDE 31

Dual hierarchy

◮ For lower bounds we need feasible solutions of the dual ◮ In the dual hierarchy optimization is over scalars ai and

positive definite kernels K ∈ C( V

≤t

  • ×

V

≤t

  • )0:

E∗

t = sup

  • s
  • i=0

N

i

  • ai : a0, . . . , as ∈ R, K ∈ C(

V

≤t

  • ×

V

≤t

  • )0,

ai − AtK ≤ w on V

i

  • for i = 0, . . . , s
  • ◮ Techniquality: we only put a linear constraint for S ∈

V

i

  • if

the points in S are not too close

◮ Strong duality holds: Et = E∗ t

slide-32
SLIDE 32

Dual hierarchy

◮ For lower bounds we need feasible solutions of the dual ◮ In the dual hierarchy optimization is over scalars ai and

positive definite kernels K ∈ C( V

≤t

  • ×

V

≤t

  • )0:

E∗

t = sup

  • s
  • i=0

N

i

  • ai : a0, . . . , as ∈ R, K ∈ C(

V

≤t

  • ×

V

≤t

0,

ai − AtK ≤ w on V

i

  • for i = 0, . . . , s
  • ◮ Techniquality: we only put a linear constraint for S ∈

V

i

  • if

the points in S are not too close

◮ Strong duality holds: Et = E∗ t ◮ If Γ acts on V and w is Γ-invariant, then we can restrict to

Γ-invariant kernels: K(γJ, γJ′) = K(J, J′) for all J, J′ ∈ V

≤t

  • (Here γ{x1, . . . , xt} = {γx1, . . . , γxt})
slide-33
SLIDE 33

Inner approximiations to the cone C( V

≤t

  • ×

V

≤t

◮ Nested chain of inner approximations:

C1 ⊆ C2 ⊆ · · · ⊆ C( V

≤t

  • ×

V

≤t

slide-34
SLIDE 34

Inner approximiations to the cone C( V

≤t

  • ×

V

≤t

◮ Nested chain of inner approximations:

C1 ⊆ C2 ⊆ · · · ⊆ C( V

≤t

  • ×

V

≤t

◮ Each cone Ci can be parametrized by a finite direct sum of

positive semidefinite matrix cones

slide-35
SLIDE 35

Inner approximiations to the cone C( V

≤t

  • ×

V

≤t

◮ Nested chain of inner approximations:

C1 ⊆ C2 ⊆ · · · ⊆ C( V

≤t

  • ×

V

≤t

◮ Each cone Ci can be parametrized by a finite direct sum of

positive semidefinite matrix cones

◮ Bochner: A kernel K ∈ C(

V

≤t

  • ×

V

≤t

0 is of the form

K(J, J′) =

  • k=0

trace(FkZk(J, J′))

◮ Fk: (infinite) positive semidefinite matrices (the Fourier

coefficients)

◮ Zk: zonal matrices corresponding to the action of Γ on

V

≤t

  • (generalizes e2πikx in the Fourier transform on the circle)
slide-36
SLIDE 36

Inner approximiations to the cone C( V

≤t

  • ×

V

≤t

◮ Nested chain of inner approximations:

C1 ⊆ C2 ⊆ · · · ⊆ C( V

≤t

  • ×

V

≤t

◮ Each cone Ci can be parametrized by a finite direct sum of

positive semidefinite matrix cones

◮ Bochner: A kernel K ∈ C(

V

≤t

  • ×

V

≤t

0 is of the form

K(J, J′) =

  • k=0

trace(FkZk(J, J′))

◮ Fk: (infinite) positive semidefinite matrices (the Fourier

coefficients)

◮ Zk: zonal matrices corresponding to the action of Γ on

V

≤t

  • (generalizes e2πikx in the Fourier transform on the circle)

◮ Define Cd by truncating the above series

slide-37
SLIDE 37

The semi-infinite semidefinite programs E∗

t,d

◮ Define E∗ t,d by replacing the cone C(

V

≤t

  • ×

V

≤t

0 in E∗ t by

the cone Cd

slide-38
SLIDE 38

The semi-infinite semidefinite programs E∗

t,d

◮ Define E∗ t,d by replacing the cone C(

V

≤t

  • ×

V

≤t

0 in E∗ t by

the cone Cd

◮ This is an optimization problem with finitely many variables

and infinitely many constraints

slide-39
SLIDE 39

The semi-infinite semidefinite programs E∗

t,d

◮ Define E∗ t,d by replacing the cone C(

V

≤t

  • ×

V

≤t

0 in E∗ t by

the cone Cd

◮ This is an optimization problem with finitely many variables

and infinitely many constraints

◮ E∗ t,d → E∗ t as d → ∞ follows from ∪∞ d=0Cd being uniformly

dense in C( V

≤t

  • ×

V

≤t

slide-40
SLIDE 40

Example: V = S1 with O(2)-invariant pair potential w

◮ The linear constraints in E∗ t,d can be written as the

nonnegativity of a trigonometric polynomial in s − 1 variables

slide-41
SLIDE 41

Example: V = S1 with O(2)-invariant pair potential w

◮ The linear constraints in E∗ t,d can be written as the

nonnegativity of a trigonometric polynomial in s − 1 variables

◮ Use trigonometric SOS characterizations [Dumitrescu 2006]

slide-42
SLIDE 42

Example: V = S1 with O(2)-invariant pair potential w

◮ The linear constraints in E∗ t,d can be written as the

nonnegativity of a trigonometric polynomial in s − 1 variables

◮ Use trigonometric SOS characterizations [Dumitrescu 2006] ◮ For the Coulomb potential (or other completely monotonic

potentials) the regular N-gon is the optimal configuration on the circle [Cohn-Kumar 2006]

slide-43
SLIDE 43

Example: V = S1 with O(2)-invariant pair potential w

◮ The linear constraints in E∗ t,d can be written as the

nonnegativity of a trigonometric polynomial in s − 1 variables

◮ Use trigonometric SOS characterizations [Dumitrescu 2006] ◮ For the Coulomb potential (or other completely monotonic

potentials) the regular N-gon is the optimal configuration on the circle [Cohn-Kumar 2006]

◮ Uses relaxation based on the 2-point correlation function

[Yudin 1992] (This is similar to E1)

slide-44
SLIDE 44

Example: V = S1 with O(2)-invariant pair potential w

◮ The linear constraints in E∗ t,d can be written as the

nonnegativity of a trigonometric polynomial in s − 1 variables

◮ Use trigonometric SOS characterizations [Dumitrescu 2006] ◮ For the Coulomb potential (or other completely monotonic

potentials) the regular N-gon is the optimal configuration on the circle [Cohn-Kumar 2006]

◮ Uses relaxation based on the 2-point correlation function

[Yudin 1992] (This is similar to E1)

◮ The bound E∗ 2 requires SOS characterizations in 3 variables

slide-45
SLIDE 45

Example: V = S1 with O(2)-invariant pair potential w

◮ The linear constraints in E∗ t,d can be written as the

nonnegativity of a trigonometric polynomial in s − 1 variables

◮ Use trigonometric SOS characterizations [Dumitrescu 2006] ◮ For the Coulomb potential (or other completely monotonic

potentials) the regular N-gon is the optimal configuration on the circle [Cohn-Kumar 2006]

◮ Uses relaxation based on the 2-point correlation function

[Yudin 1992] (This is similar to E1)

◮ The bound E∗ 2 requires SOS characterizations in 3 variables ◮ Lennard-Jones potential: Based on a sampling implementation

it appears that for e.g. N = 3 we have E1 < E2 = E

slide-46
SLIDE 46

Thank you!