High precision computations for energy minimization David de Laat - - PowerPoint PPT Presentation

high precision computations for energy minimization
SMART_READER_LITE
LIVE PREVIEW

High precision computations for energy minimization David de Laat - - PowerPoint PPT Presentation

High precision computations for energy minimization David de Laat (CWI Amsterdam) Real algebraic geometry with a view toward moment problems and optimization, 6 March 2017, MFO Energy minimization Problem: Find the ground state energy of a


slide-1
SLIDE 1

High precision computations for energy minimization

David de Laat (CWI Amsterdam) Real algebraic geometry with a view toward moment problems and optimization, 6 March 2017, MFO

slide-2
SLIDE 2

Energy minimization

◮ Problem: Find the ground state energy of a system of N

particles in a compact metric space (V, d) with pair potential h

slide-3
SLIDE 3

Energy minimization

◮ Problem: Find the ground state energy of a system of N

particles in a compact metric space (V, d) with pair potential h

◮ Example: In the Thomson problem we minimize

  • 1≤i<j≤N

1 xi − xj2

  • ver all sets {x1, . . . , xN} of N distinct points in S2 ⊆ R3
slide-4
SLIDE 4

Energy minimization

◮ Problem: Find the ground state energy of a system of N

particles in a compact metric space (V, d) with pair potential h

◮ Example: In the Thomson problem we minimize

  • 1≤i<j≤N

1 xi − xj2

  • ver all sets {x1, . . . , xN} of N distinct points in S2 ⊆ R3

◮ Here V = S2, d(x, y) = x − y2, and h(w) = 1/w

slide-5
SLIDE 5

Energy minimization

◮ Problem: Find the ground state energy of a system of N

particles in a compact metric space (V, d) with pair potential h

◮ Example: In the Thomson problem we minimize

  • 1≤i<j≤N

1 xi − xj2

  • ver all sets {x1, . . . , xN} of N distinct points in S2 ⊆ R3

◮ Here V = S2, d(x, y) = x − y2, and h(w) = 1/w ◮ Assume h(w) → ∞ as w → 0

slide-6
SLIDE 6

Energy minimization

◮ Problem: Find the ground state energy of a system of N

particles in a compact metric space (V, d) with pair potential h

◮ Example: In the Thomson problem we minimize

  • 1≤i<j≤N

1 xi − xj2

  • ver all sets {x1, . . . , xN} of N distinct points in S2 ⊆ R3

◮ Here V = S2, d(x, y) = x − y2, and h(w) = 1/w ◮ Assume h(w) → ∞ as w → 0 ◮ Use moment techniques to find lower bounds (obstructions)

slide-7
SLIDE 7

Energy minimization

◮ Problem: Find the ground state energy of a system of N

particles in a compact metric space (V, d) with pair potential h

◮ Example: In the Thomson problem we minimize

  • 1≤i<j≤N

1 xi − xj2

  • ver all sets {x1, . . . , xN} of N distinct points in S2 ⊆ R3

◮ Here V = S2, d(x, y) = x − y2, and h(w) = 1/w ◮ Assume h(w) → ∞ as w → 0 ◮ Use moment techniques to find lower bounds (obstructions) ◮ Infinite dimensional moment techniques → computations

slide-8
SLIDE 8

Energy minimization

◮ Problem: Find the ground state energy of a system of N

particles in a compact metric space (V, d) with pair potential h

◮ Example: In the Thomson problem we minimize

  • 1≤i<j≤N

1 xi − xj2

  • ver all sets {x1, . . . , xN} of N distinct points in S2 ⊆ R3

◮ Here V = S2, d(x, y) = x − y2, and h(w) = 1/w ◮ Assume h(w) → ∞ as w → 0 ◮ Use moment techniques to find lower bounds (obstructions) ◮ Infinite dimensional moment techniques → computations

(Compute sharp lower bound for the N = 5 case)

slide-9
SLIDE 9

Setup

◮ Let B be an upper bound on the minimal energy

slide-10
SLIDE 10

Setup

◮ Let B be an upper bound on the minimal energy ◮ Define a graph with vertex set V where two distinct vertices x

and y are adjacent if h(d(x, y)) > B

slide-11
SLIDE 11

Setup

◮ Let B be an upper bound on the minimal energy ◮ Define a graph with vertex set V where two distinct vertices x

and y are adjacent if h(d(x, y)) > B

◮ Let It be the set of independent sets with ≤ t elements

slide-12
SLIDE 12

Setup

◮ Let B be an upper bound on the minimal energy ◮ Define a graph with vertex set V where two distinct vertices x

and y are adjacent if h(d(x, y)) > B

◮ Let It be the set of independent sets with ≤ t elements ◮ Let I=t be the set of independent sets with t elements

slide-13
SLIDE 13

Setup

◮ Let B be an upper bound on the minimal energy ◮ Define a graph with vertex set V where two distinct vertices x

and y are adjacent if h(d(x, y)) > B

◮ Let It be the set of independent sets with ≤ t elements ◮ Let I=t be the set of independent sets with t elements ◮ These sets are compact metric spaces

slide-14
SLIDE 14

Setup

◮ Let B be an upper bound on the minimal energy ◮ Define a graph with vertex set V where two distinct vertices x

and y are adjacent if h(d(x, y)) > B

◮ Let It be the set of independent sets with ≤ t elements ◮ Let I=t be the set of independent sets with t elements ◮ These sets are compact metric spaces ◮ Define f ∈ C(IN) by

f(S) =

  • h(d(x, y))

if S = {x, y} with x = y,

  • therwise
slide-15
SLIDE 15

Setup

◮ Let B be an upper bound on the minimal energy ◮ Define a graph with vertex set V where two distinct vertices x

and y are adjacent if h(d(x, y)) > B

◮ Let It be the set of independent sets with ≤ t elements ◮ Let I=t be the set of independent sets with t elements ◮ These sets are compact metric spaces ◮ Define f ∈ C(IN) by

f(S) =

  • h(d(x, y))

if S = {x, y} with x = y,

  • therwise

◮ Ground state energy:

E = min

S∈I=N

  • P⊆S

f(P)

slide-16
SLIDE 16

Moment relaxations

◮ For S ∈ I=N, define the measure χS = R⊆S δR on IN

slide-17
SLIDE 17

Moment relaxations

◮ For S ∈ I=N, define the measure χS = R⊆S δR on IN ◮ We can use this measure to compute the energy of S

slide-18
SLIDE 18

Moment relaxations

◮ For S ∈ I=N, define the measure χS = R⊆S δR on IN ◮ We can use this measure to compute the energy of S ◮ The energy of S is given by

χS(f) =

  • f(P) dχS(P) =
  • R⊆S

f(R) =

  • {x,y}∈I=2

h(d(x, y))

slide-19
SLIDE 19

Moment relaxations

◮ For S ∈ I=N, define the measure χS = R⊆S δR on IN ◮ We can use this measure to compute the energy of S ◮ The energy of S is given by

χS(f) =

  • f(P) dχS(P) =
  • R⊆S

f(R) =

  • {x,y}∈I=2

h(d(x, y))

◮ This measure satisfies the following 3 properties:

slide-20
SLIDE 20

Moment relaxations

◮ For S ∈ I=N, define the measure χS = R⊆S δR on IN ◮ We can use this measure to compute the energy of S ◮ The energy of S is given by

χS(f) =

  • f(P) dχS(P) =
  • R⊆S

f(R) =

  • {x,y}∈I=2

h(d(x, y))

◮ This measure satisfies the following 3 properties:

◮ χS is a positive measure

slide-21
SLIDE 21

Moment relaxations

◮ For S ∈ I=N, define the measure χS = R⊆S δR on IN ◮ We can use this measure to compute the energy of S ◮ The energy of S is given by

χS(f) =

  • f(P) dχS(P) =
  • R⊆S

f(R) =

  • {x,y}∈I=2

h(d(x, y))

◮ This measure satisfies the following 3 properties:

◮ χS is a positive measure ◮ χS satisfies χS(I=i) =

N

i

  • for all i
slide-22
SLIDE 22

Moment relaxations

◮ For S ∈ I=N, define the measure χS = R⊆S δR on IN ◮ We can use this measure to compute the energy of S ◮ The energy of S is given by

χS(f) =

  • f(P) dχS(P) =
  • R⊆S

f(R) =

  • {x,y}∈I=2

h(d(x, y))

◮ This measure satisfies the following 3 properties:

◮ χS is a positive measure ◮ χS satisfies χS(I=i) =

N

i

  • for all i

◮ χS is a measure of positive type (see next slide)

slide-23
SLIDE 23

Moment relaxations

◮ For S ∈ I=N, define the measure χS = R⊆S δR on IN ◮ We can use this measure to compute the energy of S ◮ The energy of S is given by

χS(f) =

  • f(P) dχS(P) =
  • R⊆S

f(R) =

  • {x,y}∈I=2

h(d(x, y))

◮ This measure satisfies the following 3 properties:

◮ χS is a positive measure ◮ χS satisfies χS(I=i) =

N

i

  • for all i

◮ χS is a measure of positive type (see next slide)

◮ Relaxations:

Et = min

  • λ(f) : λ ∈ M(I2t) positive measure of positive type,

λ(I=i) = N

i

  • for all 0 ≤ i ≤ 2t
slide-24
SLIDE 24

Moment relaxations

◮ For S ∈ I=N, define the measure χS = R⊆S δR on IN ◮ We can use this measure to compute the energy of S ◮ The energy of S is given by

χS(f) =

  • f(P) dχS(P) =
  • R⊆S

f(R) =

  • {x,y}∈I=2

h(d(x, y))

◮ This measure satisfies the following 3 properties:

◮ χS is a positive measure ◮ χS satisfies χS(I=i) =

N

i

  • for all i

◮ χS is a measure of positive type (see next slide)

◮ Relaxations:

Et = min

  • λ(f) : λ ∈ M(I2t) positive measure of positive type,

λ(I=i) = N

i

  • for all 0 ≤ i ≤ 2t
  • ◮ Et is a min{2t, N}-point bound
slide-25
SLIDE 25

Moment relaxations

◮ For S ∈ I=N, define the measure χS = R⊆S δR on IN ◮ We can use this measure to compute the energy of S ◮ The energy of S is given by

χS(f) =

  • f(P) dχS(P) =
  • R⊆S

f(R) =

  • {x,y}∈I=2

h(d(x, y))

◮ This measure satisfies the following 3 properties:

◮ χS is a positive measure ◮ χS satisfies χS(I=i) =

N

i

  • for all i

◮ χS is a measure of positive type (see next slide)

◮ Relaxations:

Et = min

  • λ(f) : λ ∈ M(I2t) positive measure of positive type,

λ(I=i) = N

i

  • for all 0 ≤ i ≤ 2t
  • ◮ Et is a min{2t, N}-point bound

E1 ≤ E2 ≤ · · · ≤ EN = E

slide-26
SLIDE 26

Measures of positive type [L–Vallentin 2015]

◮ Operator:

At : C(It × It)sym → C(I2t), AtK(S) =

  • J,J′∈It:J∪J′=S

K(J, J′)

slide-27
SLIDE 27

Measures of positive type [L–Vallentin 2015]

◮ Operator:

At : C(It × It)sym → C(I2t), AtK(S) =

  • J,J′∈It:J∪J′=S

K(J, J′)

◮ This is an infinite dimensional version of the adjoint of the

  • pererator y → M(y) that maps a moment sequence to a

moment matrix

slide-28
SLIDE 28

Measures of positive type [L–Vallentin 2015]

◮ Operator:

At : C(It × It)sym → C(I2t), AtK(S) =

  • J,J′∈It:J∪J′=S

K(J, J′)

◮ This is an infinite dimensional version of the adjoint of the

  • pererator y → M(y) that maps a moment sequence to a

moment matrix

◮ Dual operator

A∗

t : M(I2t) → M(It × It)sym

slide-29
SLIDE 29

Measures of positive type [L–Vallentin 2015]

◮ Operator:

At : C(It × It)sym → C(I2t), AtK(S) =

  • J,J′∈It:J∪J′=S

K(J, J′)

◮ This is an infinite dimensional version of the adjoint of the

  • pererator y → M(y) that maps a moment sequence to a

moment matrix

◮ Dual operator

A∗

t : M(I2t) → M(It × It)sym ◮ Cone of positive definite kernels: C(It × It)0

slide-30
SLIDE 30

Measures of positive type [L–Vallentin 2015]

◮ Operator:

At : C(It × It)sym → C(I2t), AtK(S) =

  • J,J′∈It:J∪J′=S

K(J, J′)

◮ This is an infinite dimensional version of the adjoint of the

  • pererator y → M(y) that maps a moment sequence to a

moment matrix

◮ Dual operator

A∗

t : M(I2t) → M(It × It)sym ◮ Cone of positive definite kernels: C(It × It)0 ◮ Dual cone:

M(It×It)0 = {µ ∈ M(It×It)sym : µ(K) ≥ 0 for all K ∈ C(It×It)0}

slide-31
SLIDE 31

Measures of positive type [L–Vallentin 2015]

◮ Operator:

At : C(It × It)sym → C(I2t), AtK(S) =

  • J,J′∈It:J∪J′=S

K(J, J′)

◮ This is an infinite dimensional version of the adjoint of the

  • pererator y → M(y) that maps a moment sequence to a

moment matrix

◮ Dual operator

A∗

t : M(I2t) → M(It × It)sym ◮ Cone of positive definite kernels: C(It × It)0 ◮ Dual cone:

M(It×It)0 = {µ ∈ M(It×It)sym : µ(K) ≥ 0 for all K ∈ C(It×It)0}

◮ A measure λ ∈ M(I2t) is of positive type if

A∗

t λ ∈ M(It × It)0

slide-32
SLIDE 32

The dual hierarchy

slide-33
SLIDE 33

The dual hierarchy

E

slide-34
SLIDE 34

The dual hierarchy

Et E

slide-35
SLIDE 35

The dual hierarchy

Et E∗

t

E Dual maximization problem

slide-36
SLIDE 36

The dual hierarchy

Et E∗

t

E Dual maximization problem Strong duality holds: Et = E∗

t

slide-37
SLIDE 37

The dual hierarchy

Et E∗

t

E Dual maximization problem Strong duality holds: Et = E∗

t ◮ In E∗ t we optimize over kernels K ∈ C(It × It)0:

E∗

t = sup

2t

  • i=0

N i

  • ai : a ∈ R{0,...,2t}, K ∈ C(It × It)0,

ai + AtK(S) ≤ f(S) for S ∈ I=i and i = 0, . . . , 2t

  • ,
slide-38
SLIDE 38

The dual hierarchy

Et E∗

t

E Dual maximization problem Strong duality holds: Et = E∗

t ◮ In E∗ t we optimize over kernels K ∈ C(It × It)0:

E∗

t = sup

2t

  • i=0

N i

  • ai : a ∈ R{0,...,2t}, K ∈ C(It × It)0,

ai + AtK(S) ≤ f(S) for S ∈ I=i and i = 0, . . . , 2t

  • ,

◮ Reduce to finite dimensional variable space:

slide-39
SLIDE 39

The dual hierarchy

Et E∗

t

E Dual maximization problem Strong duality holds: Et = E∗

t ◮ In E∗ t we optimize over kernels K ∈ C(It × It)0:

E∗

t = sup

2t

  • i=0

N i

  • ai : a ∈ R{0,...,2t}, K ∈ C(It × It)0,

ai + AtK(S) ≤ f(S) for S ∈ I=i and i = 0, . . . , 2t

  • ,

◮ Reduce to finite dimensional variable space:

  • 1. Express K in terms of its Fourier coefficients
slide-40
SLIDE 40

The dual hierarchy

Et E∗

t

E Dual maximization problem Strong duality holds: Et = E∗

t ◮ In E∗ t we optimize over kernels K ∈ C(It × It)0:

E∗

t = sup

2t

  • i=0

N i

  • ai : a ∈ R{0,...,2t}, K ∈ C(It × It)0,

ai + AtK(S) ≤ f(S) for S ∈ I=i and i = 0, . . . , 2t

  • ,

◮ Reduce to finite dimensional variable space:

  • 1. Express K in terms of its Fourier coefficients
  • 2. Set all but finitely many of these coefficients to 0
slide-41
SLIDE 41

The dual hierarchy

Et E∗

t

E Dual maximization problem Strong duality holds: Et = E∗

t ◮ In E∗ t we optimize over kernels K ∈ C(It × It)0:

E∗

t = sup

2t

  • i=0

N i

  • ai : a ∈ R{0,...,2t}, K ∈ C(It × It)0,

ai + AtK(S) ≤ f(S) for S ∈ I=i and i = 0, . . . , 2t

  • ,

◮ Reduce to finite dimensional variable space:

  • 1. Express K in terms of its Fourier coefficients
  • 2. Set all but finitely many of these coefficients to 0
  • 3. Optimize over the remaining coefficients
slide-42
SLIDE 42

Harmonic analysis on subset spaces

◮ Let Γ be compact group with an action on V

slide-43
SLIDE 43

Harmonic analysis on subset spaces

◮ Let Γ be compact group with an action on V ◮ Example: Γ = O(3) and V = S2 ⊆ R3

slide-44
SLIDE 44

Harmonic analysis on subset spaces

◮ Let Γ be compact group with an action on V ◮ Example: Γ = O(3) and V = S2 ⊆ R3 ◮ Assume the metric is Γ-invariant:

d(γx, γy) = d(x, y) for all x, y ∈ V and γ ∈ Γ

slide-45
SLIDE 45

Harmonic analysis on subset spaces

◮ Let Γ be compact group with an action on V ◮ Example: Γ = O(3) and V = S2 ⊆ R3 ◮ Assume the metric is Γ-invariant:

d(γx, γy) = d(x, y) for all x, y ∈ V and γ ∈ Γ

◮ Then the action extends to an action on It by

γ∅ = ∅ and γ{x1, . . . , xt} = {γx1, . . . , γxt}

slide-46
SLIDE 46

Harmonic analysis on subset spaces

◮ Let Γ be compact group with an action on V ◮ Example: Γ = O(3) and V = S2 ⊆ R3 ◮ Assume the metric is Γ-invariant:

d(γx, γy) = d(x, y) for all x, y ∈ V and γ ∈ Γ

◮ Then the action extends to an action on It by

γ∅ = ∅ and γ{x1, . . . , xt} = {γx1, . . . , γxt}

◮ By an “averaging argument” we may assume

K ∈ C(It × It)0 to be Γ-invariant: K(γJ, γJ′) = K(J, J′) for all γ ∈ Γ and J, J′ ∈ It

slide-47
SLIDE 47

Harmonic analysis on subset spaces

◮ Fourier inversion formula:

K(J, J′) =

  • π∈ˆ

Γ mπ

  • i,j=1

ˆ K(π)i,jZπ(J, J′)i,j

slide-48
SLIDE 48

Harmonic analysis on subset spaces

◮ Fourier inversion formula:

K(J, J′) =

  • π∈ˆ

Γ mπ

  • i,j=1

ˆ K(π)i,jZπ(J, J′)i,j

◮ The Fourier coefficients ˆ

K(π) are psd matrices

slide-49
SLIDE 49

Harmonic analysis on subset spaces

◮ Fourier inversion formula:

K(J, J′) =

  • π∈ˆ

Γ mπ

  • i,j=1

ˆ K(π)i,jZπ(J, J′)i,j

◮ The Fourier coefficients ˆ

K(π) are psd matrices

◮ The Zπ(·, ·) are matrix functions that depend on Γ and It

slide-50
SLIDE 50

Harmonic analysis on subset spaces

◮ Fourier inversion formula:

K(J, J′) =

  • π∈ˆ

Γ mπ

  • i,j=1

ˆ K(π)i,jZπ(J, J′)i,j

◮ The Fourier coefficients ˆ

K(π) are psd matrices

◮ The Zπ(·, ·) are matrix functions that depend on Γ and It ◮ The action of Γ on It gives a linear action of Γ on C(It) by

γf(S) = f(γ−1S)

slide-51
SLIDE 51

Harmonic analysis on subset spaces

◮ Fourier inversion formula:

K(J, J′) =

  • π∈ˆ

Γ mπ

  • i,j=1

ˆ K(π)i,jZπ(J, J′)i,j

◮ The Fourier coefficients ˆ

K(π) are psd matrices

◮ The Zπ(·, ·) are matrix functions that depend on Γ and It ◮ The action of Γ on It gives a linear action of Γ on C(It) by

γf(S) = f(γ−1S)

◮ To construct the Zπ(·, ·) we need to decompose C(It) as a

direct sum of irreducible Γ-invariant subspaces

slide-52
SLIDE 52

Harmonic analysis on subset spaces

◮ Fourier inversion formula:

K(J, J′) =

  • π∈ˆ

Γ mπ

  • i,j=1

ˆ K(π)i,jZπ(J, J′)i,j

◮ The Fourier coefficients ˆ

K(π) are psd matrices

◮ The Zπ(·, ·) are matrix functions that depend on Γ and It ◮ The action of Γ on It gives a linear action of Γ on C(It) by

γf(S) = f(γ−1S)

◮ To construct the Zπ(·, ·) we need to decompose C(It) as a

direct sum of irreducible Γ-invariant subspaces

◮ We give procedure to do this using symmetric tensor powers

slide-53
SLIDE 53

Harmonic analysis on subset spaces

◮ Fourier inversion formula:

K(J, J′) =

  • π∈ˆ

Γ mπ

  • i,j=1

ˆ K(π)i,jZπ(J, J′)i,j

◮ The Fourier coefficients ˆ

K(π) are psd matrices

◮ The Zπ(·, ·) are matrix functions that depend on Γ and It ◮ The action of Γ on It gives a linear action of Γ on C(It) by

γf(S) = f(γ−1S)

◮ To construct the Zπ(·, ·) we need to decompose C(It) as a

direct sum of irreducible Γ-invariant subspaces

◮ We give procedure to do this using symmetric tensor powers ◮ We do this explicitly for V = S2, Γ = O(3), and t = 2

(by using Clebsch–Gordan coefficients)

slide-54
SLIDE 54

Harmonic analysis on subset spaces

◮ Fourier inversion formula:

K(J, J′) =

  • π∈ˆ

Γ mπ

  • i,j=1

ˆ K(π)i,jZπ(J, J′)i,j

◮ The Fourier coefficients ˆ

K(π) are psd matrices

◮ The Zπ(·, ·) are matrix functions that depend on Γ and It ◮ The action of Γ on It gives a linear action of Γ on C(It) by

γf(S) = f(γ−1S)

◮ To construct the Zπ(·, ·) we need to decompose C(It) as a

direct sum of irreducible Γ-invariant subspaces

◮ We give procedure to do this using symmetric tensor powers ◮ We do this explicitly for V = S2, Γ = O(3), and t = 2

(by using Clebsch–Gordan coefficients)

◮ In this way we lower bound E∗ 2 by problems with finitely many

variables and infinitely many constraints

slide-55
SLIDE 55

Invariant theory (for V = S2)

◮ These constraints are of the form

p(x1, . . . , xi) ≥ 0 for {x1, . . . , xi} ∈ I=i, where p is a polynomial whose coefficients depend linearly on the entries of the matrix variables

slide-56
SLIDE 56

Invariant theory (for V = S2)

◮ These constraints are of the form

p(x1, . . . , xi) ≥ 0 for {x1, . . . , xi} ∈ I=i, where p is a polynomial whose coefficients depend linearly on the entries of the matrix variables

◮ These polynomials satisfy

p(γx1, . . . , γxi) = p(x1, . . . , xi) for x1, . . . , xi ∈ S2 and γ ∈ O(3)

slide-57
SLIDE 57

Invariant theory (for V = S2)

◮ These constraints are of the form

p(x1, . . . , xi) ≥ 0 for {x1, . . . , xi} ∈ I=i, where p is a polynomial whose coefficients depend linearly on the entries of the matrix variables

◮ These polynomials satisfy

p(γx1, . . . , γxi) = p(x1, . . . , xi) for x1, . . . , xi ∈ S2 and γ ∈ O(3)

◮ By a theorem of invariant theory we can write p as a

polynomial in the inner products: p(x1, . . . , xi) = q(x1 · x1, x1 · x2, . . . , xi · xi)

slide-58
SLIDE 58

Invariant theory (for V = S2)

◮ These constraints are of the form

p(x1, . . . , xi) ≥ 0 for {x1, . . . , xi} ∈ I=i, where p is a polynomial whose coefficients depend linearly on the entries of the matrix variables

◮ These polynomials satisfy

p(γx1, . . . , γxi) = p(x1, . . . , xi) for x1, . . . , xi ∈ S2 and γ ∈ O(3)

◮ By a theorem of invariant theory we can write p as a

polynomial in the inner products: p(x1, . . . , xi) = q(x1 · x1, x1 · x2, . . . , xi · xi)

◮ Now we have constraints of the form

q(u1, . . . , ul) ≥ 0 for (u1, . . . , ul) ∈ some semialgebraic set

slide-59
SLIDE 59

Invariant theory

p(x1, . . . , xi) = q(x1 · x1, x1 · x2, . . . , xi · xi), deg(p) = 2d

◮ The theorem that gives the existence of q is nonconstructive

slide-60
SLIDE 60

Invariant theory

p(x1, . . . , xi) = q(x1 · x1, x1 · x2, . . . , xi · xi), deg(p) = 2d

◮ The theorem that gives the existence of q is nonconstructive ◮ Find q by solving linear system Ax = b

Rows indexed by monomials in 3i vars of degree ≤ 2d Columns indexed by monomials in i+1

2

  • vars of degree ≤ d
slide-61
SLIDE 61

Invariant theory

p(x1, . . . , xi) = q(x1 · x1, x1 · x2, . . . , xi · xi), deg(p) = 2d

◮ The theorem that gives the existence of q is nonconstructive ◮ Find q by solving linear system Ax = b

Rows indexed by monomials in 3i vars of degree ≤ 2d Columns indexed by monomials in i+1

2

  • vars of degree ≤ d

◮ For i = 4, d = 6 we get over a million rows

slide-62
SLIDE 62

Invariant theory

p(x1, . . . , xi) = q(x1 · x1, x1 · x2, . . . , xi · xi), deg(p) = 2d

◮ The theorem that gives the existence of q is nonconstructive ◮ Find q by solving linear system Ax = b

Rows indexed by monomials in 3i vars of degree ≤ 2d Columns indexed by monomials in i+1

2

  • vars of degree ≤ d

◮ For i = 4, d = 6 we get over a million rows ◮ Use custom pivoting, sparse, high precision, Cholesky

factorization algorithm

slide-63
SLIDE 63

Invariant theory

p(x1, . . . , xi) = q(x1 · x1, x1 · x2, . . . , xi · xi), deg(p) = 2d

◮ The theorem that gives the existence of q is nonconstructive ◮ Find q by solving linear system Ax = b

Rows indexed by monomials in 3i vars of degree ≤ 2d Columns indexed by monomials in i+1

2

  • vars of degree ≤ d

◮ For i = 4, d = 6 we get over a million rows ◮ Use custom pivoting, sparse, high precision, Cholesky

factorization algorithm

◮ Computing the q polynomials takes several days, but only

needs to be done once for given d

slide-64
SLIDE 64

Sums of squares characterizations

◮ Putinar: Every positive polynomial on a compact set

S = {x ∈ Rn : g1(x) ≥ 0, . . . , gm(x) ≥ 0}, where {g1, . . . , gm} has the Archimedean property, is of the form f(x) =

m

  • i=0

gi(x)si(x), where g0 = 1 and s0, . . . , sm are SOS

slide-65
SLIDE 65

Sums of squares characterizations

◮ Putinar: Every positive polynomial on a compact set

S = {x ∈ Rn : g1(x) ≥ 0, . . . , gm(x) ≥ 0}, where {g1, . . . , gm} has the Archimedean property, is of the form f(x) =

m

  • i=0

gi(x)si(x), where g0 = 1 and s0, . . . , sm are SOS

◮ The SOS polynomials si can be modeled using psd matrices

slide-66
SLIDE 66

Sums of squares characterizations

◮ Putinar: Every positive polynomial on a compact set

S = {x ∈ Rn : g1(x) ≥ 0, . . . , gm(x) ≥ 0}, where {g1, . . . , gm} has the Archimedean property, is of the form f(x) =

m

  • i=0

gi(x)si(x), where g0 = 1 and s0, . . . , sm are SOS

◮ The SOS polynomials si can be modeled using psd matrices ◮ We use this to go from infinitely many linear constraints to

finitely many semidefinite constraints

slide-67
SLIDE 67

Sums of squares characterizations

◮ Putinar: Every positive polynomial on a compact set

S = {x ∈ Rn : g1(x) ≥ 0, . . . , gm(x) ≥ 0}, where {g1, . . . , gm} has the Archimedean property, is of the form f(x) =

m

  • i=0

gi(x)si(x), where g0 = 1 and s0, . . . , sm are SOS

◮ The SOS polynomials si can be modeled using psd matrices ◮ We use this to go from infinitely many linear constraints to

finitely many semidefinite constraints

◮ In energy minimization the particles are interchangeable

slide-68
SLIDE 68

Sums of squares characterizations

◮ Putinar: Every positive polynomial on a compact set

S = {x ∈ Rn : g1(x) ≥ 0, . . . , gm(x) ≥ 0}, where {g1, . . . , gm} has the Archimedean property, is of the form f(x) =

m

  • i=0

gi(x)si(x), where g0 = 1 and s0, . . . , sm are SOS

◮ The SOS polynomials si can be modeled using psd matrices ◮ We use this to go from infinitely many linear constraints to

finitely many semidefinite constraints

◮ In energy minimization the particles are interchangeable ◮ This means

p(xσ(1), . . . , xσ(i)) = p(x1, . . . , xi) for all σ ∈ Si

slide-69
SLIDE 69

Sums of squares characterizations

◮ Putinar: Every positive polynomial on a compact set

S = {x ∈ Rn : g1(x) ≥ 0, . . . , gm(x) ≥ 0}, where {g1, . . . , gm} has the Archimedean property, is of the form f(x) =

m

  • i=0

gi(x)si(x), where g0 = 1 and s0, . . . , sm are SOS

◮ The SOS polynomials si can be modeled using psd matrices ◮ We use this to go from infinitely many linear constraints to

finitely many semidefinite constraints

◮ In energy minimization the particles are interchangeable ◮ This means

p(xσ(1), . . . , xσ(i)) = p(x1, . . . , xi) for all σ ∈ Si

◮ Additional symmetries in the q(u1, . . . , ul) polynomials

slide-70
SLIDE 70

Sums of squares characterizations

◮ Symmetrization of Putinar’s theorem to exploit the symmetry

in the particles

slide-71
SLIDE 71

Sums of squares characterizations

◮ Symmetrization of Putinar’s theorem to exploit the symmetry

in the particles

◮ Assume the set {g0, . . . , gm} is Γ-invariant

slide-72
SLIDE 72

Sums of squares characterizations

◮ Symmetrization of Putinar’s theorem to exploit the symmetry

in the particles

◮ Assume the set {g0, . . . , gm} is Γ-invariant ◮ Denote by Γgi the stabilizer subgroup of Γ with respect to gi

slide-73
SLIDE 73

Sums of squares characterizations

◮ Symmetrization of Putinar’s theorem to exploit the symmetry

in the particles

◮ Assume the set {g0, . . . , gm} is Γ-invariant ◮ Denote by Γgi the stabilizer subgroup of Γ with respect to gi

A Γ-invariant polynomial that has a Putinar representation can be written as p = m

i=0 gisi, where si is a Γgi-invariant

sum of squares polynomial

slide-74
SLIDE 74

Sums of squares characterizations

◮ Symmetrization of Putinar’s theorem to exploit the symmetry

in the particles

◮ Assume the set {g0, . . . , gm} is Γ-invariant ◮ Denote by Γgi the stabilizer subgroup of Γ with respect to gi

A Γ-invariant polynomial that has a Putinar representation can be written as p = m

i=0 gisi, where si is a Γgi-invariant

sum of squares polynomial

◮ We can represent the Γgi-invariant sum of squares

polynomials si using block diagonalized positive semidefinite matrices [Gatermann–Parillo 2004]

slide-75
SLIDE 75

Sums of squares characterizations

◮ Symmetrization of Putinar’s theorem to exploit the symmetry

in the particles

◮ Assume the set {g0, . . . , gm} is Γ-invariant ◮ Denote by Γgi the stabilizer subgroup of Γ with respect to gi

A Γ-invariant polynomial that has a Putinar representation can be written as p = m

i=0 gisi, where si is a Γgi-invariant

sum of squares polynomial

◮ We can represent the Γgi-invariant sum of squares

polynomials si using block diagonalized positive semidefinite matrices [Gatermann–Parillo 2004]

◮ For energy minimization on the sphere this yields large

reductions in solver time (Ex. 150 hours → 7 hours)

slide-76
SLIDE 76

Computations

◮ Mow we have an SDP given as high precision numbers whose

  • ptimal value lower bounds the ground state energy
slide-77
SLIDE 77

Computations

◮ Mow we have an SDP given as high precision numbers whose

  • ptimal value lower bounds the ground state energy

◮ Want to solve with high precision SDP solver

slide-78
SLIDE 78

Computations

◮ Mow we have an SDP given as high precision numbers whose

  • ptimal value lower bounds the ground state energy

◮ Want to solve with high precision SDP solver ◮ Problem 1: Free variables in the SDP → Dual SDP not

strictly feasible → Cannot solve with high precision solver

slide-79
SLIDE 79

Computations

◮ Mow we have an SDP given as high precision numbers whose

  • ptimal value lower bounds the ground state energy

◮ Want to solve with high precision SDP solver ◮ Problem 1: Free variables in the SDP → Dual SDP not

strictly feasible → Cannot solve with high precision solver

◮ Bound free variables with big M constraints

slide-80
SLIDE 80

Computations

◮ Mow we have an SDP given as high precision numbers whose

  • ptimal value lower bounds the ground state energy

◮ Want to solve with high precision SDP solver ◮ Problem 1: Free variables in the SDP → Dual SDP not

strictly feasible → Cannot solve with high precision solver

◮ Bound free variables with big M constraints ◮ Problem 2: The additional symmetry exploitation leads to

hard to predict linear dependencies in the constraints

slide-81
SLIDE 81

Computations

◮ Mow we have an SDP given as high precision numbers whose

  • ptimal value lower bounds the ground state energy

◮ Want to solve with high precision SDP solver ◮ Problem 1: Free variables in the SDP → Dual SDP not

strictly feasible → Cannot solve with high precision solver

◮ Bound free variables with big M constraints ◮ Problem 2: The additional symmetry exploitation leads to

hard to predict linear dependencies in the constraints

◮ Use QR factorization of the constraint matrix to remove these

slide-82
SLIDE 82

Computations

◮ In the Thomson problem we take

V = S2, d(x, y) = x − y2, and h(w) = 1 w

slide-83
SLIDE 83

Computations

◮ In the Thomson problem we take

V = S2, d(x, y) = x − y2, and h(w) = 1 w

◮ E∗ 1 is sharp for 2, 3, 4, 6, and 12 particles (Yudin’s LP bound)

slide-84
SLIDE 84

Computations

◮ In the Thomson problem we take

V = S2, d(x, y) = x − y2, and h(w) = 1 w

◮ E∗ 1 is sharp for 2, 3, 4, 6, and 12 particles (Yudin’s LP bound) ◮ The triangular bipiramid is optimal for N = 5 (Schwartz 2010)

slide-85
SLIDE 85

Computations

◮ In the Thomson problem we take

V = S2, d(x, y) = x − y2, and h(w) = 1 w

◮ E∗ 1 is sharp for 2, 3, 4, 6, and 12 particles (Yudin’s LP bound) ◮ The triangular bipiramid is optimal for N = 5 (Schwartz 2010) ◮ High precision SDP solver gives the first 28 decimal digits of a

lower bound on E2

slide-86
SLIDE 86

Computations

◮ In the Thomson problem we take

V = S2, d(x, y) = x − y2, and h(w) = 1 w

◮ E∗ 1 is sharp for 2, 3, 4, 6, and 12 particles (Yudin’s LP bound) ◮ The triangular bipiramid is optimal for N = 5 (Schwartz 2010) ◮ High precision SDP solver gives the first 28 decimal digits of a

lower bound on E2

◮ These all agree with the energy of the triangular bipiramid

slide-87
SLIDE 87

Computations

◮ We should be able to use this to construct an optimality

certificate for the N = 5 case of the Thomson problem, but need to replace linear algebra by Gr¨

  • bner bases
slide-88
SLIDE 88

Computations

◮ We should be able to use this to construct an optimality

certificate for the N = 5 case of the Thomson problem, but need to replace linear algebra by Gr¨

  • bner bases

◮ The system of 5 particles on S2 admits a phase transition

slide-89
SLIDE 89

Computations

◮ We should be able to use this to construct an optimality

certificate for the N = 5 case of the Thomson problem, but need to replace linear algebra by Gr¨

  • bner bases

◮ The system of 5 particles on S2 admits a phase transition ◮ Using SDP solver we see E2 is also (numerically) sharp for

many other pair potentials

slide-90
SLIDE 90

Computations

◮ We should be able to use this to construct an optimality

certificate for the N = 5 case of the Thomson problem, but need to replace linear algebra by Gr¨

  • bner bases

◮ The system of 5 particles on S2 admits a phase transition ◮ Using SDP solver we see E2 is also (numerically) sharp for

many other pair potentials

◮ Conjecture: E2 is universally sharp for 5 particles on S2

slide-91
SLIDE 91

Computations

◮ We should be able to use this to construct an optimality

certificate for the N = 5 case of the Thomson problem, but need to replace linear algebra by Gr¨

  • bner bases

◮ The system of 5 particles on S2 admits a phase transition ◮ Using SDP solver we see E2 is also (numerically) sharp for

many other pair potentials

◮ Conjecture: E2 is universally sharp for 5 particles on S2 ◮ This is the first time a four 4-bound has been computed for a

continuous problem

slide-92
SLIDE 92

Computations

◮ We should be able to use this to construct an optimality

certificate for the N = 5 case of the Thomson problem, but need to replace linear algebra by Gr¨

  • bner bases

◮ The system of 5 particles on S2 admits a phase transition ◮ Using SDP solver we see E2 is also (numerically) sharp for

many other pair potentials

◮ Conjecture: E2 is universally sharp for 5 particles on S2 ◮ This is the first time a four 4-bound has been computed for a

continuous problem

◮ Future work: apply these techniques to packing problems

slide-93
SLIDE 93

Thank you!

  • D. de Laat, Moment methods in energy minimization: New bounds for

Riesz minimal energy problems, arXiv:1610.04905.