Energy minimization via moment hierarchies David de Laat (TU Delft) - - PowerPoint PPT Presentation

energy minimization via moment hierarchies
SMART_READER_LITE
LIVE PREVIEW

Energy minimization via moment hierarchies David de Laat (TU Delft) - - PowerPoint PPT Presentation

Energy minimization via moment hierarchies David de Laat (TU Delft) ESI Workshop on Optimal Point Configurations and Applications 16 October 2014 Energy minimization What is the minimal potential energy E when we put N particles with pair


slide-1
SLIDE 1

Energy minimization via moment hierarchies

David de Laat (TU Delft)

ESI Workshop on Optimal Point Configurations and Applications 16 October 2014

slide-2
SLIDE 2

Energy minimization

◮ What is the minimal potential energy E when we put N

particles with pair potential h in a container V ?

slide-3
SLIDE 3

Energy minimization

◮ What is the minimal potential energy E when we put N

particles with pair potential h in a container V ?

◮ Example: For the Thomson problem we take

V = S2 and h({x, y}) = 1 x − y

slide-4
SLIDE 4

Energy minimization

◮ What is the minimal potential energy E when we put N

particles with pair potential h in a container V ?

◮ Example: For the Thomson problem we take

V = S2 and h({x, y}) = 1 x − y

◮ As an optimization problem:

E = min

S∈(V

N)

  • P∈(S

2)

h(P)

slide-5
SLIDE 5

Approach

◮ Configurations provide upper bounds on the optimal energy E

slide-6
SLIDE 6

Approach

◮ Configurations provide upper bounds on the optimal energy E ◮ To prove a configuration is good (or optimal) we need good

lower bounds for E

slide-7
SLIDE 7

Approach

◮ Configurations provide upper bounds on the optimal energy E ◮ To prove a configuration is good (or optimal) we need good

lower bounds for E Some systematic approaches for obtaining bounds:

◮ Linear programming bounds using the pair correlation function

[Delsarte 1973, Delsarte-Goethals-Seidel 1977, Yudin 1992]

slide-8
SLIDE 8

Approach

◮ Configurations provide upper bounds on the optimal energy E ◮ To prove a configuration is good (or optimal) we need good

lower bounds for E Some systematic approaches for obtaining bounds:

◮ Linear programming bounds using the pair correlation function

[Delsarte 1973, Delsarte-Goethals-Seidel 1977, Yudin 1992]

◮ 3-point bounds using 3-point correlation functions and

constraints arising from the stabilizer subgroup of 1 point [Schrijver 2005, Bachoc-Vallentin 2008, Cohn-Woo 2012]

slide-9
SLIDE 9

Approach

◮ Configurations provide upper bounds on the optimal energy E ◮ To prove a configuration is good (or optimal) we need good

lower bounds for E Some systematic approaches for obtaining bounds:

◮ Linear programming bounds using the pair correlation function

[Delsarte 1973, Delsarte-Goethals-Seidel 1977, Yudin 1992]

◮ 3-point bounds using 3-point correlation functions and

constraints arising from the stabilizer subgroup of 1 point [Schrijver 2005, Bachoc-Vallentin 2008, Cohn-Woo 2012]

◮ k-point bounds using stabilizer subgroup of k − 2 points

[Musin 2007]

slide-10
SLIDE 10

Approach

◮ Configurations provide upper bounds on the optimal energy E ◮ To prove a configuration is good (or optimal) we need good

lower bounds for E Some systematic approaches for obtaining bounds:

◮ Linear programming bounds using the pair correlation function

[Delsarte 1973, Delsarte-Goethals-Seidel 1977, Yudin 1992]

◮ 3-point bounds using 3-point correlation functions and

constraints arising from the stabilizer subgroup of 1 point [Schrijver 2005, Bachoc-Vallentin 2008, Cohn-Woo 2012]

◮ k-point bounds using stabilizer subgroup of k − 2 points

[Musin 2007]

◮ Hierarchy for packing problems [L.-Vallentin 2014]

slide-11
SLIDE 11

This talk

◮ Hierarchy obtained by generalizing Lasserre’s hierarchy from

combinatorial optimization to the continuous setting

slide-12
SLIDE 12

This talk

◮ Hierarchy obtained by generalizing Lasserre’s hierarchy from

combinatorial optimization to the continuous setting

◮ Finite convergence to the optimal energy

slide-13
SLIDE 13

This talk

◮ Hierarchy obtained by generalizing Lasserre’s hierarchy from

combinatorial optimization to the continuous setting

◮ Finite convergence to the optimal energy ◮ A duality theory

slide-14
SLIDE 14

This talk

◮ Hierarchy obtained by generalizing Lasserre’s hierarchy from

combinatorial optimization to the continuous setting

◮ Finite convergence to the optimal energy ◮ A duality theory ◮ Reduction to a converging sequence of semidefinite programs

slide-15
SLIDE 15

This talk

◮ Hierarchy obtained by generalizing Lasserre’s hierarchy from

combinatorial optimization to the continuous setting

◮ Finite convergence to the optimal energy ◮ A duality theory ◮ Reduction to a converging sequence of semidefinite programs ◮ Towards computations using several types of symmetry

reduction

slide-16
SLIDE 16

Approach

E

slide-17
SLIDE 17

Approach

E

Difficult minimization problem

slide-18
SLIDE 18

Approach

Et E

Difficult minimization problem

slide-19
SLIDE 19

Approach

Et E

Difficult minimization problem Relaxation to a conic program: Infinite dimensional minimization problem

slide-20
SLIDE 20

Approach

Et E E∗

t

Difficult minimization problem Relaxation to a conic program: Infinite dimensional minimization problem

slide-21
SLIDE 21

Approach

Et E E∗

t

Difficult minimization problem Relaxation to a conic program: Conic dual: Infinite dimensional minimization problem Infinite dimensional maximization problem

slide-22
SLIDE 22

Approach

Et E E∗

t

E∗

t,d

Difficult minimization problem Relaxation to a conic program: Conic dual: Infinite dimensional minimization problem Infinite dimensional maximization problem

slide-23
SLIDE 23

Approach

Et E E∗

t

E∗

t,d

Difficult minimization problem Relaxation to a conic program: Conic dual: Semi-infinite semidefinite program Infinite dimensional minimization problem Infinite dimensional maximization problem

slide-24
SLIDE 24

The minimization problem

◮ I=t (It) is the set of subsets of V which

◮ have cardinality t (≤ t) ◮ contain no points which are too close

slide-25
SLIDE 25

The minimization problem

◮ I=t (It) is the set of subsets of V which

◮ have cardinality t (≤ t) ◮ contain no points which are too close

◮ Assuming h({x, y}) → ∞ when x and y converge, we have

E = min

S∈I=N

  • P∈(S

2)

h(P)

slide-26
SLIDE 26

The minimization problem

◮ I=t (It) is the set of subsets of V which

◮ have cardinality t (≤ t) ◮ contain no points which are too close

◮ Assuming h({x, y}) → ∞ when x and y converge, we have

E = min

S∈I=N

  • P∈(S

2)

h(P)

◮ We will also assume that V is compact and h continuous

slide-27
SLIDE 27

The minimization problem

◮ I=t (It) is the set of subsets of V which

◮ have cardinality t (≤ t) ◮ contain no points which are too close

◮ Assuming h({x, y}) → ∞ when x and y converge, we have

E = min

S∈I=N

  • P∈(S

2)

h(P)

◮ We will also assume that V is compact and h continuous ◮ I=t gets its topology as a subset of a quotient of V t

slide-28
SLIDE 28

Moment hierarchy of relaxations

◮ In the relaxation Et we minimize over measures λ on the

space Is, where s = min{2t, N}

slide-29
SLIDE 29

Moment hierarchy of relaxations

◮ In the relaxation Et we minimize over measures λ on the

space Is, where s = min{2t, N} Lemma When t = N, the feasible measures λ are (generalized) convex combinations of measures χS =

  • R⊆S

δR where S ∈ I=N

slide-30
SLIDE 30

Moment hierarchy of relaxations

◮ In the relaxation Et we minimize over measures λ on the

space Is, where s = min{2t, N} Lemma When t = N, the feasible measures λ are (generalized) convex combinations of measures χS =

  • R⊆S

δR where S ∈ I=N

◮ Objective function: λ(h) =

  • I=N h(S) dλ(S)
slide-31
SLIDE 31

Moment hierarchy of relaxations

◮ In the relaxation Et we minimize over measures λ on the

space Is, where s = min{2t, N} Lemma When t = N, the feasible measures λ are (generalized) convex combinations of measures χS =

  • R⊆S

δR where S ∈ I=N

◮ Objective function: λ(h) =

  • I=N h(S) dλ(S)

◮ Moment constraints: A∗ t λ ∈ M(It × It)0

slide-32
SLIDE 32

Moment hierarchy of relaxations

◮ In the relaxation Et we minimize over measures λ on the

space Is, where s = min{2t, N} Lemma When t = N, the feasible measures λ are (generalized) convex combinations of measures χS =

  • R⊆S

δR where S ∈ I=N

◮ Objective function: λ(h) =

  • I=N h(S) dλ(S)

◮ Moment constraints: A∗ t λ ∈ M(It × It)0 ◮ Here A∗ t is an operator M(Is) → M(It × It)

slide-33
SLIDE 33

Moment hierarchy of relaxations

◮ In the relaxation Et we minimize over measures λ on the

space Is, where s = min{2t, N} Lemma When t = N, the feasible measures λ are (generalized) convex combinations of measures χS =

  • R⊆S

δR where S ∈ I=N

◮ Objective function: λ(h) =

  • I=N h(S) dλ(S)

◮ Moment constraints: A∗ t λ ∈ M(It × It)0 ◮ Here A∗ t is an operator M(Is) → M(It × It) ◮ M(It × It)0 is the cone dual to the cone C(It × It)0 of

positive kernels

slide-34
SLIDE 34

Moment hierarchy of relaxations

◮ In the relaxation Et we minimize over measures λ on the

space Is, where s = min{2t, N} Lemma When t = N, the feasible measures λ are (generalized) convex combinations of measures χS =

  • R⊆S

δR where S ∈ I=N

◮ Objective function: λ(h) =

  • I=N h(S) dλ(S)

◮ Moment constraints: A∗ t λ ∈ M(It × It)0 ◮ Here A∗ t is an operator M(Is) → M(It × It) ◮ M(It × It)0 is the cone dual to the cone C(It × It)0 of

positive kernels: µ(K) ≥ 0 for all K 0

slide-35
SLIDE 35

Moment hierarchy of relaxations

◮ In the relaxation Et we minimize over measures λ on the

space Is, where s = min{2t, N} Lemma When t = N, the feasible measures λ are (generalized) convex combinations of measures χS =

  • R⊆S

δR where S ∈ I=N

◮ Objective function: λ(h) =

  • I=N h(S) dλ(S)

◮ Moment constraints: A∗ t λ ∈ M(It × It)0 ◮ Here A∗ t is an operator M(Is) → M(It × It) ◮ M(It × It)0 is the cone dual to the cone C(It × It)0 of

positive kernels: µ(K) ≥ 0 for all K 0

◮ We have χS(h) = P∈(S

2) h(P)

slide-36
SLIDE 36

Moment hierarchy of relaxations

◮ In the relaxation Et we minimize over measures λ on the

space Is, where s = min{2t, N} Lemma When t = N, the feasible measures λ are (generalized) convex combinations of measures χS =

  • R⊆S

δR where S ∈ I=N

◮ Objective function: λ(h) =

  • I=N h(S) dλ(S)

◮ Moment constraints: A∗ t λ ∈ M(It × It)0 ◮ Here A∗ t is an operator M(Is) → M(It × It) ◮ M(It × It)0 is the cone dual to the cone C(It × It)0 of

positive kernels: µ(K) ≥ 0 for all K 0

◮ We have χS(h) = P∈(S

2) h(P)

Theorem (Finite convergence) We have E1 ≤ · · · ≤ EN = E

slide-37
SLIDE 37

Dual hierarchy

◮ Et is a minimization problem, so we need an optimal solution

to find a lower bound

slide-38
SLIDE 38

Dual hierarchy

◮ Et is a minimization problem, so we need an optimal solution

to find a lower bound

◮ The conic dual E∗ t is a maximization problem where any

feasible solution provides an upper bound

slide-39
SLIDE 39

Dual hierarchy

◮ Et is a minimization problem, so we need an optimal solution

to find a lower bound

◮ The conic dual E∗ t is a maximization problem where any

feasible solution provides an upper bound

◮ In E∗ t optimization is over scalars ai ∈ R and positive definite

kernels K ∈ C(It × It)0

slide-40
SLIDE 40

Dual hierarchy

◮ Et is a minimization problem, so we need an optimal solution

to find a lower bound

◮ The conic dual E∗ t is a maximization problem where any

feasible solution provides an upper bound

◮ In E∗ t optimization is over scalars ai ∈ R and positive definite

kernels K ∈ C(It × It)0

◮ The dual program:

E∗

t = sup

  • s
  • i=0

N

i

  • ai : a0, . . . , as ∈ R, K ∈ C(It × It)0,

ai + AtK ≤ h on I=i for i = 0, . . . , s

slide-41
SLIDE 41

Dual hierarchy

◮ Et is a minimization problem, so we need an optimal solution

to find a lower bound

◮ The conic dual E∗ t is a maximization problem where any

feasible solution provides an upper bound

◮ In E∗ t optimization is over scalars ai ∈ R and positive definite

kernels K ∈ C(It × It)0

◮ The dual program:

E∗

t = sup

  • s
  • i=0

N

i

  • ai : a0, . . . , as ∈ R, K ∈ C(It × It)0,

ai + AtK ≤ h on I=i for i = 0, . . . , s

  • ◮ Here At is the linear operator C(It × It) → C(It) given by

AtK(S) =

J,J′∈It:J∪J′=S K(J, J′)

slide-42
SLIDE 42

Dual hierarchy

◮ Et is a minimization problem, so we need an optimal solution

to find a lower bound

◮ The conic dual E∗ t is a maximization problem where any

feasible solution provides an upper bound

◮ In E∗ t optimization is over scalars ai ∈ R and positive definite

kernels K ∈ C(It × It)0

◮ The dual program:

E∗

t = sup

  • s
  • i=0

N

i

  • ai : a0, . . . , as ∈ R, K ∈ C(It × It)0,

ai + AtK ≤ h on I=i for i = 0, . . . , s

  • ◮ Here At is the linear operator C(It × It) → C(It) given by

AtK(S) =

J,J′∈It:J∪J′=S K(J, J′)

Theorem Strong duality holds: Et = E∗

t for each t

slide-43
SLIDE 43

Closing the gaps

Et E E∗

t

E∗

t,d

Difficult minimization problem Relaxation to a conic program: Conic dual: Semi-infinite semidefinite program Infinite dimensional minimization problem Infinite dimensional maximization problem

slide-44
SLIDE 44

Finite dimensional approximations to E∗

t

◮ Define E∗ t,d by replacing the cone C(It × It)0 in E∗ t by a

finite dimensional inner approximating cone Cd

slide-45
SLIDE 45

Finite dimensional approximations to E∗

t

◮ Define E∗ t,d by replacing the cone C(It × It)0 in E∗ t by a

finite dimensional inner approximating cone Cd

◮ Let e1, e2, . . . be a dense sequence in C(It) and define

Cd =

  • d
  • i,j=1

Fi,jei ⊗ ej : F ∈ Rd×d positive semidefinite

slide-46
SLIDE 46

Finite dimensional approximations to E∗

t

◮ Define E∗ t,d by replacing the cone C(It × It)0 in E∗ t by a

finite dimensional inner approximating cone Cd

◮ Let e1, e2, . . . be a dense sequence in C(It) and define

Cd =

  • d
  • i,j=1

Fi,jei ⊗ ej : F ∈ Rd×d positive semidefinite

  • Lemma

Suppose X is a compact metric space. Then the extreme rays of the cone C(X ×X)0 are precisely the kernels f ⊗f with f ∈ C(X)

slide-47
SLIDE 47

Finite dimensional approximations to E∗

t

◮ Define E∗ t,d by replacing the cone C(It × It)0 in E∗ t by a

finite dimensional inner approximating cone Cd

◮ Let e1, e2, . . . be a dense sequence in C(It) and define

Cd =

  • d
  • i,j=1

Fi,jei ⊗ ej : F ∈ Rd×d positive semidefinite

  • Lemma

Suppose X is a compact metric space. Then the extreme rays of the cone C(X ×X)0 are precisely the kernels f ⊗f with f ∈ C(X)

◮ This implies ∪∞ d=0Cd is uniformly dense in C(It × It)0

slide-48
SLIDE 48

Finite dimensional approximations to E∗

t

◮ Define E∗ t,d by replacing the cone C(It × It)0 in E∗ t by a

finite dimensional inner approximating cone Cd

◮ Let e1, e2, . . . be a dense sequence in C(It) and define

Cd =

  • d
  • i,j=1

Fi,jei ⊗ ej : F ∈ Rd×d positive semidefinite

  • Lemma

Suppose X is a compact metric space. Then the extreme rays of the cone C(X ×X)0 are precisely the kernels f ⊗f with f ∈ C(X)

◮ This implies ∪∞ d=0Cd is uniformly dense in C(It × It)0

Theorem If V is a compact metric space, then E∗

t,d → E∗ t as d → ∞ for all t

slide-49
SLIDE 49

Block diagonalization

◮ For computations use the symmetry of V and h, expressed by

the action of a group Γ, and Bochner’s theorem to block diagonalize the matrix F

slide-50
SLIDE 50

Block diagonalization

◮ For computations use the symmetry of V and h, expressed by

the action of a group Γ, and Bochner’s theorem to block diagonalize the matrix F

◮ For this we need a symmetry adapted basis of C(It)

slide-51
SLIDE 51

Block diagonalization

◮ For computations use the symmetry of V and h, expressed by

the action of a group Γ, and Bochner’s theorem to block diagonalize the matrix F

◮ For this we need a symmetry adapted basis of C(It) ◮ If t = 1 and V = S2, then

C(It) ≃ R ⊕ C(S2)

slide-52
SLIDE 52

Block diagonalization

◮ For computations use the symmetry of V and h, expressed by

the action of a group Γ, and Bochner’s theorem to block diagonalize the matrix F

◮ For this we need a symmetry adapted basis of C(It) ◮ If t = 1 and V = S2, then

C(It) ≃ R ⊕ C(S2) = R ⊕

  • k=0

Hk

slide-53
SLIDE 53

Block diagonalization

◮ For computations use the symmetry of V and h, expressed by

the action of a group Γ, and Bochner’s theorem to block diagonalize the matrix F

◮ For this we need a symmetry adapted basis of C(It) ◮ If t = 1 and V = S2, then

C(It) ≃ R ⊕ C(S2) = R ⊕

  • k=0

Hk

◮ This will block diagonalize to a diagonal matrix and we get

(something close to) Yudin’s LP bound

slide-54
SLIDE 54

Block diagonalization

◮ For computations use the symmetry of V and h, expressed by

the action of a group Γ, and Bochner’s theorem to block diagonalize the matrix F

◮ For this we need a symmetry adapted basis of C(It) ◮ If t = 1 and V = S2, then

C(It) ≃ R ⊕ C(S2) = R ⊕

  • k=0

Hk

◮ This will block diagonalize to a diagonal matrix and we get

(something close to) Yudin’s LP bound

◮ In general C(It) injects into C(V )⊙t

slide-55
SLIDE 55

Block diagonalization

◮ For computations use the symmetry of V and h, expressed by

the action of a group Γ, and Bochner’s theorem to block diagonalize the matrix F

◮ For this we need a symmetry adapted basis of C(It) ◮ If t = 1 and V = S2, then

C(It) ≃ R ⊕ C(S2) = R ⊕

  • k=0

Hk

◮ This will block diagonalize to a diagonal matrix and we get

(something close to) Yudin’s LP bound

◮ In general C(It) injects into C(V )⊙t ◮ C(V )⊙t can be written in terms of tensor products of the

irreducible subspaces of C(V )

slide-56
SLIDE 56

Block diagonalization

◮ For computations use the symmetry of V and h, expressed by

the action of a group Γ, and Bochner’s theorem to block diagonalize the matrix F

◮ For this we need a symmetry adapted basis of C(It) ◮ If t = 1 and V = S2, then

C(It) ≃ R ⊕ C(S2) = R ⊕

  • k=0

Hk

◮ This will block diagonalize to a diagonal matrix and we get

(something close to) Yudin’s LP bound

◮ In general C(It) injects into C(V )⊙t ◮ C(V )⊙t can be written in terms of tensor products of the

irreducible subspaces of C(V )

◮ If we know how to decompose C(V ) into irreducibles, and how

to decompose tensor products of those irreducibles into irreducibles, then we have a symmerty adapted basis of Vt

slide-57
SLIDE 57

The case t = 2 and V = S2

◮ We know how to these decompositions from the quantum

mechanics literature (angular momentum coupling): use Clebsch-Gordan coefficients

slide-58
SLIDE 58

The case t = 2 and V = S2

◮ We know how to these decompositions from the quantum

mechanics literature (angular momentum coupling): use Clebsch-Gordan coefficients

◮ The affine constraints in E∗ t,d are nonnegativity constraints of

a polynomial p ∈ R[x1, . . . , x4], where each xi is a vector of 3 variables (the coefficients of these polynomials depend on the entries in the block diagonalization of F)

slide-59
SLIDE 59

The case t = 2 and V = S2

◮ We know how to these decompositions from the quantum

mechanics literature (angular momentum coupling): use Clebsch-Gordan coefficients

◮ The affine constraints in E∗ t,d are nonnegativity constraints of

a polynomial p ∈ R[x1, . . . , x4], where each xi is a vector of 3 variables (the coefficients of these polynomials depend on the entries in the block diagonalization of F)

◮ We have p(γx1, . . . , γx4) = p(x1, . . . , x4) for all γ ∈ O(3)

slide-60
SLIDE 60

The case t = 2 and V = S2

◮ We know how to these decompositions from the quantum

mechanics literature (angular momentum coupling): use Clebsch-Gordan coefficients

◮ The affine constraints in E∗ t,d are nonnegativity constraints of

a polynomial p ∈ R[x1, . . . , x4], where each xi is a vector of 3 variables (the coefficients of these polynomials depend on the entries in the block diagonalization of F)

◮ We have p(γx1, . . . , γx4) = p(x1, . . . , x4) for all γ ∈ O(3) ◮ Invariant theory: there is a polynomial q such that

p(x1, . . . , x4) = q(x1 · x2, . . . , x3 · x4)

slide-61
SLIDE 61

The case t = 2 and V = S2

◮ We know how to these decompositions from the quantum

mechanics literature (angular momentum coupling): use Clebsch-Gordan coefficients

◮ The affine constraints in E∗ t,d are nonnegativity constraints of

a polynomial p ∈ R[x1, . . . , x4], where each xi is a vector of 3 variables (the coefficients of these polynomials depend on the entries in the block diagonalization of F)

◮ We have p(γx1, . . . , γx4) = p(x1, . . . , x4) for all γ ∈ O(3) ◮ Invariant theory: there is a polynomial q such that

p(x1, . . . , x4) = q(x1 · x2, . . . , x3 · x4)

◮ Model nonnegativity constraints as sum of squares constraints

using Putinar’s theorem from real algebraic geometry

slide-62
SLIDE 62

The case t = 2 and V = S2

◮ We know how to these decompositions from the quantum

mechanics literature (angular momentum coupling): use Clebsch-Gordan coefficients

◮ The affine constraints in E∗ t,d are nonnegativity constraints of

a polynomial p ∈ R[x1, . . . , x4], where each xi is a vector of 3 variables (the coefficients of these polynomials depend on the entries in the block diagonalization of F)

◮ We have p(γx1, . . . , γx4) = p(x1, . . . , x4) for all γ ∈ O(3) ◮ Invariant theory: there is a polynomial q such that

p(x1, . . . , x4) = q(x1 · x2, . . . , x3 · x4)

◮ Model nonnegativity constraints as sum of squares constraints

using Putinar’s theorem from real algebraic geometry

◮ A sum of squares polynomial s can be written as

s(x) = v(x)TQv(x), where Q is a positive semidefinite matrix and v(x) a vector containing all monomials up to some degree

slide-63
SLIDE 63

More symmetry

◮ More symmetry: p(x1, . . . , x4) = p(xσ(1), . . . , xσ(4)) for all

permutations σ ∈ S4

slide-64
SLIDE 64

More symmetry

◮ More symmetry: p(x1, . . . , x4) = p(xσ(1), . . . , xσ(4)) for all

permutations σ ∈ S4

◮ This means that q is symmetric under a subgroup of S6

slide-65
SLIDE 65

More symmetry

◮ More symmetry: p(x1, . . . , x4) = p(xσ(1), . . . , xσ(4)) for all

permutations σ ∈ S4

◮ This means that q is symmetric under a subgroup of S6 ◮ Use this to block diagonalize the positive semidefinite

matrices showing up in the sums of squares characterizations

slide-66
SLIDE 66

More symmetry

◮ More symmetry: p(x1, . . . , x4) = p(xσ(1), . . . , xσ(4)) for all

permutations σ ∈ S4

◮ This means that q is symmetric under a subgroup of S6 ◮ Use this to block diagonalize the positive semidefinite

matrices showing up in the sums of squares characterizations

◮ We give a symmetrized version of Putinar’s theorem using the

method of Gatermann and Parillo for symmetry reduction in sums of squares characterizations

slide-67
SLIDE 67

More symmetry

◮ More symmetry: p(x1, . . . , x4) = p(xσ(1), . . . , xσ(4)) for all

permutations σ ∈ S4

◮ This means that q is symmetric under a subgroup of S6 ◮ Use this to block diagonalize the positive semidefinite

matrices showing up in the sums of squares characterizations

◮ We give a symmetrized version of Putinar’s theorem using the

method of Gatermann and Parillo for symmetry reduction in sums of squares characterizations

◮ Significant simplifications in the semidefinite programs

slide-68
SLIDE 68

More symmetry

◮ More symmetry: p(x1, . . . , x4) = p(xσ(1), . . . , xσ(4)) for all

permutations σ ∈ S4

◮ This means that q is symmetric under a subgroup of S6 ◮ Use this to block diagonalize the positive semidefinite

matrices showing up in the sums of squares characterizations

◮ We give a symmetrized version of Putinar’s theorem using the

method of Gatermann and Parillo for symmetry reduction in sums of squares characterizations

◮ Significant simplifications in the semidefinite programs ◮ Not clear yet whether we can compute E∗ 2,d for large enough d

(with current SDP solvers) to get improved bounds for S2

slide-69
SLIDE 69

More symmetry

◮ More symmetry: p(x1, . . . , x4) = p(xσ(1), . . . , xσ(4)) for all

permutations σ ∈ S4

◮ This means that q is symmetric under a subgroup of S6 ◮ Use this to block diagonalize the positive semidefinite

matrices showing up in the sums of squares characterizations

◮ We give a symmetrized version of Putinar’s theorem using the

method of Gatermann and Parillo for symmetry reduction in sums of squares characterizations

◮ Significant simplifications in the semidefinite programs ◮ Not clear yet whether we can compute E∗ 2,d for large enough d

(with current SDP solvers) to get improved bounds for S2

◮ Toy example: E1 is not sharp for 3 points on S1 with the

Lennard-Jones potential

slide-70
SLIDE 70

More symmetry

◮ More symmetry: p(x1, . . . , x4) = p(xσ(1), . . . , xσ(4)) for all

permutations σ ∈ S4

◮ This means that q is symmetric under a subgroup of S6 ◮ Use this to block diagonalize the positive semidefinite

matrices showing up in the sums of squares characterizations

◮ We give a symmetrized version of Putinar’s theorem using the

method of Gatermann and Parillo for symmetry reduction in sums of squares characterizations

◮ Significant simplifications in the semidefinite programs ◮ Not clear yet whether we can compute E∗ 2,d for large enough d

(with current SDP solvers) to get improved bounds for S2

◮ Toy example: E1 is not sharp for 3 points on S1 with the

Lennard-Jones potential

◮ Using a reduction to 3 variables using trigonometric

polynomials we compute that E2 = E (up to solver precision)

slide-71
SLIDE 71

Thank you!