SLIDE 1
Energy minimization via moment hierarchies David de Laat (TU Delft) - - PowerPoint PPT Presentation
Energy minimization via moment hierarchies David de Laat (TU Delft) - - PowerPoint PPT Presentation
Energy minimization via moment hierarchies David de Laat (TU Delft) ESI Workshop on Optimal Point Configurations and Applications 16 October 2014 Energy minimization What is the minimal potential energy E when we put N particles with pair
SLIDE 2
SLIDE 3
Energy minimization
◮ What is the minimal potential energy E when we put N
particles with pair potential h in a container V ?
◮ Example: For the Thomson problem we take
V = S2 and h({x, y}) = 1 x − y
SLIDE 4
Energy minimization
◮ What is the minimal potential energy E when we put N
particles with pair potential h in a container V ?
◮ Example: For the Thomson problem we take
V = S2 and h({x, y}) = 1 x − y
◮ As an optimization problem:
E = min
S∈(V
N)
- P∈(S
2)
h(P)
SLIDE 5
Approach
◮ Configurations provide upper bounds on the optimal energy E
SLIDE 6
Approach
◮ Configurations provide upper bounds on the optimal energy E ◮ To prove a configuration is good (or optimal) we need good
lower bounds for E
SLIDE 7
Approach
◮ Configurations provide upper bounds on the optimal energy E ◮ To prove a configuration is good (or optimal) we need good
lower bounds for E Some systematic approaches for obtaining bounds:
◮ Linear programming bounds using the pair correlation function
[Delsarte 1973, Delsarte-Goethals-Seidel 1977, Yudin 1992]
SLIDE 8
Approach
◮ Configurations provide upper bounds on the optimal energy E ◮ To prove a configuration is good (or optimal) we need good
lower bounds for E Some systematic approaches for obtaining bounds:
◮ Linear programming bounds using the pair correlation function
[Delsarte 1973, Delsarte-Goethals-Seidel 1977, Yudin 1992]
◮ 3-point bounds using 3-point correlation functions and
constraints arising from the stabilizer subgroup of 1 point [Schrijver 2005, Bachoc-Vallentin 2008, Cohn-Woo 2012]
SLIDE 9
Approach
◮ Configurations provide upper bounds on the optimal energy E ◮ To prove a configuration is good (or optimal) we need good
lower bounds for E Some systematic approaches for obtaining bounds:
◮ Linear programming bounds using the pair correlation function
[Delsarte 1973, Delsarte-Goethals-Seidel 1977, Yudin 1992]
◮ 3-point bounds using 3-point correlation functions and
constraints arising from the stabilizer subgroup of 1 point [Schrijver 2005, Bachoc-Vallentin 2008, Cohn-Woo 2012]
◮ k-point bounds using stabilizer subgroup of k − 2 points
[Musin 2007]
SLIDE 10
Approach
◮ Configurations provide upper bounds on the optimal energy E ◮ To prove a configuration is good (or optimal) we need good
lower bounds for E Some systematic approaches for obtaining bounds:
◮ Linear programming bounds using the pair correlation function
[Delsarte 1973, Delsarte-Goethals-Seidel 1977, Yudin 1992]
◮ 3-point bounds using 3-point correlation functions and
constraints arising from the stabilizer subgroup of 1 point [Schrijver 2005, Bachoc-Vallentin 2008, Cohn-Woo 2012]
◮ k-point bounds using stabilizer subgroup of k − 2 points
[Musin 2007]
◮ Hierarchy for packing problems [L.-Vallentin 2014]
SLIDE 11
This talk
◮ Hierarchy obtained by generalizing Lasserre’s hierarchy from
combinatorial optimization to the continuous setting
SLIDE 12
This talk
◮ Hierarchy obtained by generalizing Lasserre’s hierarchy from
combinatorial optimization to the continuous setting
◮ Finite convergence to the optimal energy
SLIDE 13
This talk
◮ Hierarchy obtained by generalizing Lasserre’s hierarchy from
combinatorial optimization to the continuous setting
◮ Finite convergence to the optimal energy ◮ A duality theory
SLIDE 14
This talk
◮ Hierarchy obtained by generalizing Lasserre’s hierarchy from
combinatorial optimization to the continuous setting
◮ Finite convergence to the optimal energy ◮ A duality theory ◮ Reduction to a converging sequence of semidefinite programs
SLIDE 15
This talk
◮ Hierarchy obtained by generalizing Lasserre’s hierarchy from
combinatorial optimization to the continuous setting
◮ Finite convergence to the optimal energy ◮ A duality theory ◮ Reduction to a converging sequence of semidefinite programs ◮ Towards computations using several types of symmetry
reduction
SLIDE 16
Approach
E
SLIDE 17
Approach
E
Difficult minimization problem
SLIDE 18
Approach
Et E
Difficult minimization problem
SLIDE 19
Approach
Et E
Difficult minimization problem Relaxation to a conic program: Infinite dimensional minimization problem
SLIDE 20
Approach
Et E E∗
t
Difficult minimization problem Relaxation to a conic program: Infinite dimensional minimization problem
SLIDE 21
Approach
Et E E∗
t
Difficult minimization problem Relaxation to a conic program: Conic dual: Infinite dimensional minimization problem Infinite dimensional maximization problem
SLIDE 22
Approach
Et E E∗
t
E∗
t,d
Difficult minimization problem Relaxation to a conic program: Conic dual: Infinite dimensional minimization problem Infinite dimensional maximization problem
SLIDE 23
Approach
Et E E∗
t
E∗
t,d
Difficult minimization problem Relaxation to a conic program: Conic dual: Semi-infinite semidefinite program Infinite dimensional minimization problem Infinite dimensional maximization problem
SLIDE 24
The minimization problem
◮ I=t (It) is the set of subsets of V which
◮ have cardinality t (≤ t) ◮ contain no points which are too close
SLIDE 25
The minimization problem
◮ I=t (It) is the set of subsets of V which
◮ have cardinality t (≤ t) ◮ contain no points which are too close
◮ Assuming h({x, y}) → ∞ when x and y converge, we have
E = min
S∈I=N
- P∈(S
2)
h(P)
SLIDE 26
The minimization problem
◮ I=t (It) is the set of subsets of V which
◮ have cardinality t (≤ t) ◮ contain no points which are too close
◮ Assuming h({x, y}) → ∞ when x and y converge, we have
E = min
S∈I=N
- P∈(S
2)
h(P)
◮ We will also assume that V is compact and h continuous
SLIDE 27
The minimization problem
◮ I=t (It) is the set of subsets of V which
◮ have cardinality t (≤ t) ◮ contain no points which are too close
◮ Assuming h({x, y}) → ∞ when x and y converge, we have
E = min
S∈I=N
- P∈(S
2)
h(P)
◮ We will also assume that V is compact and h continuous ◮ I=t gets its topology as a subset of a quotient of V t
SLIDE 28
Moment hierarchy of relaxations
◮ In the relaxation Et we minimize over measures λ on the
space Is, where s = min{2t, N}
SLIDE 29
Moment hierarchy of relaxations
◮ In the relaxation Et we minimize over measures λ on the
space Is, where s = min{2t, N} Lemma When t = N, the feasible measures λ are (generalized) convex combinations of measures χS =
- R⊆S
δR where S ∈ I=N
SLIDE 30
Moment hierarchy of relaxations
◮ In the relaxation Et we minimize over measures λ on the
space Is, where s = min{2t, N} Lemma When t = N, the feasible measures λ are (generalized) convex combinations of measures χS =
- R⊆S
δR where S ∈ I=N
◮ Objective function: λ(h) =
- I=N h(S) dλ(S)
SLIDE 31
Moment hierarchy of relaxations
◮ In the relaxation Et we minimize over measures λ on the
space Is, where s = min{2t, N} Lemma When t = N, the feasible measures λ are (generalized) convex combinations of measures χS =
- R⊆S
δR where S ∈ I=N
◮ Objective function: λ(h) =
- I=N h(S) dλ(S)
◮ Moment constraints: A∗ t λ ∈ M(It × It)0
SLIDE 32
Moment hierarchy of relaxations
◮ In the relaxation Et we minimize over measures λ on the
space Is, where s = min{2t, N} Lemma When t = N, the feasible measures λ are (generalized) convex combinations of measures χS =
- R⊆S
δR where S ∈ I=N
◮ Objective function: λ(h) =
- I=N h(S) dλ(S)
◮ Moment constraints: A∗ t λ ∈ M(It × It)0 ◮ Here A∗ t is an operator M(Is) → M(It × It)
SLIDE 33
Moment hierarchy of relaxations
◮ In the relaxation Et we minimize over measures λ on the
space Is, where s = min{2t, N} Lemma When t = N, the feasible measures λ are (generalized) convex combinations of measures χS =
- R⊆S
δR where S ∈ I=N
◮ Objective function: λ(h) =
- I=N h(S) dλ(S)
◮ Moment constraints: A∗ t λ ∈ M(It × It)0 ◮ Here A∗ t is an operator M(Is) → M(It × It) ◮ M(It × It)0 is the cone dual to the cone C(It × It)0 of
positive kernels
SLIDE 34
Moment hierarchy of relaxations
◮ In the relaxation Et we minimize over measures λ on the
space Is, where s = min{2t, N} Lemma When t = N, the feasible measures λ are (generalized) convex combinations of measures χS =
- R⊆S
δR where S ∈ I=N
◮ Objective function: λ(h) =
- I=N h(S) dλ(S)
◮ Moment constraints: A∗ t λ ∈ M(It × It)0 ◮ Here A∗ t is an operator M(Is) → M(It × It) ◮ M(It × It)0 is the cone dual to the cone C(It × It)0 of
positive kernels: µ(K) ≥ 0 for all K 0
SLIDE 35
Moment hierarchy of relaxations
◮ In the relaxation Et we minimize over measures λ on the
space Is, where s = min{2t, N} Lemma When t = N, the feasible measures λ are (generalized) convex combinations of measures χS =
- R⊆S
δR where S ∈ I=N
◮ Objective function: λ(h) =
- I=N h(S) dλ(S)
◮ Moment constraints: A∗ t λ ∈ M(It × It)0 ◮ Here A∗ t is an operator M(Is) → M(It × It) ◮ M(It × It)0 is the cone dual to the cone C(It × It)0 of
positive kernels: µ(K) ≥ 0 for all K 0
◮ We have χS(h) = P∈(S
2) h(P)
SLIDE 36
Moment hierarchy of relaxations
◮ In the relaxation Et we minimize over measures λ on the
space Is, where s = min{2t, N} Lemma When t = N, the feasible measures λ are (generalized) convex combinations of measures χS =
- R⊆S
δR where S ∈ I=N
◮ Objective function: λ(h) =
- I=N h(S) dλ(S)
◮ Moment constraints: A∗ t λ ∈ M(It × It)0 ◮ Here A∗ t is an operator M(Is) → M(It × It) ◮ M(It × It)0 is the cone dual to the cone C(It × It)0 of
positive kernels: µ(K) ≥ 0 for all K 0
◮ We have χS(h) = P∈(S
2) h(P)
Theorem (Finite convergence) We have E1 ≤ · · · ≤ EN = E
SLIDE 37
Dual hierarchy
◮ Et is a minimization problem, so we need an optimal solution
to find a lower bound
SLIDE 38
Dual hierarchy
◮ Et is a minimization problem, so we need an optimal solution
to find a lower bound
◮ The conic dual E∗ t is a maximization problem where any
feasible solution provides an upper bound
SLIDE 39
Dual hierarchy
◮ Et is a minimization problem, so we need an optimal solution
to find a lower bound
◮ The conic dual E∗ t is a maximization problem where any
feasible solution provides an upper bound
◮ In E∗ t optimization is over scalars ai ∈ R and positive definite
kernels K ∈ C(It × It)0
SLIDE 40
Dual hierarchy
◮ Et is a minimization problem, so we need an optimal solution
to find a lower bound
◮ The conic dual E∗ t is a maximization problem where any
feasible solution provides an upper bound
◮ In E∗ t optimization is over scalars ai ∈ R and positive definite
kernels K ∈ C(It × It)0
◮ The dual program:
E∗
t = sup
- s
- i=0
N
i
- ai : a0, . . . , as ∈ R, K ∈ C(It × It)0,
ai + AtK ≤ h on I=i for i = 0, . . . , s
SLIDE 41
Dual hierarchy
◮ Et is a minimization problem, so we need an optimal solution
to find a lower bound
◮ The conic dual E∗ t is a maximization problem where any
feasible solution provides an upper bound
◮ In E∗ t optimization is over scalars ai ∈ R and positive definite
kernels K ∈ C(It × It)0
◮ The dual program:
E∗
t = sup
- s
- i=0
N
i
- ai : a0, . . . , as ∈ R, K ∈ C(It × It)0,
ai + AtK ≤ h on I=i for i = 0, . . . , s
- ◮ Here At is the linear operator C(It × It) → C(It) given by
AtK(S) =
J,J′∈It:J∪J′=S K(J, J′)
SLIDE 42
Dual hierarchy
◮ Et is a minimization problem, so we need an optimal solution
to find a lower bound
◮ The conic dual E∗ t is a maximization problem where any
feasible solution provides an upper bound
◮ In E∗ t optimization is over scalars ai ∈ R and positive definite
kernels K ∈ C(It × It)0
◮ The dual program:
E∗
t = sup
- s
- i=0
N
i
- ai : a0, . . . , as ∈ R, K ∈ C(It × It)0,
ai + AtK ≤ h on I=i for i = 0, . . . , s
- ◮ Here At is the linear operator C(It × It) → C(It) given by
AtK(S) =
J,J′∈It:J∪J′=S K(J, J′)
Theorem Strong duality holds: Et = E∗
t for each t
SLIDE 43
Closing the gaps
Et E E∗
t
E∗
t,d
Difficult minimization problem Relaxation to a conic program: Conic dual: Semi-infinite semidefinite program Infinite dimensional minimization problem Infinite dimensional maximization problem
SLIDE 44
Finite dimensional approximations to E∗
t
◮ Define E∗ t,d by replacing the cone C(It × It)0 in E∗ t by a
finite dimensional inner approximating cone Cd
SLIDE 45
Finite dimensional approximations to E∗
t
◮ Define E∗ t,d by replacing the cone C(It × It)0 in E∗ t by a
finite dimensional inner approximating cone Cd
◮ Let e1, e2, . . . be a dense sequence in C(It) and define
Cd =
- d
- i,j=1
Fi,jei ⊗ ej : F ∈ Rd×d positive semidefinite
SLIDE 46
Finite dimensional approximations to E∗
t
◮ Define E∗ t,d by replacing the cone C(It × It)0 in E∗ t by a
finite dimensional inner approximating cone Cd
◮ Let e1, e2, . . . be a dense sequence in C(It) and define
Cd =
- d
- i,j=1
Fi,jei ⊗ ej : F ∈ Rd×d positive semidefinite
- Lemma
Suppose X is a compact metric space. Then the extreme rays of the cone C(X ×X)0 are precisely the kernels f ⊗f with f ∈ C(X)
SLIDE 47
Finite dimensional approximations to E∗
t
◮ Define E∗ t,d by replacing the cone C(It × It)0 in E∗ t by a
finite dimensional inner approximating cone Cd
◮ Let e1, e2, . . . be a dense sequence in C(It) and define
Cd =
- d
- i,j=1
Fi,jei ⊗ ej : F ∈ Rd×d positive semidefinite
- Lemma
Suppose X is a compact metric space. Then the extreme rays of the cone C(X ×X)0 are precisely the kernels f ⊗f with f ∈ C(X)
◮ This implies ∪∞ d=0Cd is uniformly dense in C(It × It)0
SLIDE 48
Finite dimensional approximations to E∗
t
◮ Define E∗ t,d by replacing the cone C(It × It)0 in E∗ t by a
finite dimensional inner approximating cone Cd
◮ Let e1, e2, . . . be a dense sequence in C(It) and define
Cd =
- d
- i,j=1
Fi,jei ⊗ ej : F ∈ Rd×d positive semidefinite
- Lemma
Suppose X is a compact metric space. Then the extreme rays of the cone C(X ×X)0 are precisely the kernels f ⊗f with f ∈ C(X)
◮ This implies ∪∞ d=0Cd is uniformly dense in C(It × It)0
Theorem If V is a compact metric space, then E∗
t,d → E∗ t as d → ∞ for all t
SLIDE 49
Block diagonalization
◮ For computations use the symmetry of V and h, expressed by
the action of a group Γ, and Bochner’s theorem to block diagonalize the matrix F
SLIDE 50
Block diagonalization
◮ For computations use the symmetry of V and h, expressed by
the action of a group Γ, and Bochner’s theorem to block diagonalize the matrix F
◮ For this we need a symmetry adapted basis of C(It)
SLIDE 51
Block diagonalization
◮ For computations use the symmetry of V and h, expressed by
the action of a group Γ, and Bochner’s theorem to block diagonalize the matrix F
◮ For this we need a symmetry adapted basis of C(It) ◮ If t = 1 and V = S2, then
C(It) ≃ R ⊕ C(S2)
SLIDE 52
Block diagonalization
◮ For computations use the symmetry of V and h, expressed by
the action of a group Γ, and Bochner’s theorem to block diagonalize the matrix F
◮ For this we need a symmetry adapted basis of C(It) ◮ If t = 1 and V = S2, then
C(It) ≃ R ⊕ C(S2) = R ⊕
∞
- k=0
Hk
SLIDE 53
Block diagonalization
◮ For computations use the symmetry of V and h, expressed by
the action of a group Γ, and Bochner’s theorem to block diagonalize the matrix F
◮ For this we need a symmetry adapted basis of C(It) ◮ If t = 1 and V = S2, then
C(It) ≃ R ⊕ C(S2) = R ⊕
∞
- k=0
Hk
◮ This will block diagonalize to a diagonal matrix and we get
(something close to) Yudin’s LP bound
SLIDE 54
Block diagonalization
◮ For computations use the symmetry of V and h, expressed by
the action of a group Γ, and Bochner’s theorem to block diagonalize the matrix F
◮ For this we need a symmetry adapted basis of C(It) ◮ If t = 1 and V = S2, then
C(It) ≃ R ⊕ C(S2) = R ⊕
∞
- k=0
Hk
◮ This will block diagonalize to a diagonal matrix and we get
(something close to) Yudin’s LP bound
◮ In general C(It) injects into C(V )⊙t
SLIDE 55
Block diagonalization
◮ For computations use the symmetry of V and h, expressed by
the action of a group Γ, and Bochner’s theorem to block diagonalize the matrix F
◮ For this we need a symmetry adapted basis of C(It) ◮ If t = 1 and V = S2, then
C(It) ≃ R ⊕ C(S2) = R ⊕
∞
- k=0
Hk
◮ This will block diagonalize to a diagonal matrix and we get
(something close to) Yudin’s LP bound
◮ In general C(It) injects into C(V )⊙t ◮ C(V )⊙t can be written in terms of tensor products of the
irreducible subspaces of C(V )
SLIDE 56
Block diagonalization
◮ For computations use the symmetry of V and h, expressed by
the action of a group Γ, and Bochner’s theorem to block diagonalize the matrix F
◮ For this we need a symmetry adapted basis of C(It) ◮ If t = 1 and V = S2, then
C(It) ≃ R ⊕ C(S2) = R ⊕
∞
- k=0
Hk
◮ This will block diagonalize to a diagonal matrix and we get
(something close to) Yudin’s LP bound
◮ In general C(It) injects into C(V )⊙t ◮ C(V )⊙t can be written in terms of tensor products of the
irreducible subspaces of C(V )
◮ If we know how to decompose C(V ) into irreducibles, and how
to decompose tensor products of those irreducibles into irreducibles, then we have a symmerty adapted basis of Vt
SLIDE 57
The case t = 2 and V = S2
◮ We know how to these decompositions from the quantum
mechanics literature (angular momentum coupling): use Clebsch-Gordan coefficients
SLIDE 58
The case t = 2 and V = S2
◮ We know how to these decompositions from the quantum
mechanics literature (angular momentum coupling): use Clebsch-Gordan coefficients
◮ The affine constraints in E∗ t,d are nonnegativity constraints of
a polynomial p ∈ R[x1, . . . , x4], where each xi is a vector of 3 variables (the coefficients of these polynomials depend on the entries in the block diagonalization of F)
SLIDE 59
The case t = 2 and V = S2
◮ We know how to these decompositions from the quantum
mechanics literature (angular momentum coupling): use Clebsch-Gordan coefficients
◮ The affine constraints in E∗ t,d are nonnegativity constraints of
a polynomial p ∈ R[x1, . . . , x4], where each xi is a vector of 3 variables (the coefficients of these polynomials depend on the entries in the block diagonalization of F)
◮ We have p(γx1, . . . , γx4) = p(x1, . . . , x4) for all γ ∈ O(3)
SLIDE 60
The case t = 2 and V = S2
◮ We know how to these decompositions from the quantum
mechanics literature (angular momentum coupling): use Clebsch-Gordan coefficients
◮ The affine constraints in E∗ t,d are nonnegativity constraints of
a polynomial p ∈ R[x1, . . . , x4], where each xi is a vector of 3 variables (the coefficients of these polynomials depend on the entries in the block diagonalization of F)
◮ We have p(γx1, . . . , γx4) = p(x1, . . . , x4) for all γ ∈ O(3) ◮ Invariant theory: there is a polynomial q such that
p(x1, . . . , x4) = q(x1 · x2, . . . , x3 · x4)
SLIDE 61
The case t = 2 and V = S2
◮ We know how to these decompositions from the quantum
mechanics literature (angular momentum coupling): use Clebsch-Gordan coefficients
◮ The affine constraints in E∗ t,d are nonnegativity constraints of
a polynomial p ∈ R[x1, . . . , x4], where each xi is a vector of 3 variables (the coefficients of these polynomials depend on the entries in the block diagonalization of F)
◮ We have p(γx1, . . . , γx4) = p(x1, . . . , x4) for all γ ∈ O(3) ◮ Invariant theory: there is a polynomial q such that
p(x1, . . . , x4) = q(x1 · x2, . . . , x3 · x4)
◮ Model nonnegativity constraints as sum of squares constraints
using Putinar’s theorem from real algebraic geometry
SLIDE 62
The case t = 2 and V = S2
◮ We know how to these decompositions from the quantum
mechanics literature (angular momentum coupling): use Clebsch-Gordan coefficients
◮ The affine constraints in E∗ t,d are nonnegativity constraints of
a polynomial p ∈ R[x1, . . . , x4], where each xi is a vector of 3 variables (the coefficients of these polynomials depend on the entries in the block diagonalization of F)
◮ We have p(γx1, . . . , γx4) = p(x1, . . . , x4) for all γ ∈ O(3) ◮ Invariant theory: there is a polynomial q such that
p(x1, . . . , x4) = q(x1 · x2, . . . , x3 · x4)
◮ Model nonnegativity constraints as sum of squares constraints
using Putinar’s theorem from real algebraic geometry
◮ A sum of squares polynomial s can be written as
s(x) = v(x)TQv(x), where Q is a positive semidefinite matrix and v(x) a vector containing all monomials up to some degree
SLIDE 63
More symmetry
◮ More symmetry: p(x1, . . . , x4) = p(xσ(1), . . . , xσ(4)) for all
permutations σ ∈ S4
SLIDE 64
More symmetry
◮ More symmetry: p(x1, . . . , x4) = p(xσ(1), . . . , xσ(4)) for all
permutations σ ∈ S4
◮ This means that q is symmetric under a subgroup of S6
SLIDE 65
More symmetry
◮ More symmetry: p(x1, . . . , x4) = p(xσ(1), . . . , xσ(4)) for all
permutations σ ∈ S4
◮ This means that q is symmetric under a subgroup of S6 ◮ Use this to block diagonalize the positive semidefinite
matrices showing up in the sums of squares characterizations
SLIDE 66
More symmetry
◮ More symmetry: p(x1, . . . , x4) = p(xσ(1), . . . , xσ(4)) for all
permutations σ ∈ S4
◮ This means that q is symmetric under a subgroup of S6 ◮ Use this to block diagonalize the positive semidefinite
matrices showing up in the sums of squares characterizations
◮ We give a symmetrized version of Putinar’s theorem using the
method of Gatermann and Parillo for symmetry reduction in sums of squares characterizations
SLIDE 67
More symmetry
◮ More symmetry: p(x1, . . . , x4) = p(xσ(1), . . . , xσ(4)) for all
permutations σ ∈ S4
◮ This means that q is symmetric under a subgroup of S6 ◮ Use this to block diagonalize the positive semidefinite
matrices showing up in the sums of squares characterizations
◮ We give a symmetrized version of Putinar’s theorem using the
method of Gatermann and Parillo for symmetry reduction in sums of squares characterizations
◮ Significant simplifications in the semidefinite programs
SLIDE 68
More symmetry
◮ More symmetry: p(x1, . . . , x4) = p(xσ(1), . . . , xσ(4)) for all
permutations σ ∈ S4
◮ This means that q is symmetric under a subgroup of S6 ◮ Use this to block diagonalize the positive semidefinite
matrices showing up in the sums of squares characterizations
◮ We give a symmetrized version of Putinar’s theorem using the
method of Gatermann and Parillo for symmetry reduction in sums of squares characterizations
◮ Significant simplifications in the semidefinite programs ◮ Not clear yet whether we can compute E∗ 2,d for large enough d
(with current SDP solvers) to get improved bounds for S2
SLIDE 69
More symmetry
◮ More symmetry: p(x1, . . . , x4) = p(xσ(1), . . . , xσ(4)) for all
permutations σ ∈ S4
◮ This means that q is symmetric under a subgroup of S6 ◮ Use this to block diagonalize the positive semidefinite
matrices showing up in the sums of squares characterizations
◮ We give a symmetrized version of Putinar’s theorem using the
method of Gatermann and Parillo for symmetry reduction in sums of squares characterizations
◮ Significant simplifications in the semidefinite programs ◮ Not clear yet whether we can compute E∗ 2,d for large enough d
(with current SDP solvers) to get improved bounds for S2
◮ Toy example: E1 is not sharp for 3 points on S1 with the
Lennard-Jones potential
SLIDE 70
More symmetry
◮ More symmetry: p(x1, . . . , x4) = p(xσ(1), . . . , xσ(4)) for all
permutations σ ∈ S4
◮ This means that q is symmetric under a subgroup of S6 ◮ Use this to block diagonalize the positive semidefinite
matrices showing up in the sums of squares characterizations
◮ We give a symmetrized version of Putinar’s theorem using the
method of Gatermann and Parillo for symmetry reduction in sums of squares characterizations
◮ Significant simplifications in the semidefinite programs ◮ Not clear yet whether we can compute E∗ 2,d for large enough d
(with current SDP solvers) to get improved bounds for S2
◮ Toy example: E1 is not sharp for 3 points on S1 with the
Lennard-Jones potential
◮ Using a reduction to 3 variables using trigonometric
polynomials we compute that E2 = E (up to solver precision)
SLIDE 71