The quest of efficiency and certification in polynomial optimization - - PowerPoint PPT Presentation

the quest of efficiency and certification in polynomial
SMART_READER_LITE
LIVE PREVIEW

The quest of efficiency and certification in polynomial optimization - - PowerPoint PPT Presentation

The quest of efficiency and certification in polynomial optimization Victor Magron , CNRSLAAS SPOT, Toulouse 4 November 2019 6 5 1 4 3 2 The Moment-Sums of Squares Hierarchy NP-hard NON CONVEX Problem f = inf f ( x ) Theory


slide-1
SLIDE 1

The quest of efficiency and certification in polynomial optimization

Victor Magron, CNRS–LAAS SPOT, Toulouse 4 November 2019

6 4 5 1 2 3

slide-2
SLIDE 2

The Moment-Sums of Squares Hierarchy

NP-hard NON CONVEX Problem f ⋆ = inf f (x) Theory (Primal) (Dual) inf

  • f dµ

sup λ with µ proba ⇒

INFINITE LP

⇐ with f − λ 0

Victor Magron The quest of efficiency and certification in polynomial optimization 1 / 42

slide-3
SLIDE 3

The Moment-Sums of Squares Hierarchy

NP-hard NON CONVEX Problem f ⋆ = inf f (x) Practice (Primal Relaxation) (Dual Strengthening) moments

  • xα dµ

f − λ = sum of squares finite number ⇒ SDP ⇐ fixed degree LASSERRE’S HIERARCHY of CONVEX PROBLEMS ↑ f ∗ [Lasserre/Parrilo 01] degree d & n vars Numeric solvers

= ⇒ (n+2d

n ) SDP VARIABLES

= ⇒ Approx Certificate

Victor Magron The quest of efficiency and certification in polynomial optimization 1 / 42

slide-4
SLIDE 4

Success Stories: Lasserre’s Hierarchy

MODELING POWER: Cast as ∞-dimensional LP over measures STATIC Polynomial Optimization Optimal Powerflow n ≃ 103 [Josz et al 16] DYNAMICAL Polynomial Optimization Regions of attraction [Henrion-Korda 14] Reachable sets [Magron et al 17]

!

APPROXIMATE OPTIMIZATION BOUNDS!

Victor Magron The quest of efficiency and certification in polynomial optimization 2 / 42

slide-5
SLIDE 5

Success Stories: Certified Optimization

Kepler’s Conjecture(1611)

The max density of sphere packings is π/ √ 18

Flyspeck : Formalizing the proof of Kepler by T.Hales (1994) Verification of thousands of “tight” nonlinear inequalities Seminal Paper:

Hales, Adams, Bauer, Dang, Harrison, Hoang, Kaliszyk, M., Mclaughlin, Nguyen, Nguyen, Nipkow, Obua, Pleso, Rute, Solovyev, Ta, Tran, Trieu, Urban, Vu & Zumkeller, Forum of Mathematics, Pi, 5 2017

CONTRIBUTION: (Non)-Polynomial optimization to verify Flyspeck inequalities

Victor Magron The quest of efficiency and certification in polynomial optimization 3 / 42

slide-6
SLIDE 6

Exploiting Sparsity Certified Polynomial Optimization

slide-7
SLIDE 7

Exploiting Sparsity Certified Polynomial Optimization

slide-8
SLIDE 8

SDP for Polynomial Optimization

NP hard General Problem: f ∗ := min

x∈X f (x)

Semialgebraic set X := {x ∈ Rn : g1(x) 0, . . . , gl(x) 0}

Victor Magron The quest of efficiency and certification in polynomial optimization 4 / 42

slide-9
SLIDE 9

SDP for Polynomial Optimization

NP hard General Problem: f ∗ := min

x∈X f (x)

Semialgebraic set X := {x ∈ Rn : g1(x) 0, . . . , gl(x) 0} X = [0, 1]2 = {x ∈ R2 : x1(1 − x1) 0, x2(1 − x2) 0}

Victor Magron The quest of efficiency and certification in polynomial optimization 4 / 42

slide-10
SLIDE 10

SDP for Polynomial Optimization

NP hard General Problem: f ∗ := min

x∈X f (x)

Semialgebraic set X := {x ∈ Rn : g1(x) 0, . . . , gl(x) 0} X = [0, 1]2 = {x ∈ R2 : x1(1 − x1) 0, x2(1 − x2) 0}

f

  • x1x2 =

−1 8 +

σ0

  • 1

2

  • x1 + x2 − 1

2 2 +

σ1

  • 1

2

g1

  • x1(1 − x1) +

σ2

  • 1

2

g2

  • x2(1 − x2)

Victor Magron The quest of efficiency and certification in polynomial optimization 4 / 42

slide-11
SLIDE 11

SDP for Polynomial Optimization

NP hard General Problem: f ∗ := min

x∈X f (x)

Semialgebraic set X := {x ∈ Rn : g1(x) 0, . . . , gl(x) 0} X = [0, 1]2 = {x ∈ R2 : x1(1 − x1) 0, x2(1 − x2) 0}

f

  • x1x2 =

−1 8 +

σ0

  • 1

2

  • x1 + x2 − 1

2 2 +

σ1

  • 1

2

g1

  • x1(1 − x1) +

σ2

  • 1

2

g2

  • x2(1 − x2)

Sums of squares (SOS) σj

Victor Magron The quest of efficiency and certification in polynomial optimization 4 / 42

slide-12
SLIDE 12

SDP for Polynomial Optimization

NP hard General Problem: f ∗ := min

x∈X f (x)

Semialgebraic set X := {x ∈ Rn : g1(x) 0, . . . , gl(x) 0} X = [0, 1]2 = {x ∈ R2 : x1(1 − x1) 0, x2(1 − x2) 0}

f

  • x1x2 =

−1 8 +

σ0

  • 1

2

  • x1 + x2 − 1

2 2 +

σ1

  • 1

2

g1

  • x1(1 − x1) +

σ2

  • 1

2

g2

  • x2(1 − x2)

Sums of squares (SOS) σj Bounded degree: Qr(X) :=

  • σ0 + ∑l

j=1 σjgj, with deg σj gj 2r

  • Victor Magron

The quest of efficiency and certification in polynomial optimization 4 / 42

slide-13
SLIDE 13

SDP for Polynomial Optimization

Hierarchy of SDP relaxations: λr := sup

λ

  • λ : f − λ ∈ Qr(X)
  • Convergence guarantees λr ↑ f ∗ [Lasserre 01]

Can be computed with SDP solvers (CSDP, SDPA, MOSEK) “No Free Lunch” Rule: (n+2d

n ) SDP variables

Victor Magron The quest of efficiency and certification in polynomial optimization 5 / 42

slide-14
SLIDE 14

Sparse Polynomial Optimization [Waki, Lasserre 06]

Correlative sparsity pattern (csp) of vars x2x5 + x3x6 − x2x3 − x5x6 + x1(−x1 + x2 + x3 − x4 + x5 + x6)

6 4 5 1 2 3

Victor Magron The quest of efficiency and certification in polynomial optimization 6 / 42

slide-15
SLIDE 15

Sparse Polynomial Optimization [Waki, Lasserre 06]

Correlative sparsity pattern (csp) of vars x2x5 + x3x6 − x2x3 − x5x6 + x1(−x1 + x2 + x3 − x4 + x5 + x6)

6 4 5 1 2 3

1 Index sets I1, . . . , Ip 2 Average size κ ❀ (κ+2d κ ) vars

I1 = {1, 4} I2 = {1, 2, 3, 5} I3 = {1, 3, 5, 6} Dense SDP: 210 vars Sparse SDP: 115 vars

Victor Magron The quest of efficiency and certification in polynomial optimization 6 / 42

slide-16
SLIDE 16

Sparse Polynomial Optimization [Waki, Lasserre 06]

Sparse f = f1 + · · · + fp with fk ∈ R[x, Ik] Sparse K = {x : gj(x) 0} with gj ∈ R[x, Ik(j)] for some k(j) Additional constraints nk − ∑i∈Ik x2

i 0 in K

Victor Magron The quest of efficiency and certification in polynomial optimization 7 / 42

slide-17
SLIDE 17

Sparse Polynomial Optimization [Waki, Lasserre 06]

Sparse f = f1 + · · · + fp with fk ∈ R[x, Ik] Sparse K = {x : gj(x) 0} with gj ∈ R[x, Ik(j)] for some k(j) Additional constraints nk − ∑i∈Ik x2

i 0 in K

RUNNING INTERSECTION PROPERTY (RIP) ∀k = 1, . . . , p − 1 Ik+1 ∩

  • jk

Ij ⊆ Ii for some i k

Victor Magron The quest of efficiency and certification in polynomial optimization 7 / 42

slide-18
SLIDE 18

Sparse Polynomial Optimization [Waki, Lasserre 06]

Sparse f = f1 + · · · + fp with fk ∈ R[x, Ik] Sparse K = {x : gj(x) 0} with gj ∈ R[x, Ik(j)] for some k(j) Additional constraints nk − ∑i∈Ik x2

i 0 in K

RUNNING INTERSECTION PROPERTY (RIP) ∀k = 1, . . . , p − 1 Ik+1 ∩

  • jk

Ij ⊆ Ii for some i k Theorem: Sparse Putinar’s representation [Lasserre 06] f > 0 on K + RIP = ⇒ f = σ01 + · · · + σ0p +

m

j=1

σjgj with σ0k ∈ Σ[x, Ik], σj ∈ Σ[x, Ik(j)]

Victor Magron The quest of efficiency and certification in polynomial optimization 7 / 42

slide-19
SLIDE 19

Sparse Examples

Chained singular function: fcs = ∑

i∈J

((xi + 10xi+1)2 + 5(xi+2 − xi+3)2 + (xi+1 − 2xi+2)4 +10(xi − xi+3)4) where J = {1, 3, 4, . . . , n − 3} and n is a multiple of 4 Ik = {k, k + 1, k + 2, k + 3}

Victor Magron The quest of efficiency and certification in polynomial optimization 8 / 42

slide-20
SLIDE 20

Sparse Examples

Chained singular function: fcs = ∑

i∈J

((xi + 10xi+1)2 + 5(xi+2 − xi+3)2 + (xi+1 − 2xi+2)4 +10(xi − xi+3)4) where J = {1, 3, 4, . . . , n − 3} and n is a multiple of 4 Ik = {k, k + 1, k + 2, k + 3} Generalized Rosenbrock function: fgR = 1 +

n−1

i=1

  • 100(xi+1 − x2

i )2 + (1 − xi+1)2

Ik = {k, k + 1}

Victor Magron The quest of efficiency and certification in polynomial optimization 8 / 42

slide-21
SLIDE 21

Roundoff Errors

Exact: f (x) := x1x2 + x3x4 Floating-point: ˆ f (x, e) := [x1x2(1 + e1) + x3x4(1 + e2)](1 + e3) x ∈ X , | ei | 2−δ δ = 24 (single) or 53 (double)

Victor Magron The quest of efficiency and certification in polynomial optimization 9 / 42

slide-22
SLIDE 22

Roundoff Errors

Input: exact f (x), floating-point ˆ f (x, e) Output: Bounds for f − ˆ f

1: Error r(x, e) := f (x) − ˆ

f (x, e) = ∑

α

rα(e)xα

2: Decompose r(x, e) = ℓ(x, e) + h(x, e), ℓ linear in e 3: Bound h(x, e) with interval arithmetic 4: Bound ℓ(x, e) with SPARSE SUMS OF SQUARES

Victor Magron The quest of efficiency and certification in polynomial optimization 10 / 42

slide-23
SLIDE 23

Exploiting Sparsity for Roundoff Error Bounds

l(x, e) = ∑m

i=1 si(x)ei

I1, . . . , Im correspond to {x, e1}, . . . , {x, em} Dense relaxation: (n+m+2d

n+m )

SDP variables Sparse relaxation: m(n+1+2d

n+1 )

SDP variables

Victor Magron The quest of efficiency and certification in polynomial optimization 11 / 42

slide-24
SLIDE 24

Preliminary Comparisons

f (x) := x2x5 + x3x6 − x2x3 − x5x6 + x1(−x1 + x2 + x3 − x4 + x5 + x6) x ∈ [4.00, 6.36]6 , e ∈ [−ǫ, ǫ]15 , ǫ = 2−53 Dense SDP: (6+15+4

6+15 ) = 12650 variables ❀ Out of memory

Victor Magron The quest of efficiency and certification in polynomial optimization 12 / 42

slide-25
SLIDE 25

Preliminary Comparisons

f (x) := x2x5 + x3x6 − x2x3 − x5x6 + x1(−x1 + x2 + x3 − x4 + x5 + x6) x ∈ [4.00, 6.36]6 , e ∈ [−ǫ, ǫ]15 , ǫ = 2−53 Dense SDP: (6+15+4

6+15 ) = 12650 variables ❀ Out of memory

Sparse SDP Real2Float tool: 15(6+1+4

6+1 ) = 4950 ❀ 759ǫ

Victor Magron The quest of efficiency and certification in polynomial optimization 12 / 42

slide-26
SLIDE 26

Preliminary Comparisons

f (x) := x2x5 + x3x6 − x2x3 − x5x6 + x1(−x1 + x2 + x3 − x4 + x5 + x6) x ∈ [4.00, 6.36]6 , e ∈ [−ǫ, ǫ]15 , ǫ = 2−53 Dense SDP: (6+15+4

6+15 ) = 12650 variables ❀ Out of memory

Sparse SDP Real2Float tool: 15(6+1+4

6+1 ) = 4950 ❀ 759ǫ

Interval arithmetic: 922ǫ (10 × less CPU)

Victor Magron The quest of efficiency and certification in polynomial optimization 12 / 42

slide-27
SLIDE 27

Preliminary Comparisons

f (x) := x2x5 + x3x6 − x2x3 − x5x6 + x1(−x1 + x2 + x3 − x4 + x5 + x6) x ∈ [4.00, 6.36]6 , e ∈ [−ǫ, ǫ]15 , ǫ = 2−53 Dense SDP: (6+15+4

6+15 ) = 12650 variables ❀ Out of memory

Sparse SDP Real2Float tool: 15(6+1+4

6+1 ) = 4950 ❀ 759ǫ

Interval arithmetic: 922ǫ (10 × less CPU) Symbolic Taylor FPTaylor tool: 721ǫ (21 × more CPU)

Victor Magron The quest of efficiency and certification in polynomial optimization 12 / 42

slide-28
SLIDE 28

Preliminary Comparisons

f (x) := x2x5 + x3x6 − x2x3 − x5x6 + x1(−x1 + x2 + x3 − x4 + x5 + x6) x ∈ [4.00, 6.36]6 , e ∈ [−ǫ, ǫ]15 , ǫ = 2−53 Dense SDP: (6+15+4

6+15 ) = 12650 variables ❀ Out of memory

Sparse SDP Real2Float tool: 15(6+1+4

6+1 ) = 4950 ❀ 759ǫ

Interval arithmetic: 922ǫ (10 × less CPU) Symbolic Taylor FPTaylor tool: 721ǫ (21 × more CPU) SMT-based rosa tool: 762ǫ (19 × more CPU)

Victor Magron The quest of efficiency and certification in polynomial optimization 12 / 42

slide-29
SLIDE 29

Preliminary Comparisons

R e a l 2 F l

  • a

t r

  • s

a F P T a y l

  • r

200 400 600 800 1,000 759ǫ 762ǫ 721ǫ CPU Time Error Bound (ǫ)

Victor Magron The quest of efficiency and certification in polynomial optimization 12 / 42

slide-30
SLIDE 30

Comparison with rosa

Relative bound precision Relative execution time

a b c d e f g h i j k l m

  • p

q r t u v w x y z 10 100 −10 1 −1 0.5 −0.5

Victor Magron The quest of efficiency and certification in polynomial optimization 13 / 42

slide-31
SLIDE 31

Comparison with FPTaylor

Relative bound precision Relative execution time

a b c d e f g h i jk l m n

  • p

q r t u v w x α β γ δ 10 100 −10 1 −1 0.5 −0.5

Victor Magron The quest of efficiency and certification in polynomial optimization 14 / 42

slide-32
SLIDE 32

Noncommutative (NC) Polynomials

Symmetric Matrix variables Xi, Yj

f = X1(Y1 +Y2 +Y3) + X2(Y1 +Y2 −Y3) + X3(Y1 −Y2) − X1 − 2Y1 −Y2

with X1X2 = X2X1, involution (X1Y3)⋆ = Y3X1

Victor Magron The quest of efficiency and certification in polynomial optimization 15 / 42

slide-33
SLIDE 33

Noncommutative (NC) Polynomials

Symmetric Matrix variables Xi, Yj

f = X1(Y1 +Y2 +Y3) + X2(Y1 +Y2 −Y3) + X3(Y1 −Y2) − X1 − 2Y1 −Y2

with X1X2 = X2X1, involution (X1Y3)⋆ = Y3X1

Constraints K = {(X, Y) : Xi, Yj 0, X2

i = Xi, Y2 j = Yj, XiYj = YjXi}

Victor Magron The quest of efficiency and certification in polynomial optimization 15 / 42

slide-34
SLIDE 34

Noncommutative (NC) Polynomials

Symmetric Matrix variables Xi, Yj

f = X1(Y1 +Y2 +Y3) + X2(Y1 +Y2 −Y3) + X3(Y1 −Y2) − X1 − 2Y1 −Y2

with X1X2 = X2X1, involution (X1Y3)⋆ = Y3X1

Constraints K = {(X, Y) : Xi, Yj 0, X2

i = Xi, Y2 j = Yj, XiYj = YjXi}

MINIMAL EIGENVALUE OPTIMIZATION λmin = inf { f (X, Y)v, v : (X, Y) ∈ K, v = 1}

Victor Magron The quest of efficiency and certification in polynomial optimization 15 / 42

slide-35
SLIDE 35

Noncommutative (NC) Polynomials

Symmetric Matrix variables Xi, Yj

f = X1(Y1 +Y2 +Y3) + X2(Y1 +Y2 −Y3) + X3(Y1 −Y2) − X1 − 2Y1 −Y2

with X1X2 = X2X1, involution (X1Y3)⋆ = Y3X1

Constraints K = {(X, Y) : Xi, Yj 0, X2

i = Xi, Y2 j = Yj, XiYj = YjXi}

MINIMAL EIGENVALUE OPTIMIZATION λmin = inf { f (X, Y)v, v : (X, Y) ∈ K, v = 1} = sup λ s.t. f (X, Y) − λI 0 , ∀(X, Y) ∈ K

Victor Magron The quest of efficiency and certification in polynomial optimization 15 / 42

slide-36
SLIDE 36

Putinar’s Representation

“Archimedean” constraint in K = {X : gj(X) 0}: N − ∑i X2

i 0

Theorem: NC Putinar’s representation [Helton-McCullough 02] f ≻ 0 on K = ⇒ f = ∑

i

s⋆

i si + ∑ j ∑ i

t⋆

jigjtji with si, tji ∈ RX

Victor Magron The quest of efficiency and certification in polynomial optimization 16 / 42

slide-37
SLIDE 37

Putinar’s Representation

“Archimedean” constraint in K = {X : gj(X) 0}: N − ∑i X2

i 0

Theorem: NC Putinar’s representation [Helton-McCullough 02] f ≻ 0 on K = ⇒ f = ∑

i

s⋆

i si + ∑ j ∑ i

t⋆

jigjtji with si, tji ∈ RX

NC variant of Lasserre’s Hierarchy for λmin: replace “f − λI 0 on K” by f − λI = ∑i s⋆

i si + ∑j ∑i t⋆ jigjtji

with si, tji of bounded degrees

Victor Magron The quest of efficiency and certification in polynomial optimization 16 / 42

slide-38
SLIDE 38

Sparse Putinar’s Representation

Sparse f = f1 + · · · + fp with fk ∈ RX, Ik Sparse K = {X : gj(X) 0} with gj ∈ RX, Ik(j) for some k(j) Additional constraints nk − ∑i∈Ik X2

i 0 in K

Victor Magron The quest of efficiency and certification in polynomial optimization 17 / 42

slide-39
SLIDE 39

Sparse Putinar’s Representation

Sparse f = f1 + · · · + fp with fk ∈ RX, Ik Sparse K = {X : gj(X) 0} with gj ∈ RX, Ik(j) for some k(j) Additional constraints nk − ∑i∈Ik X2

i 0 in K

RUNNING INTERSECTION PROPERTY (RIP) ∀k = 1, . . . , p − 1 Ik+1 ∩

  • jk

Ij ⊆ Ii for some i k

Victor Magron The quest of efficiency and certification in polynomial optimization 17 / 42

slide-40
SLIDE 40

Sparse Putinar’s Representation

Sparse f = f1 + · · · + fp with fk ∈ RX, Ik Sparse K = {X : gj(X) 0} with gj ∈ RX, Ik(j) for some k(j) Additional constraints nk − ∑i∈Ik X2

i 0 in K

RUNNING INTERSECTION PROPERTY (RIP) ∀k = 1, . . . , p − 1 Ik+1 ∩

  • jk

Ij ⊆ Ii for some i k Theorem: Sparse Putinar’s representation [Klep-M.-Povh 19] f ≻ 0 on K + RIP = ⇒ f = ∑

k ∑ i

s⋆

kiski + ∑ j ∑ i

tji⋆gjtji with ski ∈ RX, Ik, tji ∈ RX, Ik(j)

Victor Magron The quest of efficiency and certification in polynomial optimization 17 / 42

slide-41
SLIDE 41

Sparse Example: I3322 Bell Inequality

Entanglement in quantum mechanics → upper bounds for violation levels of Bell inequalities

Victor Magron The quest of efficiency and certification in polynomial optimization 18 / 42

slide-42
SLIDE 42

Sparse Example: I3322 Bell Inequality

Entanglement in quantum mechanics → upper bounds for violation levels of Bell inequalities → upper bounds on λmax of f on K

Victor Magron The quest of efficiency and certification in polynomial optimization 18 / 42

slide-43
SLIDE 43

Sparse Example: I3322 Bell Inequality

Entanglement in quantum mechanics → upper bounds for violation levels of Bell inequalities → upper bounds on λmax of f on K I3322 Bell inequality

f = X1(Y1 +Y2 +Y3) + X2(Y1 +Y2 −Y3) + X3(Y1 −Y2) − X1 − 2Y1 −Y2 K = {(X, Y) : Xi, Yj 0, X2

i = Xi, Y2 j = Yj, XiYj = YjXi}

Victor Magron The quest of efficiency and certification in polynomial optimization 18 / 42

slide-44
SLIDE 44

Sparse Example: I3322 Bell Inequality

Entanglement in quantum mechanics → upper bounds for violation levels of Bell inequalities → upper bounds on λmax of f on K I3322 Bell inequality

f = X1(Y1 +Y2 +Y3) + X2(Y1 +Y2 −Y3) + X3(Y1 −Y2) − X1 − 2Y1 −Y2 K = {(X, Y) : Xi, Yj 0, X2

i = Xi, Y2 j = Yj, XiYj = YjXi}

Ik → {X1, X2, X3, Yk}

Victor Magron The quest of efficiency and certification in polynomial optimization 18 / 42

slide-45
SLIDE 45

Sparse Example: I3322 Bell Inequality

Entanglement in quantum mechanics → upper bounds for violation levels of Bell inequalities → upper bounds on λmax of f on K I3322 Bell inequality

f = X1(Y1 +Y2 +Y3) + X2(Y1 +Y2 −Y3) + X3(Y1 −Y2) − X1 − 2Y1 −Y2 K = {(X, Y) : Xi, Yj 0, X2

i = Xi, Y2 j = Yj, XiYj = YjXi}

Ik → {X1, X2, X3, Yk} level sparse dense [Pál-Vértesi 18] 2 0.2550008 0.2509397

Victor Magron The quest of efficiency and certification in polynomial optimization 18 / 42

slide-46
SLIDE 46

Sparse Example: I3322 Bell Inequality

Entanglement in quantum mechanics → upper bounds for violation levels of Bell inequalities → upper bounds on λmax of f on K I3322 Bell inequality

f = X1(Y1 +Y2 +Y3) + X2(Y1 +Y2 −Y3) + X3(Y1 −Y2) − X1 − 2Y1 −Y2 K = {(X, Y) : Xi, Yj 0, X2

i = Xi, Y2 j = Yj, XiYj = YjXi}

Ik → {X1, X2, X3, Yk} level sparse dense [Pál-Vértesi 18] 2 0.2550008 0.2509397 3 0.2511592 0.2508756

Victor Magron The quest of efficiency and certification in polynomial optimization 18 / 42

slide-47
SLIDE 47

Sparse Example: I3322 Bell Inequality

Entanglement in quantum mechanics → upper bounds for violation levels of Bell inequalities → upper bounds on λmax of f on K I3322 Bell inequality

f = X1(Y1 +Y2 +Y3) + X2(Y1 +Y2 −Y3) + X3(Y1 −Y2) − X1 − 2Y1 −Y2 K = {(X, Y) : Xi, Yj 0, X2

i = Xi, Y2 j = Yj, XiYj = YjXi}

Ik → {X1, X2, X3, Yk} level sparse dense [Pál-Vértesi 18] 2 0.2550008 0.2509397 3 0.2511592 0.2508756 3’ 0.2508754

Victor Magron The quest of efficiency and certification in polynomial optimization 18 / 42

slide-48
SLIDE 48

Sparse Example: I3322 Bell Inequality

Entanglement in quantum mechanics → upper bounds for violation levels of Bell inequalities → upper bounds on λmax of f on K I3322 Bell inequality

f = X1(Y1 +Y2 +Y3) + X2(Y1 +Y2 −Y3) + X3(Y1 −Y2) − X1 − 2Y1 −Y2 K = {(X, Y) : Xi, Yj 0, X2

i = Xi, Y2 j = Yj, XiYj = YjXi}

Ik → {X1, X2, X3, Yk} level sparse dense [Pál-Vértesi 18] 2 0.2550008 0.2509397 3 0.2511592 0.2508756 3’ 0.2508754 4 0.2508917

Victor Magron The quest of efficiency and certification in polynomial optimization 18 / 42

slide-49
SLIDE 49

Sparse Example: I3322 Bell Inequality

Entanglement in quantum mechanics → upper bounds for violation levels of Bell inequalities → upper bounds on λmax of f on K I3322 Bell inequality

f = X1(Y1 +Y2 +Y3) + X2(Y1 +Y2 −Y3) + X3(Y1 −Y2) − X1 − 2Y1 −Y2 K = {(X, Y) : Xi, Yj 0, X2

i = Xi, Y2 j = Yj, XiYj = YjXi}

Ik → {X1, X2, X3, Yk} level sparse dense [Pál-Vértesi 18] 2 0.2550008 0.2509397 3 0.2511592 0.2508756 3’ 0.2508754 4 0.2508917 5 0.2508763

Victor Magron The quest of efficiency and certification in polynomial optimization 18 / 42

slide-50
SLIDE 50

Sparse Example: I3322 Bell Inequality

Entanglement in quantum mechanics → upper bounds for violation levels of Bell inequalities → upper bounds on λmax of f on K I3322 Bell inequality

f = X1(Y1 +Y2 +Y3) + X2(Y1 +Y2 −Y3) + X3(Y1 −Y2) − X1 − 2Y1 −Y2 K = {(X, Y) : Xi, Yj 0, X2

i = Xi, Y2 j = Yj, XiYj = YjXi}

Ik → {X1, X2, X3, Yk} level sparse dense [Pál-Vértesi 18] 2 0.2550008 0.2509397 3 0.2511592 0.2508756 3’ 0.2508754 4 0.2508917 5 0.2508763 6 0.2508753977180 !!!!!

Victor Magron The quest of efficiency and certification in polynomial optimization 18 / 42

slide-51
SLIDE 51

Exploiting Sparsity Certified Polynomial Optimization

slide-52
SLIDE 52

Certified Polynomial Optimization

X = (X1, . . . , Xn) co-NP hard problem: check f 0 on K f ∈ Q[X]

NP hard problem: min{ f (x) : x ∈ K}

Victor Magron The quest of efficiency and certification in polynomial optimization 19 / 42

slide-53
SLIDE 53

Certified Polynomial Optimization

X = (X1, . . . , Xn) co-NP hard problem: check f 0 on K f ∈ Q[X]

NP hard problem: min{ f (x) : x ∈ K}

1 Unconstrained K = Rn 2 Constrained

K = {x ∈ Rn : g1(x) 0, . . . , gl(x) 0} gj ∈ Q[X]

deg f, deg gj d

Victor Magron The quest of efficiency and certification in polynomial optimization 19 / 42

slide-54
SLIDE 54

Certified Polynomial Optimization

X = (X1, . . . , Xn) co-NP hard problem: check f 0 on K f ∈ Q[X]

NP hard problem: min{ f (x) : x ∈ K}

1 Unconstrained K = Rn 2 Constrained

K = {x ∈ Rn : g1(x) 0, . . . , gl(x) 0} gj ∈ Q[X]

deg f, deg gj d [Collins 75] CAD doubly exp. in n poly. in d [Grigoriev-Vorobjov 88, Basu-Pollack-Roy 98] Critical points singly exponential time (l + 1) τ dO (n)

Victor Magron The quest of efficiency and certification in polynomial optimization 19 / 42

slide-55
SLIDE 55

Certified Polynomial Optimization

Sums of squares (SOS) σ = h12 + · · · + hp2

Victor Magron The quest of efficiency and certification in polynomial optimization 20 / 42

slide-56
SLIDE 56

Certified Polynomial Optimization

Sums of squares (SOS) σ = h12 + · · · + hp2 HILBERT 17TH PROBLEM: f SOS of rational functions? [Artin 27] YES!

Victor Magron The quest of efficiency and certification in polynomial optimization 20 / 42

slide-57
SLIDE 57

Certified Polynomial Optimization

Sums of squares (SOS) σ = h12 + · · · + hp2 HILBERT 17TH PROBLEM: f SOS of rational functions? [Artin 27] YES! [Lasserre/Parrilo 01] Numerical solvers compute σ Semidefinite programming (SDP) approximate certificates

Victor Magron The quest of efficiency and certification in polynomial optimization 20 / 42

slide-58
SLIDE 58

Certified Polynomial Optimization

Sums of squares (SOS) σ = h12 + · · · + hp2 HILBERT 17TH PROBLEM: f SOS of rational functions? [Artin 27] YES! [Lasserre/Parrilo 01] Numerical solvers compute σ Semidefinite programming (SDP) approximate certificates ≃ → = The Question of Exact Certification How to go from approximate to exact certification?

Victor Magron The quest of efficiency and certification in polynomial optimization 20 / 42

slide-59
SLIDE 59

Motivation

Positivity certificates Stability proofs of critical control systems (Lyapunov) Certified function evaluation [Chevillard et. al 11] Formal verification of real inequalities [Hales et. al 15]: COQ HOL-LIGHT

Victor Magron The quest of efficiency and certification in polynomial optimization 21 / 42

slide-60
SLIDE 60

Decomposing Nonnegative Polynomials

1 Polya’s representation

f =

σ (X2

1+···+X2 n)D

positive definite form f [Reznick 95]

2 Hilbert-Artin’s representation

f = σ

h2

f 0 [Artin 27]

3 Putinar’s representation

f = σ0 + σ1 g1 + · · · + σl gl f > 0 on compact K deg σi 2D [Putinar 93]

Victor Magron The quest of efficiency and certification in polynomial optimization 22 / 42

slide-61
SLIDE 61

Decomposing Nonnegative Polynomials

Deciding polynomial nonnegativity f (a, b) = a2 − 2ab + b2 0 f (a, b) =

  • a

b z1 z2 z2 z3

  • a

b

  • a2 − 2ab + b2 = z1a2 + 2z2ab + z3b2

(A z = d)

z1 z2 z2 z3

  • =

1

  • F1

z1 + 1 1

  • F2

z2 + 1

  • F3

z3

  • F0

Victor Magron The quest of efficiency and certification in polynomial optimization 23 / 42

slide-62
SLIDE 62

Decomposing Nonnegative Polynomials

Choose a cost c e.g. (1, 0, 1) and solve SDP min

z

c

⊤z

s.t.

i

Fi zi F0 , A z = d Solution

z1 z2 z2 z3

  • =

1 −1 −1 1

  • (eigenvalues 0 and 2)

a2 − 2ab + b2 =

  • a

b 1 −1 −1 1

  • a

b

  • = (a − b)2

Solving SDP = ⇒ Finding SUMS OF SQUARES certificates

Victor Magron The quest of efficiency and certification in polynomial optimization 24 / 42

slide-63
SLIDE 63

From Approximate to Exact Solutions

APPROXIMATE SOLUTIONS sum of squares of a2 − 2ab + b2? (1.00001a − 0.99998b)2! a2 − 2ab + b2 ≃ (1.00001a − 0.99998b)2 a2 − 2ab + b2 = 1.0000200001a2 − 1.9999799996ab + 0.9999600004b2 ≃ → = ?

Victor Magron The quest of efficiency and certification in polynomial optimization 25 / 42

slide-64
SLIDE 64

Rational SOS Decompositions

Let f ∈ R[X] and f 0 on R (n = 1) Theorem There exist f1, f2 ∈ R[X] s.t. f = f12 + f22.

Victor Magron The quest of efficiency and certification in polynomial optimization 26 / 42

slide-65
SLIDE 65

Rational SOS Decompositions

Let f ∈ R[X] and f 0 on R (n = 1) Theorem There exist f1, f2 ∈ R[X] s.t. f = f12 + f22.

Proof.

f = h2(q + ir)(q − ir)

Victor Magron The quest of efficiency and certification in polynomial optimization 26 / 42

slide-66
SLIDE 66

Rational SOS Decompositions

Let f ∈ R[X] and f 0 on R (n = 1) Theorem There exist f1, f2 ∈ R[X] s.t. f = f12 + f22.

Proof.

f = h2(q + ir)(q − ir) Examples

1 + X + X2 =

  • X + 1

2 2 + √ 3 2 2 1 + X + X2 + X3 + X4 =

  • X2 + 1

2 X + 1 + √ 5 4 2 + 10 + 2 √ 5 +

  • 10 − 2

√ 5 4 X +

  • 10 − 2

√ 5 4 2

Victor Magron The quest of efficiency and certification in polynomial optimization 26 / 42

slide-67
SLIDE 67

Rational SOS Decompositions

f ∈ Q[X] ∩ ˚ Σ[X] (interior of the SOS cone) Existence Question Does there exist fi ∈ Q[X], ci ∈ Q>0 s.t. f = ∑i ci fi2?

Victor Magron The quest of efficiency and certification in polynomial optimization 27 / 42

slide-68
SLIDE 68

Rational SOS Decompositions

f ∈ Q[X] ∩ ˚ Σ[X] (interior of the SOS cone) Existence Question Does there exist fi ∈ Q[X], ci ∈ Q>0 s.t. f = ∑i ci fi2? Examples

1 + X + X2 =

  • X + 1

2 2 + √ 3 2 2 = 1

  • X + 1

2 2 + 3 4 (1)2 1 + X + X2 + X3 + X4 =

  • X2 + 1

2 X + 1 + √ 5 4 2 + 10 + 2 √ 5 +

  • 10 − 2

√ 5 4 X +

  • 10 − 2

√ 5 4 2 = ???

Victor Magron The quest of efficiency and certification in polynomial optimization 27 / 42

slide-69
SLIDE 69

Round & Project Algorithm [Peyrl-Parrilo 08]

Σ f

f ∈ ˚ Σ[X] with deg f = 2D

Victor Magron The quest of efficiency and certification in polynomial optimization 28 / 42

slide-70
SLIDE 70

Round & Project Algorithm [Peyrl-Parrilo 08]

Σ f

f ∈ ˚ Σ[X] with deg f = 2D Find ˜ G with SDP at tolerance ˜ δ satisfying f (X) ≃ vDT(X) ˜ G vD(X) ˜ G ≻ 0 vD(X): vector of monomials of deg D

Victor Magron The quest of efficiency and certification in polynomial optimization 28 / 42

slide-71
SLIDE 71

Round & Project Algorithm [Peyrl-Parrilo 08]

Σ f

f ∈ ˚ Σ[X] with deg f = 2D Find ˜ G with SDP at tolerance ˜ δ satisfying f (X) ≃ vDT(X) ˜ G vD(X) ˜ G ≻ 0 vD(X): vector of monomials of deg D Exact G = ⇒ fγ = ∑α′+β′=γ Gα′,β′

Victor Magron The quest of efficiency and certification in polynomial optimization 28 / 42

slide-72
SLIDE 72

Round & Project Algorithm [Peyrl-Parrilo 08]

Σ f

f ∈ ˚ Σ[X] with deg f = 2D Find ˜ G with SDP at tolerance ˜ δ satisfying f (X) ≃ vDT(X) ˜ G vD(X) ˜ G ≻ 0 vD(X): vector of monomials of deg D Exact G = ⇒ fγ = ∑α′+β′=γ Gα′,β′ fα+β = ∑α′+β′=α+β Gα′,β′

Victor Magron The quest of efficiency and certification in polynomial optimization 28 / 42

slide-73
SLIDE 73

Round & Project Algorithm [Peyrl-Parrilo 08]

Σ f

f ∈ ˚ Σ[X] with deg f = 2D Find ˜ G with SDP at tolerance ˜ δ satisfying f (X) ≃ vDT(X) ˜ G vD(X) ˜ G ≻ 0 vD(X): vector of monomials of deg D Exact G = ⇒ fγ = ∑α′+β′=γ Gα′,β′ fα+β = ∑α′+β′=α+β Gα′,β′

1 Rounding step ˆ

G ← round ˜ G, ˆ δ

  • Victor Magron

The quest of efficiency and certification in polynomial optimization 28 / 42

slide-74
SLIDE 74

Round & Project Algorithm [Peyrl-Parrilo 08]

Σ f

f ∈ ˚ Σ[X] with deg f = 2D Find ˜ G with SDP at tolerance ˜ δ satisfying f (X) ≃ vDT(X) ˜ G vD(X) ˜ G ≻ 0 vD(X): vector of monomials of deg D Exact G = ⇒ fγ = ∑α′+β′=γ Gα′,β′ fα+β = ∑α′+β′=α+β Gα′,β′

1 Rounding step ˆ

G ← round ˜ G, ˆ δ

  • 2 Projection step

Gα,β ← ˆ Gα,β −

1 η(α+β)(∑α′+β′=α+β ˆ

Gα′,β′ − fα+β)

Victor Magron The quest of efficiency and certification in polynomial optimization 28 / 42

slide-75
SLIDE 75

Round & Project Algorithm [Peyrl-Parrilo 08]

Σ f

f ∈ ˚ Σ[X] with deg f = 2D Find ˜ G with SDP at tolerance ˜ δ satisfying f (X) ≃ vDT(X) ˜ G vD(X) ˜ G ≻ 0 vD(X): vector of monomials of deg D Exact G = ⇒ fγ = ∑α′+β′=γ Gα′,β′ fα+β = ∑α′+β′=α+β Gα′,β′

1 Rounding step ˆ

G ← round ˜ G, ˆ δ

  • 2 Projection step

Gα,β ← ˆ Gα,β −

1 η(α+β)(∑α′+β′=α+β ˆ

Gα′,β′ − fα+β) Small enough ˜ δ, ˆ δ = ⇒ f (X) = vDT(X) G vD(X) and G 0

Victor Magron The quest of efficiency and certification in polynomial optimization 28 / 42

slide-76
SLIDE 76

One Answer when K = {x ∈ Rn : gj(x) 0}

Hybrid SYMBOLIC/NUMERIC methods Magron-Allamigeon-Gaubert-Werner 14 f ≃ ˜ σ0 + ˜ σ1 g1 + · · · + ˜ σm gm u = f − ˜ σ0 + ˜ σ1 g1 + · · · + ˜ σm gm

Victor Magron The quest of efficiency and certification in polynomial optimization 29 / 42

slide-77
SLIDE 77

One Answer when K = {x ∈ Rn : gj(x) 0}

Hybrid SYMBOLIC/NUMERIC methods Magron-Allamigeon-Gaubert-Werner 14 f ≃ ˜ σ0 + ˜ σ1 g1 + · · · + ˜ σm gm u = f − ˜ σ0 + ˜ σ1 g1 + · · · + ˜ σm gm ≃ → = ∀x ∈ [0, 1]n, u(x) −ε minK f ε when ε → 0 COMPLEXITY? Compact K ⊆ [0, 1]n

Victor Magron The quest of efficiency and certification in polynomial optimization 29 / 42

slide-78
SLIDE 78

From Approximate to Exact Solutions

Win TWO-PLAYER GAME

Σ f

sum of squares of f ? ≃ Output!

Victor Magron The quest of efficiency and certification in polynomial optimization 30 / 42

slide-79
SLIDE 79

From Approximate to Exact Solutions

Win TWO-PLAYER GAME

Σ f

Hybrid Symbolic/Numeric Algorithms sum of squares of f − ε? ≃ Output! Error Compensation ≃ → =

Victor Magron The quest of efficiency and certification in polynomial optimization 30 / 42

slide-80
SLIDE 80

From Approximate to Exact Solutions

Exact SOS Exact SONC/SAGE

Σ f CSONC f CSAGE f

Victor Magron The quest of efficiency and certification in polynomial optimization 31 / 42

slide-81
SLIDE 81

Software: RealCertify and POEM

Exact optimization via SOS: RealCertify Maple & arbitrary precision SDP solver SDPA-GMP [Nakata 10] univsos n = 1 multivsos n > 1 Exact optimization via SONC/SAGE: POEM Python (SymPy) & geometric programming/relative entropy ECOS [Domahidi-Chu-Boyd 13]

Victor Magron The quest of efficiency and certification in polynomial optimization 31 / 42

slide-82
SLIDE 82

intsos with n 1: Perturbation

Σ f

PERTURBATION idea Approximate SOS Decomposition f (X) - ε ∑α∈P/2 X2α = ˜ σ + u

Victor Magron The quest of efficiency and certification in polynomial optimization 32 / 42

slide-83
SLIDE 83

intsos with n = 1 [Chevillard et. al 11]

p ∈ Q[X], deg p = d = 2k, p > 0

x p p = 1 + X + X2 + X3 + X4

Victor Magron The quest of efficiency and certification in polynomial optimization 33 / 42

slide-84
SLIDE 84

intsos with n = 1 [Chevillard et. al 11]

p ∈ Q[X], deg p = d = 2k, p > 0 PERTURB: find ε ∈ Q s.t. pε := p − ε

k

i=0

X2i > 0

x p

1 4(1 + x2 + x4)

pε p = 1 + X + X2 + X3 + X4 ε = 1 4 p > 1 4 (1 + X2 + X4)

Victor Magron The quest of efficiency and certification in polynomial optimization 33 / 42

slide-85
SLIDE 85

intsos with n = 1 [Chevillard et. al 11]

p ∈ Q[X], deg p = d = 2k, p > 0 PERTURB: find ε ∈ Q s.t. pε := p − ε

k

i=0

X2i > 0 SDP Approximation: p − ε

k

i=0

X2i = ˜ σ + u ABSORB: small enough ui = ⇒ ε ∑k

i=0 X2i + u SOS x p

1 4(1 + x2 + x4)

pε p = 1 + X + X2 + X3 + X4 ε = 1 4 p > 1 4 (1 + X2 + X4)

Victor Magron The quest of efficiency and certification in polynomial optimization 33 / 42

slide-86
SLIDE 86

intsos with n = 1 and SDP Approximation

Input f 0 ∈ Q[X] of degree d 2, ε ∈ Q>0, δ ∈ N>0 Output: SOS decomposition with coefficients in Q

pε ←p − ε

k

i=0

X2i ε ← ε 2 ˜ σ ←sdp(pε, δ) u ←pε − ˜ σ δ ←2δ (p, h) ← sqrfree( f ) f h, ˜ σ, ε, u while pε ≤ 0 while ε < |u2i+1| + |u2i−1| 2 − u2i

Victor Magron The quest of efficiency and certification in polynomial optimization 34 / 42

slide-87
SLIDE 87

intsos with n = 1: Absorbtion

X = 1

2

(X + 1)2 − 1 − X2 −X = 1

2

(X − 1)2 − 1 − X2

Victor Magron The quest of efficiency and certification in polynomial optimization 35 / 42

slide-88
SLIDE 88

intsos with n = 1: Absorbtion

X = 1

2

(X + 1)2 − 1 − X2 −X = 1

2

(X − 1)2 − 1 − X2 u2i+1X2i+1 = |u2i+1| 2 (Xi+1 + sgn (u2i+1)Xi)2 − X2i − X2i+2

Victor Magron The quest of efficiency and certification in polynomial optimization 35 / 42

slide-89
SLIDE 89

intsos with n = 1: Absorbtion

X = 1

2

(X + 1)2 − 1 − X2 −X = 1

2

(X − 1)2 − 1 − X2 u2i+1X2i+1 = |u2i+1| 2 (Xi+1 + sgn (u2i+1)Xi)2 − X2i − X2i+2

u ε ∑k

i=0 X2i

· · · 2i − 2 2i − 1 2i 2i + 1 2i + 2 · · · ε ε ε

Victor Magron The quest of efficiency and certification in polynomial optimization 35 / 42

slide-90
SLIDE 90

intsos with n = 1: Absorbtion

X = 1

2

(X + 1)2 − 1 − X2 −X = 1

2

(X − 1)2 − 1 − X2 u2i+1X2i+1 = |u2i+1| 2 (Xi+1 + sgn (u2i+1)Xi)2 − X2i − X2i+2

u ε ∑k

i=0 X2i

· · · 2i − 2 2i − 1 2i 2i + 1 2i + 2 · · · ε ε ε

ε |u2i+1| + |u2i−1| 2 − u2i = ⇒ ε

k

i=0

X2i + u SOS

Victor Magron The quest of efficiency and certification in polynomial optimization 35 / 42

slide-91
SLIDE 91

intsos with n 1: Absorbtion f (X) - ε ∑α∈P/2 X2α = ˜ σ + u Choice of P?

x y 1 2 3 4 5 1 2 3 4 5 6 u1,3 ε ε xy3 = 1

2(x + y3)2 − x2+y6 2

Victor Magron The quest of efficiency and certification in polynomial optimization 36 / 42

slide-92
SLIDE 92

intsos with n 1: Absorbtion f (X) - ε ∑α∈P/2 X2α = ˜ σ + u Choice of P?

x y 1 2 3 4 5 1 2 3 4 5 6 u1,3 ε ε xy3 = 1

2(xy + y2)2 − x2y2+y4 2

Victor Magron The quest of efficiency and certification in polynomial optimization 36 / 42

slide-93
SLIDE 93

intsos with n 1: Absorbtion f (X) - ε ∑α∈P/2 X2α = ˜ σ + u Choice of P?

x y 1 2 3 4 5 1 2 3 4 5 6 u1,3 ε ε xy3 = 1

2(xy2 + y)2 − x2y4+y2 2

Victor Magron The quest of efficiency and certification in polynomial optimization 36 / 42

slide-94
SLIDE 94

intsos with n 1: Absorbtion f (X) - ε ∑α∈P/2 X2α = ˜ σ + u Choice of P?

f = 4x4y6 + x2 − xy2 + y2 spt( f ) = {(4, 6), (2, 0), (1, 2), (0, 2)} Newton Polytope P = conv (spt( f )) Squares in SOS decomposition ⊆ P

2 ∩ Nn

[Reznick 78]

Victor Magron The quest of efficiency and certification in polynomial optimization 36 / 42

slide-95
SLIDE 95

Algorithm intsos

Input f ∈ Q[X] ∩ ˚ Σ[X] of degree d, ε ∈ Q>0, δ ∈ N>0 Output: SOS decomposition with coefficients in Q

fε ← f − ε ∑

α∈P/2

X2α ε ← ε 2 ˜ σ ←sdp( fε, δ) u ← fε − ˜ σ δ ←2δ P ← conv (spt( f )) f h, ˜ σ, ε, u while fε ≤ 0 while u + ε ∑

α∈P/2

X2α / ∈ Σ

Victor Magron The quest of efficiency and certification in polynomial optimization 37 / 42

slide-96
SLIDE 96

Algorithm intsos

Theorem (Exact Certification Cost in ˚ Σ) f ∈ Q[X] ∩ ˚ Σ[X] with deg f = d = 2k and bit size τ = ⇒ intsos terminates with SOS output of bit size τ dO (n)

Victor Magron The quest of efficiency and certification in polynomial optimization 38 / 42

slide-97
SLIDE 97

Algorithm intsos

Theorem (Exact Certification Cost in ˚ Σ) f ∈ Q[X] ∩ ˚ Σ[X] with deg f = d = 2k and bit size τ = ⇒ intsos terminates with SOS output of bit size τ dO (n)

Victor Magron The quest of efficiency and certification in polynomial optimization 38 / 42

slide-98
SLIDE 98

Algorithm Polyasos

f positive definite form has Polya’s representation: f = σ (X1 + · · · + Xn)2D with σ ∈ Σ[X]

Victor Magron The quest of efficiency and certification in polynomial optimization 39 / 42

slide-99
SLIDE 99

Algorithm Polyasos

f positive definite form has Polya’s representation: f = σ (X1 + · · · + Xn)2D with σ ∈ Σ[X] Theorem f (X1 + · · · + Xn)2D ∈ Σ[X] = ⇒ f (X1 + · · · + Xn)2D+2 ∈ ˚ Σ[X]

Victor Magron The quest of efficiency and certification in polynomial optimization 39 / 42

slide-100
SLIDE 100

Algorithm Polyasos

f positive definite form has Polya’s representation: f = σ (X1 + · · · + Xn)2D with σ ∈ Σ[X] Theorem f (X1 + · · · + Xn)2D ∈ Σ[X] = ⇒ f (X1 + · · · + Xn)2D+2 ∈ ˚ Σ[X] Apply Algorithm intsos on f (X1 + · · · + Xn)2D+2

Victor Magron The quest of efficiency and certification in polynomial optimization 39 / 42

slide-101
SLIDE 101

Algorithm Polyasos

f positive definite form has Polya’s representation: f = σ (X1 + · · · + Xn)2D with σ ∈ Σ[X] Theorem f (X1 + · · · + Xn)2D ∈ Σ[X] = ⇒ f (X1 + · · · + Xn)2D+2 ∈ ˚ Σ[X] Apply Algorithm intsos on f (X1 + · · · + Xn)2D+2 Theorem (Exact Certification Cost of Polya’s representations) f ∈ Q[X] positive definite form with deg f = d and bit size τ = ⇒ D 2τ dO (n)

OUTPUT BIT SIZE = τ DO (n)

Victor Magron The quest of efficiency and certification in polynomial optimization 39 / 42

slide-102
SLIDE 102

Algorithm Putinarsos

Assumption: ∃i s.t. gi = 1 − X2

2

f > 0 on K := {x : gj(x) 0} has Putinar’s representation: f = σ0 + ∑

j

σj gj with σj ∈ Σ[X] , deg σj 2D

Victor Magron The quest of efficiency and certification in polynomial optimization 40 / 42

slide-103
SLIDE 103

Algorithm Putinarsos

Assumption: ∃i s.t. gi = 1 − X2

2

f > 0 on K := {x : gj(x) 0} has Putinar’s representation: f = σ0 + ∑

j

σj gj with σj ∈ Σ[X] , deg σj 2D Theorem [M.-Safey El Din 18] f = ˚ σ0 + ∑

j

˚ σj gj with ˚ σj ∈ ˚ Σ[X] , deg ˚ σj 2D

Victor Magron The quest of efficiency and certification in polynomial optimization 40 / 42

slide-104
SLIDE 104

Algorithm Putinarsos

Assumption: ∃i s.t. gi = 1 − X2

2

f > 0 on K := {x : gj(x) 0} has Putinar’s representation: f = σ0 + ∑

j

σj gj with σj ∈ Σ[X] , deg σj 2D Theorem [M.-Safey El Din 18] f = ˚ σ0 + ∑

j

˚ σj gj with ˚ σj ∈ ˚ Σ[X] , deg ˚ σj 2D ABSORBTION as in Algorithm intsos: u = fε − ˜ σ0 − ∑j ˜ σj gj

Victor Magron The quest of efficiency and certification in polynomial optimization 40 / 42

slide-105
SLIDE 105

Algorithm Putinarsos

Assumption: ∃i s.t. gi = 1 − X2

2

f > 0 on K := {x : gj(x) 0} has Putinar’s representation: f = σ0 + ∑

j

σj gj with σj ∈ Σ[X] , deg σj 2D Theorem [M.-Safey El Din 18] f = ˚ σ0 + ∑

j

˚ σj gj with ˚ σj ∈ ˚ Σ[X] , deg ˚ σj 2D ABSORBTION as in Algorithm intsos: u = fε − ˜ σ0 − ∑j ˜ σj gj OUTPUT BIT SIZE = τ DO (n)

Victor Magron The quest of efficiency and certification in polynomial optimization 40 / 42

slide-106
SLIDE 106

SOS Benchmarks

Round & Project (SOS) [Peyrl-Parrilo] RAGLib (critical points) [Safey El Din] SamplePoints (CAD) [Moreno Maza-Alvandi et al.]

n d RealCertify RoundProject RAGLib CAD τ1 (bits) t1 (s) τ2 (bits) t2 (s) t3 (s) t4 (s) 2 20 745 419 110. 78 949 497 141. 0.16 0.03 3 8 17 232 0.35 18 831 0.29 0.15 0.03 2 4 1 866 0.03 1 031 0.04 0.09 0.01 6 4 56 890 0.34 475 359 0.54 598. − 10 4 344 347 2.45 8 374 082 4.59 − −

Victor Magron The quest of efficiency and certification in polynomial optimization 41 / 42

slide-107
SLIDE 107

Conclusion and Perspectives

1 Quest for efficiency

Commutative: Roundoff error n ≃ 102 Noncommutative: Minimal eigenvalue n ≃ 20 − 30

Victor Magron The quest of efficiency and certification in polynomial optimization 42 / 42

slide-108
SLIDE 108

Conclusion and Perspectives

1 Quest for efficiency

Commutative: Roundoff error n ≃ 102 Noncommutative: Minimal eigenvalue n ≃ 20 − 30

2 Quest for certification

Input f of deg d & bitsize τ ⇒ Output SOS of bitsize τdO (n)

Victor Magron The quest of efficiency and certification in polynomial optimization 42 / 42

slide-109
SLIDE 109

Conclusion and Perspectives

1 Quest for efficiency

Commutative: Roundoff error n ≃ 102 Noncommutative: Minimal eigenvalue n ≃ 20 − 30

2 Quest for certification

Input f of deg d & bitsize τ ⇒ Output SOS of bitsize τdO (n)

Symmetric noncommutative problems? Certification of minimal eigenvalues Bell inequalities

Victor Magron The quest of efficiency and certification in polynomial optimization 42 / 42

slide-110
SLIDE 110

Conclusion and Perspectives

1 Quest for efficiency

Commutative: Roundoff error n ≃ 102 Noncommutative: Minimal eigenvalue n ≃ 20 − 30

2 Quest for certification

Input f of deg d & bitsize τ ⇒ Output SOS of bitsize τdO (n)

Symmetric noncommutative problems? Certification of minimal eigenvalues Bell inequalities APPLICATIONS IN QUANTUM PHYSICS

Quantum games: number of mutually unbiased bases in dim 6,

OPEN FOR SEVERAL DECADES!!

symmetric Ground state energy of hamiltonians symmetric & sparse Inflation for quantum correlations symmetric & sparse

Victor Magron The quest of efficiency and certification in polynomial optimization 42 / 42

slide-111
SLIDE 111

Thank you for your attention!

https://homepages.laas.fr/vmagron

Klep, M. & Povh. Sparse Noncommutative Polynomial Optimization, arxiv:1909.00569 NCSOStools M., Constantinides & Donaldson. Certified Roundoff Error Bounds Using Semidefinite Programming, TOMS. arxiv:1507.03331 Real2Float M., Safey El Din & Schweighofer. Algorithms for Weighted Sums of Squares Decomposition of Non-negative Univariate Polynomials, JSC. arxiv:1706.03941 RealCertify

  • M. & Safey El Din. On Exact Polya and Putinar’s Representations,

ISSAC’18. arxiv:1802.10339 RealCertify

  • M. & Safey El Din. RealCertify: a Maple package for certifying

non-negativity, ISSAC’18. arxiv:1805.02201 RealCertify M., Seidler & de Wolff. Exact optimization via sums of nonnegative circuits and AM/GM exponentials, ISSAC’19. arxiv:1902.02123 POEM