Low-Degree Hardness of Random Optimization Problems Alex Wein - - PowerPoint PPT Presentation

low degree hardness of random optimization problems alex
SMART_READER_LITE
LIVE PREVIEW

Low-Degree Hardness of Random Optimization Problems Alex Wein - - PowerPoint PPT Presentation

Low-Degree Hardness of Random Optimization Problems Alex Wein Courant Institute, New York University Joint work with: David Gamarnik Aukosh Jagannath MIT Waterloo 1 / 18 Random Optimization Problems Examples: 2 / 18 Random Optimization


slide-1
SLIDE 1

Low-Degree Hardness of Random Optimization Problems Alex Wein

Courant Institute, New York University Joint work with: David Gamarnik

MIT

Aukosh Jagannath

Waterloo

1 / 18

slide-2
SLIDE 2

Random Optimization Problems

Examples:

2 / 18

slide-3
SLIDE 3

Random Optimization Problems

Examples: ◮ Max clique in a random graph

2 / 18

slide-4
SLIDE 4

Random Optimization Problems

Examples: ◮ Max clique in a random graph ◮ Max-k-SAT on a random formula

2 / 18

slide-5
SLIDE 5

Random Optimization Problems

Examples: ◮ Max clique in a random graph ◮ Max-k-SAT on a random formula ◮ Maximizing a random degree-p polynomial over the sphere

2 / 18

slide-6
SLIDE 6

Random Optimization Problems

Examples: ◮ Max clique in a random graph ◮ Max-k-SAT on a random formula ◮ Maximizing a random degree-p polynomial over the sphere Note: no planted solution

2 / 18

slide-7
SLIDE 7

Random Optimization Problems

Examples: ◮ Max clique in a random graph ◮ Max-k-SAT on a random formula ◮ Maximizing a random degree-p polynomial over the sphere Note: no planted solution Q: What is the typical value of the optimum (OPT)?

2 / 18

slide-8
SLIDE 8

Random Optimization Problems

Examples: ◮ Max clique in a random graph ◮ Max-k-SAT on a random formula ◮ Maximizing a random degree-p polynomial over the sphere Note: no planted solution Q: What is the typical value of the optimum (OPT)? Q: What objective value can be reached algorithmically (ALG)?

2 / 18

slide-9
SLIDE 9

Random Optimization Problems

Examples: ◮ Max clique in a random graph ◮ Max-k-SAT on a random formula ◮ Maximizing a random degree-p polynomial over the sphere Note: no planted solution Q: What is the typical value of the optimum (OPT)? Q: What objective value can be reached algorithmically (ALG)? Q: In cases where it seems hard to reach a particular objective value, can we understand why?

2 / 18

slide-10
SLIDE 10

Random Optimization Problems

Examples: ◮ Max clique in a random graph ◮ Max-k-SAT on a random formula ◮ Maximizing a random degree-p polynomial over the sphere Note: no planted solution Q: What is the typical value of the optimum (OPT)? Q: What objective value can be reached algorithmically (ALG)? Q: In cases where it seems hard to reach a particular objective value, can we understand why? In a unified way?

2 / 18

slide-11
SLIDE 11

Max Independent Set

Example (max independent set): given sparse graph G(n, d/n), max

S⊆[n] |S|

s.t. S independent

3 / 18

slide-12
SLIDE 12

Max Independent Set

Example (max independent set): given sparse graph G(n, d/n), max

S⊆[n] |S|

s.t. S independent OPT = 2 log d d n

[Frieze ’90]

3 / 18

slide-13
SLIDE 13

Max Independent Set

Example (max independent set): given sparse graph G(n, d/n), max

S⊆[n] |S|

s.t. S independent OPT = 2 log d d n ALG = log d d n

[Frieze ’90]

3 / 18

slide-14
SLIDE 14

Max Independent Set

Example (max independent set): given sparse graph G(n, d/n), max

S⊆[n] |S|

s.t. S independent OPT = 2 log d d n ALG = log d d n

[Frieze ’90]

[Karp ’76]: Is there a better algorithm?

3 / 18

slide-15
SLIDE 15

Max Independent Set

Example (max independent set): given sparse graph G(n, d/n), max

S⊆[n] |S|

s.t. S independent OPT = 2 log d d n ALG = log d d n

[Frieze ’90]

[Karp ’76]: Is there a better algorithm?

Structural evidence suggests no!

[Achlioptas, Coja-Oghlan ’08; Coja-Oghlan, Efthymiou ’10]

3 / 18

slide-16
SLIDE 16

Max Independent Set

Example (max independent set): given sparse graph G(n, d/n), max

S⊆[n] |S|

s.t. S independent OPT = 2 log d d n ALG = log d d n

[Frieze ’90]

[Karp ’76]: Is there a better algorithm?

Structural evidence suggests no!

[Achlioptas, Coja-Oghlan ’08; Coja-Oghlan, Efthymiou ’10]

Local algorithms achieve value ALG and no better

[Gamarnik, Sudan ’13; Rahman, Vir´ ag ’14]

3 / 18

slide-17
SLIDE 17

Spherical Spin Glass

Example (spherical p-spin model): for Y ∈ R⊗p i.i.d. N(0, 1), max

v=1

1 √nY , v⊗p

4 / 18

slide-18
SLIDE 18

Spherical Spin Glass

Example (spherical p-spin model): for Y ∈ R⊗p i.i.d. N(0, 1), max

v=1

1 √nY , v⊗p OPT = Θ(1)

[Auffinger, Ben Arous, ˇ Cern´ y ’13]

4 / 18

slide-19
SLIDE 19

Spherical Spin Glass

Example (spherical p-spin model): for Y ∈ R⊗p i.i.d. N(0, 1), max

v=1

1 √nY , v⊗p OPT = Θ(1)

[Auffinger, Ben Arous, ˇ Cern´ y ’13]

ALG = Θ(1)

[Subag ’18]

4 / 18

slide-20
SLIDE 20

Spherical Spin Glass

Example (spherical p-spin model): for Y ∈ R⊗p i.i.d. N(0, 1), max

v=1

1 √nY , v⊗p OPT = Θ(1)

[Auffinger, Ben Arous, ˇ Cern´ y ’13]

ALG = Θ(1)

[Subag ’18]

ALG < OPT (for p ≥ 3)

4 / 18

slide-21
SLIDE 21

Spherical Spin Glass

Example (spherical p-spin model): for Y ∈ R⊗p i.i.d. N(0, 1), max

v=1

1 √nY , v⊗p OPT = Θ(1)

[Auffinger, Ben Arous, ˇ Cern´ y ’13]

ALG = Θ(1)

[Subag ’18]

ALG < OPT (for p ≥ 3) Approximate message passing (AMP) algorithms achieve value ALG and no better

[El Alaoui, Montanari, Sellke ’20]

4 / 18

slide-22
SLIDE 22

What’s Missing?

How to give the best “evidence” that there are no better algorithms?

5 / 18

slide-23
SLIDE 23

What’s Missing?

How to give the best “evidence” that there are no better algorithms? Prior work rules out certain classes of algorithms (local, AMP), but do we expect these to be optimal?

5 / 18

slide-24
SLIDE 24

What’s Missing?

How to give the best “evidence” that there are no better algorithms? Prior work rules out certain classes of algorithms (local, AMP), but do we expect these to be optimal? ◮ AMP is not optimal for tensor PCA

[Montanari, Richard ’14]

5 / 18

slide-25
SLIDE 25

What’s Missing?

How to give the best “evidence” that there are no better algorithms? Prior work rules out certain classes of algorithms (local, AMP), but do we expect these to be optimal? ◮ AMP is not optimal for tensor PCA

[Montanari, Richard ’14]

Would like a unified framework for lower bounds

5 / 18

slide-26
SLIDE 26

What’s Missing?

How to give the best “evidence” that there are no better algorithms? Prior work rules out certain classes of algorithms (local, AMP), but do we expect these to be optimal? ◮ AMP is not optimal for tensor PCA

[Montanari, Richard ’14]

Would like a unified framework for lower bounds ◮ Local algorithms only make sense on sparse graphs

5 / 18

slide-27
SLIDE 27

What’s Missing?

How to give the best “evidence” that there are no better algorithms? Prior work rules out certain classes of algorithms (local, AMP), but do we expect these to be optimal? ◮ AMP is not optimal for tensor PCA

[Montanari, Richard ’14]

Would like a unified framework for lower bounds ◮ Local algorithms only make sense on sparse graphs Solution: lower bounds against a larger class of algorithms (low-degree polynomials) that contains both local and AMP algorithms

5 / 18

slide-28
SLIDE 28

The Low-Degree Polynomial Framework

Study a restricted class of algorithms: low-degree polynomials

6 / 18

slide-29
SLIDE 29

The Low-Degree Polynomial Framework

Study a restricted class of algorithms: low-degree polynomials ◮ Multivariate polynomial f : RM → RN

6 / 18

slide-30
SLIDE 30

The Low-Degree Polynomial Framework

Study a restricted class of algorithms: low-degree polynomials ◮ Multivariate polynomial f : RM → RN

◮ Input: e.g. graph Y ∈ {0, 1}(

n 2)

6 / 18

slide-31
SLIDE 31

The Low-Degree Polynomial Framework

Study a restricted class of algorithms: low-degree polynomials ◮ Multivariate polynomial f : RM → RN

◮ Input: e.g. graph Y ∈ {0, 1}(

n 2)

◮ Output: e.g. b ∈ {0, 1}

6 / 18

slide-32
SLIDE 32

The Low-Degree Polynomial Framework

Study a restricted class of algorithms: low-degree polynomials ◮ Multivariate polynomial f : RM → RN

◮ Input: e.g. graph Y ∈ {0, 1}(

n 2)

◮ Output: e.g. b ∈ {0, 1} or v ∈ Rn

6 / 18

slide-33
SLIDE 33

The Low-Degree Polynomial Framework

Study a restricted class of algorithms: low-degree polynomials ◮ Multivariate polynomial f : RM → RN

◮ Input: e.g. graph Y ∈ {0, 1}(

n 2)

◮ Output: e.g. b ∈ {0, 1} or v ∈ Rn

◮ “Low” means O(log n) where n is dimension

6 / 18

slide-34
SLIDE 34

The Low-Degree Polynomial Framework

Study a restricted class of algorithms: low-degree polynomials ◮ Multivariate polynomial f : RM → RN

◮ Input: e.g. graph Y ∈ {0, 1}(

n 2)

◮ Output: e.g. b ∈ {0, 1} or v ∈ Rn

◮ “Low” means O(log n) where n is dimension Examples of low-degree algorithms:

6 / 18

slide-35
SLIDE 35

The Low-Degree Polynomial Framework

Study a restricted class of algorithms: low-degree polynomials ◮ Multivariate polynomial f : RM → RN

◮ Input: e.g. graph Y ∈ {0, 1}(

n 2)

◮ Output: e.g. b ∈ {0, 1} or v ∈ Rn

◮ “Low” means O(log n) where n is dimension Examples of low-degree algorithms: input Y ∈ Rn×n

6 / 18

slide-36
SLIDE 36

The Low-Degree Polynomial Framework

Study a restricted class of algorithms: low-degree polynomials ◮ Multivariate polynomial f : RM → RN

◮ Input: e.g. graph Y ∈ {0, 1}(

n 2)

◮ Output: e.g. b ∈ {0, 1} or v ∈ Rn

◮ “Low” means O(log n) where n is dimension Examples of low-degree algorithms: input Y ∈ Rn×n ◮ Power iteration: Y k1

6 / 18

slide-37
SLIDE 37

The Low-Degree Polynomial Framework

Study a restricted class of algorithms: low-degree polynomials ◮ Multivariate polynomial f : RM → RN

◮ Input: e.g. graph Y ∈ {0, 1}(

n 2)

◮ Output: e.g. b ∈ {0, 1} or v ∈ Rn

◮ “Low” means O(log n) where n is dimension Examples of low-degree algorithms: input Y ∈ Rn×n ◮ Power iteration: Y k1 ◮ Approximate message passing

6 / 18

slide-38
SLIDE 38

The Low-Degree Polynomial Framework

Study a restricted class of algorithms: low-degree polynomials ◮ Multivariate polynomial f : RM → RN

◮ Input: e.g. graph Y ∈ {0, 1}(

n 2)

◮ Output: e.g. b ∈ {0, 1} or v ∈ Rn

◮ “Low” means O(log n) where n is dimension Examples of low-degree algorithms: input Y ∈ Rn×n ◮ Power iteration: Y k1 ◮ Approximate message passing ◮ Local algorithms on sparse graphs

6 / 18

slide-39
SLIDE 39

The Low-Degree Polynomial Framework

Study a restricted class of algorithms: low-degree polynomials ◮ Multivariate polynomial f : RM → RN

◮ Input: e.g. graph Y ∈ {0, 1}(

n 2)

◮ Output: e.g. b ∈ {0, 1} or v ∈ Rn

◮ “Low” means O(log n) where n is dimension Examples of low-degree algorithms: input Y ∈ Rn×n ◮ Power iteration: Y k1 ◮ Approximate message passing ◮ Local algorithms on sparse graphs ◮ Or any of the above applied to ˜ Y = g(Y )

6 / 18

slide-40
SLIDE 40

Planted Problems

For problems with a planted signal, the low-degree framework is already well-established

[Barak, Hopkins, Kelner, Kothari, Moitra, Potechin ’16] [Hopkins, Steurer ’17] [Hopkins, Kothari, Potechin, Raghavendra, Schramm, Steurer ’17] [Hopkins ’18] (PhD thesis)

7 / 18

slide-41
SLIDE 41

Planted Problems

For problems with a planted signal, the low-degree framework is already well-established

[Barak, Hopkins, Kelner, Kothari, Moitra, Potechin ’16] [Hopkins, Steurer ’17] [Hopkins, Kothari, Potechin, Raghavendra, Schramm, Steurer ’17] [Hopkins ’18] (PhD thesis)

Example (planted clique): G(n, 1/2) with planted k-clique ◮ Detection ◮ Recovery

7 / 18

slide-42
SLIDE 42

Planted Problems (Continued)

For all of these planted problems...

8 / 18

slide-43
SLIDE 43

Planted Problems (Continued)

For all of these planted problems...

planted clique, sparse PCA, community detection, tensor PCA, spiked Wigner/Wishart, planted submatrix, planted dense subgraph, ...

8 / 18

slide-44
SLIDE 44

Planted Problems (Continued)

For all of these planted problems...

planted clique, sparse PCA, community detection, tensor PCA, spiked Wigner/Wishart, planted submatrix, planted dense subgraph, ...

...it is the case that

8 / 18

slide-45
SLIDE 45

Planted Problems (Continued)

For all of these planted problems...

planted clique, sparse PCA, community detection, tensor PCA, spiked Wigner/Wishart, planted submatrix, planted dense subgraph, ...

...it is the case that ◮ the best known poly-time algorithms are captured by O(log n)-degree polynomials (spectral/AMP)

8 / 18

slide-46
SLIDE 46

Planted Problems (Continued)

For all of these planted problems...

planted clique, sparse PCA, community detection, tensor PCA, spiked Wigner/Wishart, planted submatrix, planted dense subgraph, ...

...it is the case that ◮ the best known poly-time algorithms are captured by O(log n)-degree polynomials (spectral/AMP) ◮ low-degree polynomials fail in the “hard” regime

[BHKKMP16,HS17,HKPRSS17,Hop18,BKW19,KWB19,DKWB19]

8 / 18

slide-47
SLIDE 47

Planted Problems (Continued)

For all of these planted problems...

planted clique, sparse PCA, community detection, tensor PCA, spiked Wigner/Wishart, planted submatrix, planted dense subgraph, ...

...it is the case that ◮ the best known poly-time algorithms are captured by O(log n)-degree polynomials (spectral/AMP) ◮ low-degree polynomials fail in the “hard” regime

[BHKKMP16,HS17,HKPRSS17,Hop18,BKW19,KWB19,DKWB19]

This work: extend low-degree framework to non-planted setting

8 / 18

slide-48
SLIDE 48

Planted Problems (Continued)

For all of these planted problems...

planted clique, sparse PCA, community detection, tensor PCA, spiked Wigner/Wishart, planted submatrix, planted dense subgraph, ...

...it is the case that ◮ the best known poly-time algorithms are captured by O(log n)-degree polynomials (spectral/AMP) ◮ low-degree polynomials fail in the “hard” regime

[BHKKMP16,HS17,HKPRSS17,Hop18,BKW19,KWB19,DKWB19]

This work: extend low-degree framework to non-planted setting Other frameworks: sum-of-squares, statistical query model

8 / 18

slide-49
SLIDE 49

Planted Problems (Continued)

For all of these planted problems...

planted clique, sparse PCA, community detection, tensor PCA, spiked Wigner/Wishart, planted submatrix, planted dense subgraph, ...

...it is the case that ◮ the best known poly-time algorithms are captured by O(log n)-degree polynomials (spectral/AMP) ◮ low-degree polynomials fail in the “hard” regime

[BHKKMP16,HS17,HKPRSS17,Hop18,BKW19,KWB19,DKWB19]

This work: extend low-degree framework to non-planted setting Other frameworks: sum-of-squares, statistical query model “Robustness”: Gaussian elimination for XOR-SAT

8 / 18

slide-50
SLIDE 50

Spherical Spin Glass: Results

Example (spherical p-spin model): for Y ∈ R⊗p i.i.d. N(0, 1), max

v=1

1 √nY , v⊗p ALG < OPT (for p ≥ 3)

9 / 18

slide-51
SLIDE 51

Spherical Spin Glass: Results

Example (spherical p-spin model): for Y ∈ R⊗p i.i.d. N(0, 1), max

v=1

1 √nY , v⊗p ALG < OPT (for p ≥ 3) Result: no low-degree polynomial can achieve value OPT − ǫ

9 / 18

slide-52
SLIDE 52

Spherical Spin Glass: Results

Example (spherical p-spin model): for Y ∈ R⊗p i.i.d. N(0, 1), max

v=1

1 √nY , v⊗p ALG < OPT (for p ≥ 3) Result: no low-degree polynomial can achieve value OPT − ǫ Theorem [Gamarnik, Jagannath, W. ’20] Let p ≥ 4 be even. For some ǫ > 0, no f : R⊗p → Rn of degree polylog(n) achieves both of the following with probability 1 − exp(−nΩ(1)):

9 / 18

slide-53
SLIDE 53

Spherical Spin Glass: Results

Example (spherical p-spin model): for Y ∈ R⊗p i.i.d. N(0, 1), max

v=1

1 √nY , v⊗p ALG < OPT (for p ≥ 3) Result: no low-degree polynomial can achieve value OPT − ǫ Theorem [Gamarnik, Jagannath, W. ’20] Let p ≥ 4 be even. For some ǫ > 0, no f : R⊗p → Rn of degree polylog(n) achieves both of the following with probability 1 − exp(−nΩ(1)): ◮ Objective: H(f (Y )) ≥ OPT − ǫ

9 / 18

slide-54
SLIDE 54

Spherical Spin Glass: Results

Example (spherical p-spin model): for Y ∈ R⊗p i.i.d. N(0, 1), max

v=1

1 √nY , v⊗p ALG < OPT (for p ≥ 3) Result: no low-degree polynomial can achieve value OPT − ǫ Theorem [Gamarnik, Jagannath, W. ’20] Let p ≥ 4 be even. For some ǫ > 0, no f : R⊗p → Rn of degree polylog(n) achieves both of the following with probability 1 − exp(−nΩ(1)): ◮ Objective: H(f (Y )) ≥ OPT − ǫ ◮ Normalization: f (Y ) ≈ 1

9 / 18

slide-55
SLIDE 55

Max Independent Set: Results

Example (max independent set): given sparse graph G(n, d/n), max

S⊆[n] |S|

s.t. S independent

10 / 18

slide-56
SLIDE 56

Max Independent Set: Results

Example (max independent set): given sparse graph G(n, d/n), max

S⊆[n] |S|

s.t. S independent OPT = 2 log d d n

10 / 18

slide-57
SLIDE 57

Max Independent Set: Results

Example (max independent set): given sparse graph G(n, d/n), max

S⊆[n] |S|

s.t. S independent OPT = 2 log d d n ALG = log d d n

10 / 18

slide-58
SLIDE 58

Max Independent Set: Results

Example (max independent set): given sparse graph G(n, d/n), max

S⊆[n] |S|

s.t. S independent OPT = 2 log d d n ALG = log d d n Result: no low-degree polynomial can achieve (1 +

1 √ 2) log d d n

10 / 18

slide-59
SLIDE 59

Max Independent Set: Results

Example (max independent set): given sparse graph G(n, d/n), max

S⊆[n] |S|

s.t. S independent OPT = 2 log d d n ALG = log d d n Result: no low-degree polynomial can achieve (1 +

1 √ 2) log d d n

Theorem [Gamarnik, Jagannath, W. ’20] No polynomial f : {0, 1}(n

2) → Rn of degree polylog(n) achieves

both of the following with probability 1 − exp(−nΩ(1)):

10 / 18

slide-60
SLIDE 60

Max Independent Set: Results

Example (max independent set): given sparse graph G(n, d/n), max

S⊆[n] |S|

s.t. S independent OPT = 2 log d d n ALG = log d d n Result: no low-degree polynomial can achieve (1 +

1 √ 2) log d d n

Theorem [Gamarnik, Jagannath, W. ’20] No polynomial f : {0, 1}(n

2) → Rn of degree polylog(n) achieves

both of the following with probability 1 − exp(−nΩ(1)): ◮ fi(Y ) ∈ [0, 1/3] ∪ [2/3, 1] for most i

10 / 18

slide-61
SLIDE 61

Max Independent Set: Results

Example (max independent set): given sparse graph G(n, d/n), max

S⊆[n] |S|

s.t. S independent OPT = 2 log d d n ALG = log d d n Result: no low-degree polynomial can achieve (1 +

1 √ 2) log d d n

Theorem [Gamarnik, Jagannath, W. ’20] No polynomial f : {0, 1}(n

2) → Rn of degree polylog(n) achieves

both of the following with probability 1 − exp(−nΩ(1)): ◮ fi(Y ) ∈ [0, 1/3] ∪ [2/3, 1] for most i ◮ {i : fi(Y ) ∈ [2/3, 1]} is a near-indep set of size (1 +

1 √ 2) log d d n

10 / 18

slide-62
SLIDE 62

Max Independent Set: Results

Example (max independent set): given sparse graph G(n, d/n), max

S⊆[n] |S|

s.t. S independent OPT = 2 log d d n ALG = log d d n Result: no low-degree polynomial can achieve (1 +

1 √ 2) log d d n

Theorem [Gamarnik, Jagannath, W. ’20] No polynomial f : {0, 1}(n

2) → Rn of degree polylog(n) achieves

both of the following with probability 1 − exp(−nΩ(1)): ◮ fi(Y ) ∈ [0, 1/3] ∪ [2/3, 1] for most i ◮ {i : fi(Y ) ∈ [2/3, 1]} is a near-indep set of size (1 +

1 √ 2) log d d n

Forthcoming: improve 1 +

1 √ 2 → 1 + ǫ

(optimal)

10 / 18

slide-63
SLIDE 63

Optimization [Gamarnik, Jagannath, W. ’20]

How to prove failure of low-degree polynomials?

11 / 18

slide-64
SLIDE 64

Optimization [Gamarnik, Jagannath, W. ’20]

How to prove failure of low-degree polynomials? For problems with a planted signal: ◮ Detection: linear algebra

[BHKKMP’16; HS’17; HKPRSS’17]

◮ Recovery: Jensen + linear algebra

[Schramm, W. ’20]

11 / 18

slide-65
SLIDE 65

Optimization [Gamarnik, Jagannath, W. ’20]

How to prove failure of low-degree polynomials? For problems with a planted signal: ◮ Detection: linear algebra

[BHKKMP’16; HS’17; HKPRSS’17]

◮ Recovery: Jensen + linear algebra

[Schramm, W. ’20]

For random optimization problems, need different approach:

11 / 18

slide-66
SLIDE 66

Optimization [Gamarnik, Jagannath, W. ’20]

How to prove failure of low-degree polynomials? For problems with a planted signal: ◮ Detection: linear algebra

[BHKKMP’16; HS’17; HKPRSS’17]

◮ Recovery: Jensen + linear algebra

[Schramm, W. ’20]

For random optimization problems, need different approach: ◮ Stability of low-degree polynomials

11 / 18

slide-67
SLIDE 67

Optimization [Gamarnik, Jagannath, W. ’20]

How to prove failure of low-degree polynomials? For problems with a planted signal: ◮ Detection: linear algebra

[BHKKMP’16; HS’17; HKPRSS’17]

◮ Recovery: Jensen + linear algebra

[Schramm, W. ’20]

For random optimization problems, need different approach: ◮ Stability of low-degree polynomials ◮ Overlap gap property (OGP)

[Gamarnik, Sudan ’13] [Chen, Gamarnik, Panchenko, Rahman ’17] [Gamarnik, Jagannath ’19]

11 / 18

slide-68
SLIDE 68

Low-Degree Polynomials are Stable

12 / 18

slide-69
SLIDE 69

Low-Degree Polynomials are Stable

Theorem Let Y , Y ′ be ρ-correlated samples from N(0, Im)

12 / 18

slide-70
SLIDE 70

Low-Degree Polynomials are Stable

Theorem Let Y , Y ′ be ρ-correlated samples from N(0, Im) Let f : Rm → Rn have degree ≤ D

12 / 18

slide-71
SLIDE 71

Low-Degree Polynomials are Stable

Theorem Let Y , Y ′ be ρ-correlated samples from N(0, Im) Let f : Rm → Rn have degree ≤ D Normalization EY f (Y )2 = 1

12 / 18

slide-72
SLIDE 72

Low-Degree Polynomials are Stable

Theorem Let Y , Y ′ be ρ-correlated samples from N(0, Im) Let f : Rm → Rn have degree ≤ D Normalization EY f (Y )2 = 1 Then for any t ≥ (6e)D, Pr

  • f (Y ) − f (Y ′)2 ≥ 2t(1 − ρD)
  • ≤ exp
  • − D

3e t1/D

  • 12 / 18
slide-73
SLIDE 73

Low-Degree Polynomials are Stable

Theorem Let Y , Y ′ be ρ-correlated samples from N(0, Im) Let f : Rm → Rn have degree ≤ D Normalization EY f (Y )2 = 1 Then for any t ≥ (6e)D, Pr

  • f (Y ) − f (Y ′)2 ≥ 2t(1 − ρD)
  • ≤ exp
  • − D

3e t1/D

  • Proof: low-degree polynomials have

◮ Low noise sensitivty ◮ Low total influence ◮ Hypercontractivity

12 / 18

slide-74
SLIDE 74

Low-Degree Polynomials are Stable (Binary Case)

13 / 18

slide-75
SLIDE 75

Low-Degree Polynomials are Stable (Binary Case)

Y ∼ i.i.d. Bernoulli(p)

13 / 18

slide-76
SLIDE 76

Low-Degree Polynomials are Stable (Binary Case)

Y ∼ i.i.d. Bernoulli(p) Interpolation path: Y (0) Y (1) Y (2) · · · Y (m−1) Y (m)

13 / 18

slide-77
SLIDE 77

Low-Degree Polynomials are Stable (Binary Case)

Y ∼ i.i.d. Bernoulli(p) Interpolation path: Y (0) Y (1) Y (2) · · · Y (m−1) Y (m) f : {0, 1}m → Rn degree D

13 / 18

slide-78
SLIDE 78

Low-Degree Polynomials are Stable (Binary Case)

Y ∼ i.i.d. Bernoulli(p) Interpolation path: Y (0) Y (1) Y (2) · · · Y (m−1) Y (m) f : {0, 1}m → Rn degree D Definition: Index i is “c-bad” if f (Y (i)) − f (Y (i−1))2 > c E

Y f (Y )2

13 / 18

slide-79
SLIDE 79

Low-Degree Polynomials are Stable (Binary Case)

Y ∼ i.i.d. Bernoulli(p) Interpolation path: Y (0) Y (1) Y (2) · · · Y (m−1) Y (m) f : {0, 1}m → Rn degree D Definition: Index i is “c-bad” if f (Y (i)) − f (Y (i−1))2 > c E

Y f (Y )2

Theorem Pr

Y (0),...,Y (m) [∄ c-bad i] ≥ p4D/c

13 / 18

slide-80
SLIDE 80

Low-Degree Polynomials are Stable (Binary Case)

Y ∼ i.i.d. Bernoulli(p) Interpolation path: Y (0) Y (1) Y (2) · · · Y (m−1) Y (m) f : {0, 1}m → Rn degree D Definition: Index i is “c-bad” if f (Y (i)) − f (Y (i−1))2 > c E

Y f (Y )2

Theorem Pr

Y (0),...,Y (m) [∄ c-bad i] ≥ p4D/c

With non-trivial probability (over path), f ’s output is “smooth”

13 / 18

slide-81
SLIDE 81

Overlap Gap Property

Overlap gap property (OGP): with high probability, Y ∼ G(n, d/n) has no occurrence of

14 / 18

slide-82
SLIDE 82

Overlap Gap Property

Overlap gap property (OGP): with high probability, Y ∼ G(n, d/n) has no occurrence of ◮ S, T independent sets

14 / 18

slide-83
SLIDE 83

Overlap Gap Property

Overlap gap property (OGP): with high probability, Y ∼ G(n, d/n) has no occurrence of ◮ S, T independent sets ◮ |S|, |T| ≈ (1 +

1 √ 2)Φ

14 / 18

slide-84
SLIDE 84

Overlap Gap Property

Overlap gap property (OGP): with high probability, Y ∼ G(n, d/n) has no occurrence of ◮ S, T independent sets ◮ |S|, |T| ≈ (1 +

1 √ 2)Φ

◮ |S ∩ T| ≈ Φ

14 / 18

slide-85
SLIDE 85

Overlap Gap Property

Overlap gap property (OGP): with high probability, Y ∼ G(n, d/n) has no occurrence of ◮ S, T independent sets ◮ |S|, |T| ≈ (1 +

1 √ 2)Φ

◮ |S ∩ T| ≈ Φ Proof: first moment method [Gamarnik, Sudan ’13]

14 / 18

slide-86
SLIDE 86

Ensemble OGP

Ensemble OGP: with high probability, ∀i, j on the interpolation path Y (0) Y (1) Y (2) · · · Y (m−1) Y (m) there is no occurrence of ◮ S independent set in Y (i) ◮ T independent set in Y (j) ◮ |S|, |T| ≈ (1 +

1 √ 2)Φ

◮ |S ∩ T| ≈ Φ

15 / 18

slide-87
SLIDE 87

Putting it Together

Proof that low-degree polynomials fail:

16 / 18

slide-88
SLIDE 88

Putting it Together

Proof that low-degree polynomials fail: Suppose f (Y ) outputs independent sets of size (1 +

1 √ 2)Φ

16 / 18

slide-89
SLIDE 89

Putting it Together

Proof that low-degree polynomials fail: Suppose f (Y ) outputs independent sets of size (1 +

1 √ 2)Φ

Y (0) Y (1) Y (2) · · · Y (m−1) Y (m)

16 / 18

slide-90
SLIDE 90

Putting it Together

Proof that low-degree polynomials fail: Suppose f (Y ) outputs independent sets of size (1 +

1 √ 2)Φ

Y (0) Y (1) Y (2) · · · Y (m−1) Y (m) Separation: f (Y (0)) and f (Y (m)) are “far apart”

16 / 18

slide-91
SLIDE 91

Putting it Together

Proof that low-degree polynomials fail: Suppose f (Y ) outputs independent sets of size (1 +

1 √ 2)Φ

Y (0) Y (1) Y (2) · · · Y (m−1) Y (m) Separation: f (Y (0)) and f (Y (m)) are “far apart” Stability: with probability n−D, there are no big “jumps” f (Y (i)) → f (Y (i+1))

16 / 18

slide-92
SLIDE 92

Putting it Together

Proof that low-degree polynomials fail: Suppose f (Y ) outputs independent sets of size (1 +

1 √ 2)Φ

Y (0) Y (1) Y (2) · · · Y (m−1) Y (m) Separation: f (Y (0)) and f (Y (m)) are “far apart” Stability: with probability n−D, there are no big “jumps” f (Y (i)) → f (Y (i+1)) Contradicts OGP

16 / 18

slide-93
SLIDE 93

Comments

17 / 18

slide-94
SLIDE 94

Comments

◮ Improvement to (1 + ǫ) log d

d n

◮ Inspired by [Rahman, Vir´ ag ’14]

17 / 18

slide-95
SLIDE 95

Comments

◮ Improvement to (1 + ǫ) log d

d n

◮ Inspired by [Rahman, Vir´ ag ’14]

◮ Proof of OGP for p-spin (for p ≥ 4 even)

[Chen, Sen ’15; Auffinger, Chen ’17]

17 / 18

slide-96
SLIDE 96

Comments

◮ Improvement to (1 + ǫ) log d

d n

◮ Inspired by [Rahman, Vir´ ag ’14]

◮ Proof of OGP for p-spin (for p ≥ 4 even)

[Chen, Sen ’15; Auffinger, Chen ’17]

◮ Langevin dynamics

17 / 18

slide-97
SLIDE 97

Comments

◮ Improvement to (1 + ǫ) log d

d n

◮ Inspired by [Rahman, Vir´ ag ’14]

◮ Proof of OGP for p-spin (for p ≥ 4 even)

[Chen, Sen ’15; Auffinger, Chen ’17]

◮ Langevin dynamics ◮ Connections between heuristics

◮ OGP → Low-Degree

17 / 18

slide-98
SLIDE 98

References for the Low-Degree Framework

◮ Detection (survey article)

Notes on Computational Hardness of Hypothesis Testing: Predictions using the Low-Degree Likelihood Ratio Kunisky, W., Bandeira arXiv:1907.11636

◮ Recovery

Computational Barriers to Estimation from Low-Degree Polynomials Schramm, W. arXiv:2008.02269

◮ Optimization

Low-Degree Hardness of Random Optimization Problems Gamarnik, Jagannath, W. arXiv:2004.12063

18 / 18