algorithms and lower bounds for de morgan formulas of low
play

Algorithms and lower bounds for de-Morgan formulas of low- - PowerPoint PPT Presentation

Algorithms and lower bounds for de-Morgan formulas of low- communication leaf gates Sajin Koroth (Simon Fraser University) Joint with Valentine Zhenjian Dimitrios Igor Carboni Kabanets Lu Myrisiotis Oliveira Outline Background


  1. Algorithms and lower bounds for de-Morgan formulas of low- communication leaf gates Sajin Koroth (Simon Fraser University) Joint with Valentine Zhenjian Dimitrios Igor Carboni Kabanets Lu Myrisiotis Oliveira

  2. Outline • Background • Circuit model : Formula [ s ] ∘ 𝒣 • Prior work • Results • Lower bounds • PRG’s • SAT algorithm’s • Learning algorithms • Overview of the lower bound technique

  3. Parallel vs Sequential computation • Most of linear algebra can be done in parallel • Gaussian elimination is an outlier • Intuitively its an inherently sequential procedure • There are theoretical reasons to believe so • There is an e ffi cient sequential algorithm

  4. P vs NC 1 Class P of poly-time solvable problems Are there problems with e ffi cient sequential algorithms which do not have e ffi cient parallel algorithms ? Modeled as circuits

  5. Circuit complexity Internal gates • Complexity parameters : • Size : # of gates • Depth : length of the longest path from root to leaf • Fan in : 2, Fan out • Formulas : Leaf gates • Underlying DAG is a tree • No reuse of computation • Depth = log ( Size )

  6. Circuit complexity Class = Poly-Size Formulas NC 1 F • E ffi cient parallel computation (formally CREW PRAM): size ( F ) = n O (1) • Polynomially many processors depth ( F ) = O (log n ) • Logarithmic computation time x 1 x 2 x n x 5 In formula, depth ( F ) = O (log size ( F ))

  7. Circuit complexity P vs rephrased NC 1 • A Boolean function (candidates: Perfect matching, Gaussian elimination etc) f • That can be computed in poly-time ( ) f ∈ P • Any de-Morgan formula computing it has super-poly size ( f ∉ NC 1 )

  8. P vs NC 1 State of the art • Andreev’87 : Ω ( n 2.5 − o (1) ) for a function in called the Andreev function P • Also, Andreev’87 : Ω ( n 1+ Γ− o (1) ) , where is the shrinkage exponent Γ • Paterson and Zwick’93 : Γ ≥ 1.63 • Hastad’98 (breakthrough) : Γ ≥ 2 − o (1) • Tal’14 : Γ = 2 Ω ( log 2 n log log n ) n 3 • Best l.b. for Andreev’s function (Tal’14) : Ω ( log n (log log n ) 2 ) n 3 • Best l.b. for a function in (Tal’16) : P

  9. Cubic formula lower bounds Andreev’s function f ( , x 1 x 2 x 3 x n ⋯⋯ 2 log n = n Truth Table of a bit function ( ) log n h ) = y 1 y 2 y 3 y n ⋯⋯ h z log n z 2 z 1 ⊕ ⊕ ⊕ n log n y n − y n y y 2 n y 1 y 2 y 3 y ⋯⋯ ⋯⋯ ⋯⋯ n n n log n log n log n log n

  10. Cubic formula lower bounds Hastad’s result Ω ( log 2 n log log n ) n 3 • (Tal’14) : • Doesn’t work if there are parity gates at bottom

  11. Our Model Augmenting de-Morgan formulas • de-Morgan formulas : leaf gates, input literals • Our model : leaf gates, low communication functions Leaf gates of low cc Leaf gates

  12. Our model Reformulation • Formula [ s ] ∘ 𝒣 • Size s de-Morgan formula • : A family of Boolean functions 𝒣 • Leaf gates are functions g ∈ 𝒣 • Our model : • - low communication complexity Boolean functions 𝒣 s = ˜ • O ( n 2 )

  13. Communication complexity • Yao’s 2-party model m 1 m 2 • Input divided into 2 parts x , y m k y x f ( x , y ) • Goal : compute f ( x , y ) with minimal communication

  14. Our model Complexity of Andreev’s function f ( , x 1 x 2 x 3 x n ⋯⋯ 2 log n = n Truth Table of a bit function ( ) log n h ) = y 1 y 2 y 3 y n ⋯⋯ n de-Morgan formula of size log n h Leaf gates z log n z 2 z 1 Communication complexity Of Parity = 2 bits ⊕ ⊕ ⊕ n log n y n − y n y y 2 n y 1 y 2 y 3 y ⋯⋯ ⋯⋯ ⋯⋯ n n n log n log n log n log n

  15. Our model Prior work - Bipartite Formulas F • Input is divided into two parts, x , y • Every leaf can gate can access any Boolean function of either or but not x y both Communication complexity Of a bipartite function = 1 bit • Models a well known measure - graph complexity g 1 g 2 g 3 g s • Tal’16: Bipartite formula complexity of ˜ Ω ( n 2 ) is IP n x 1 x 2 x 3 x n y 1 y 2 y 3 y n ⋯⋯ ⋯⋯ • Earlier methods could not do super linear

  16. Our model Connection to Hardness Magnification N = 2 n • : Given the truth table of a function on bits ( ) MCSP N [ k ] f n • Yes : if has a circuit of size at most f k • No : otherwise • Meta computational problem with connections to Crypto, learning theory, circuit complexity etc • OPS’19: • If there exists an such that MCSP N [2 o ( n ) ] Formula [ N 1+ ϵ ] ∘ XOR is not in ϵ • then, NP ∉ NC 1

  17. Our model Connection to PRG for polytopes • Polytope : AND of LTF’s • LTF : sign ( w 1 x 1 + … + w n x n − θ ) • w 1 , …, w n , θ ∈ ℝ • Ex : 3 x 1 + 4 x 2 + 5 x 7 ≥ 12 • Nisan’94 : Randomized communication complexity O (log n ) • PRG’s for polytopes : Approximate volume computation

  18. Our model Interesting low communication bottom gates • Bipartite functions • Parities • LTF’s (Linear threshold functions) • PTF’s (Polynomial threshold functions)

  19. Our results Target function - Generalized inner product x 1 1 x 1 2 x 1 x 1 ⋯⋯ • Generalization of binary inner n 3 k product ∧ ∧ ∧ IP n ( x , y ) = ∑ x i y i x 2 2 x 2 x 2 1 x 2 ⋯⋯ • n 3 k i ∈ [ n ] n ( x 1 , x 2 , …, x k ) = ∑ ∧ ∧ ∧ i ∈ [ n / k ] ∏ x j GIP k • i j ∈ [ k ] x k 2 x k x k 1 x k ⋯⋯ n 3 k ∧ ∧ ∧ z 1 z 2 z k ⊕ 0/1

  20. Our results Lower bound • Let GIP k be computed on average by , F ∈ Formula [ s ] ∘ 𝒣 n x [ F ( x ) = GIP k • That is, Pr n ( x )] ≥ 1/2 + ϵ s = Ω ( ϵ /2 n 2 ( 𝒣 ) ⋅ log 2 (1/ ϵ ) ) n 2 • Then, k 2 ⋅ 16 k ⋅ R k ϵ /2 n 2 R k • : Randomized communication of with error in the ϵ /2 n 2 ( 𝒣 ) 𝒣 number on forehead communication complexity model

  21. Our results MCSP lower bounds s = ˜ O ( n 2 ) • If MCSP N [2 cn ] is computed , then Formula [ s ] ∘ XOR • Contrast : OPS’19: MCSP N [2 o ( n ) ] • If there exists an such that is not in ϵ Formula [ N 1+ ϵ ] ∘ XOR • then, NP ∉ NC 1 MCSP N [2 o ( n ) ] • Our techniques cannot handle

  22. Our results PRG • A pseudo random generator is said to fool a function class if G ℱ ϵ z ∈ {0,1} l ( n ) [ f ( G ( z )) = 1 ] − x ∈ {0,1} n [ f ( x ) = 1 ] Pr Pr ≤ ϵ • • is any function from f ℱ G : {0,1} l ( n ) → {0,1} n • • is the seed, z l ( n ) ⋘ n • Smaller the seed length compared to the better n

  23. Our results PRG • Parities at the bottom can make things harder. • AC 0 best known PRG seed length poly (log n ) AC 0 ∘ XOR • best known only (1 − o (1)) n

  24. Our results PRG • There is a PRG that -fools Formula [ s ] ∘ XOR ϵ • Seed length : O ( s ⋅ log s ⋅ log(1/ ϵ ) + log n ) • Seed length is optimal, unless lower bound can be improved

  25. Our results PRG • Natural generalization to Formula [ s ] ∘ 𝒣 • There is a PRG that -fools Formula [ s ] ∘ 𝒣 ϵ s ⋅ ( R k − NIH • Seed length : n / k + O ( ( 𝒣 ) + log s ) ⋅ log(1/ ϵ ) + log k ) ⋅ log k ϵ /6 s • Number in hand

  26. Our results PRG - Corollaries • (Ours + Vio15) : There is a PRG O ( n 1/2 ⋅ m 1/4 ⋅ log n ⋅ log( n / ϵ )) • Seed length : • {0,1} n -fools intersection of halfspaces over m ϵ • Our results beats earlier results when and m = O ( n ) ϵ ≤ 1/ n

  27. Our results PRG - Corollaries • There is a PRG O ( n 1/2 ⋅ s 1/4 ⋅ log n ⋅ log( n / ϵ )) • Seed length : • -fools Formula [ s ] ∘ SYM ϵ • First of its kind • Blackbox counting algorithm (Whitebox due to CW19)

  28. Our results SAT Algorithm • Given circuit class 𝒟 • Circuit SAT : Given , is there an , x C ( x ) = 1 C ∈ 𝒟 • #Circuit SAT : Given , how many , C ∈ 𝒟 x C ( x ) = 1

  29. Our results SAT Algorithm • Randomized #SAT algorithm for Formula [ s ] ∘ 𝒣 • Running time 2 n − t 1/2 n for LTFs log n t = Ω • s ⋅ log 2 s ⋅ R 2 1/3 ( 𝒣 ) • First of its kind #SAT for unbounded depth Boolean circuits with PTF’s at the bottom

  30. Our results Learning algorithm • There is PAC-learning algorithm • Learns Formula [ n 2 − γ ] ∘ XOR • Accuracy : , Confidence : ϵ δ • Time complexity : poly (2 n /log n ,1/ ϵ , log(1/ δ )) • Formula [ n 2 − γ ] 2 o ( n ) can be learned in [Rei11] • Crypto connection: • is assumed to compute PRFs (BIP+18) MOD 3 ∘ XOR • If true, Formula [ n 2.8 ] ∘ XOR 2 o ( n ) can’t be learned in time

  31. Lower bound technique Outline • GIP k cannot even be weakly approximated by low communication n complexity functions • Weakness of : Size formula can be “approximated” by Formula [ s ] ∘ 𝒣 s degree polynomial s • GIP k is weakly approximated by a collection of leaf gates n

  32. Lower bound technique Part I • GIP k cannot even be weakly approximated by low communication n complexity functions • In the number on forehead model • Protocol computes GIP k with error (uniform distribution) ϵ n n /4 k − log(1/(1 − 2 ϵ )) • Then commn.comp >

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend