bounds on reliable boolean function computation with
play

Bounds on Reliable Boolean Function Computation with Noisy Gates - - PowerPoint PPT Presentation

Bounds on Reliable Boolean Function Computation with Noisy Gates - R. L. Dobrushin & S. I. Ortyukov, 1977 - N. Pippenger, 1985 - P . G acs & A. G al, 1994 Presenter: Da Wang 6.454 Graduate Seminar in Area I EECS, MIT Oct. 5,


  1. Bounds on Reliable Boolean Function Computation with Noisy Gates - R. L. Dobrushin & S. I. Ortyukov, 1977 - N. Pippenger, 1985 - P . G´ acs & A. G´ al, 1994 Presenter: Da Wang 6.454 Graduate Seminar in Area I EECS, MIT Oct. 5, 2011 0 / 33

  2. Question Given a network of noisy logic gates, what is the redundancy required if we want to compute the a Boolean function reliably? noisy: gates produce the wrong output independently with error probability no more than ε . reliably: the value computed by the entire circuit is correct with probability at least 1 − δ redundancy: minimum #gates needed for reliable computation in noisy circuit minimum #gates needed for reliable computation in noiseless circuit ◮ noisy/noiseless complexity ◮ may depend on the function of interest ◮ upper bound: achievability ◮ lower bound: converse 1 / 33

  3. Part I Lower Bounds for the Complexity of Reliable Boolean Circuits with Noisy Gates 2 / 33

  4. History of development [Dobrushin & Ortyukov 1977] ◮ Contains all the key ideas ◮ Proofs for a few lemmas are incorrect [Pippenger & Stamoulis & Tsitsiklis 1990] ◮ Pointed out the errors in [DO1977] ◮ Provide proofs for the case of computing the parity function [G´ acs & G´ al 1994] ◮ Follow the ideas in [DO1977] and provide correct proofs ◮ Also prove some stronger results In this talk We will mainly follow the presentation in [G´ acs & G´ al 1994]. 3 / 33

  5. Problem formulation System Model Boolean circuit C Gate g a function g : { 0 , 1 } n g → { 0 , 1 } a directed acycic graph ◮ n g : fan-in of the gate node ∼ gate edge ∼ in/out of a gate Basis Φ Assumptions a set of possible gate functions each gate g has constant number of fan-ins n g . e.g., Φ = { AND, OR, XOR } f can be represented by complete basis compositions of gate functions in for circuit C : Φ C Φ C . maximum fan-in in C : n (Φ C ) 4 / 33

  6. Problem formulation Error models ( ε, p ) Gate error Circuit error A gate fails if its output value C ( x ) : random variable for output for z ∈ { 0 , 1 } n g is different from of circuit C on input x . g ( z ) A circuit computes f with error gates fail independently with probability at most p if ◮ fixed probability ε P [ C ( x ) � = f ( x )] ≤ p used for lower bound proof ◮ probability at most ε for any input x . ε ∈ (0 , 1 / 2) 5 / 33

  7. Problem formulation Sensitivity of a Boolean function Let f : { 0 , 1 } n → { 0 , 1 } be a Boolean function with binary input vector x = ( x 1 , x 2 , . . . , x n ) . Let x l be a binary vector that differs from x only in the l -th bit, i.e., � x i i � = l x l i = i = l . ¬ x i f is sensitive to the l th bit on x if f ( x l ) � = f ( x ) . Sensitivity of f on x : #bits in x that f is sensitive to. ◮ “effecitive” input size Sensitivity of f : maximum over all x . 6 / 33

  8. Asymptotic notations f ( n ) = O ( g ( n )) : � f ( n ) � � � lim sup � < ∞ , � � g ( n ) n →∞ � f ( n ) = Ω ( g ( n )) : � � f ( n ) � � lim inf � ≥ 1 , � � g ( n ) n →∞ � f ( n ) = Θ ( g ( n )) : f ( n ) = O ( g ( n )) and f ( n ) = Ω ( g ( n )) 7 / 33

  9. Main results Theorem: number of gates for reliable computation ◮ Let ε and p be any constants such that ε ∈ (0 , 1 / 2) , p ∈ (0 , 1 / 2) . ◮ Let f be any Boolean function with sensitivity s . Under the error model ( ε, p ) , the number of gates of the curcuit is Ω ( s log s ) . Corollary: redundancy of noisy computation For any Boolean function of n variables and with O ( n ) noiseless complexity and Ω ( n ) sensitivity, the redundancy of noisy computation is Ω (log n ) . ◮ e.g., nonconstant symmetric function of n variables has redundancy Ω (log n ) 8 / 33

  10. Equivalence result for wire failures Lemma 3.1 in Dobrushin&Ortyukov ◮ Let ε ∈ (0 , 1 / 2) and δ ∈ [0 , ε/n (Φ C )] . ◮ Let y and t be the vector that a gate receives when the wire fail and does not fail respectively. For any gate g in the circuit C there exists unique values η g ( y , δ ) such that if ◮ the wires of C fails independently with error probability δ , and ◮ the gate g fails with probability η g ( y , δ ) when receiving input y , then the probability that the output of g is different from g ( t ) is equal to ε . Insights Independent gate failures can be “simulated” by independently wire failures and corresponding gate failures. These two failure modes are equivalent in the sense that the circuit C computes f with the same error probability. 9 / 33

  11. “Noisy-wires” version of the main result Theorem ◮ Let ε and p be any constants such that ε ∈ (0 , 1 / 2) , p ∈ (0 , 1 / 2) . ◮ Let f be any Boolean function with sensitivity s . Let C be a circuit such that ◮ its wires fail independently with fixed probability δ , and ◮ each gate fails independently with probability η g ( y , δ ) when receiving y . Suppose C computes f with error probability at most p . Then the number of gates of the curcuit is Ω ( s log s ) . 10 / 33

  12. Error analysis Function and circuit inputs Maximal sensitive set S for f s > 0 : sensitivity of f z : an input vector with s bits that f is sensitive to ◮ an input vector where f has maximum sensitivity S : the set of sensitive bits in z ◮ key object B l : edges originated from l -th input 1 m l � | B l | 2 e.g. ◮ l = 3 3 ◮ B l f ( z ) 4 ◮ m l = 3 11 / 33

  13. Error analysis Wire failures For β ⊂ B l , let H ( β ) be the event that w 1 for wires in B l , only those in β fail. w 2 input l Let w 3 � H ( β ) C ( z l ) = f ( z l ) � β l � arg max � � P β ⊂ B l B l = { w 1 , w 2 , w 3 } β = { w 2 } ◮ the best failing set for input z l Let H l � H ( B l \ β l ) Fact 1 � H ( β l ) C ( z l ) = f ( z l ) � � � P [ C ( z ) � = f ( z ) | H l ] = P Proof ◮ f is sensitive to z l ◮ ¬ z l ⇔ “flip” all wires in B l β l is the worst non-failing set for input z 12 / 33

  14. Error analysis Error probability given wire failures Fact 2 � H ( β l ) C ( z l ) = f ( z l ) � � � P ≥ 1 − p Proof C ( z l ) = f ( z l ) ◮ P � � ≥ 1 − p � H ( β ) ◮ β l maximizes P C ( z l ) = f ( z l ) � � � Fact 1 & 2 ⇒ Fact 3 For each l ∈ S , P [ C ( z ) � = f ( z ) | H l ] ≥ 1 − p where { H l , l ∈ S } are independent events. Furthermore, Lemma 4.3 in [G´ acs&G´ al 1994] shows � � � ≥ (1 − √ p ) 2 � � P C ( z ) � = f ( z ) H l � � � l ∈ S The error probability given H l or � l ∈ S H l is relatively large. 13 / 33

  15. Error analysis Bounds on wire failure probabilities Note p ≥ P [ C ( z ) � = f ( z )] � � � �� � � � ≥ P C ( z ) � = f ( z ) H l P H l � � � l ∈ S l ∈ S Fact 3 implies Fact 4 �� � p H l ≤ P (1 − √ p ) 2 l ∈ S which implies (via Lemma 4.1 in [G´ acs&G´ al 1994]), Fact 5 �� � � � � p H l ≥ 1 − P [ H l ] P (1 − √ p ) 2 l ∈ S l ∈ S 14 / 33

  16. Error analysis Bounds on the total number of sensitive wires Fact 6 P [ H l ] = (1 − δ ) | β l | δ m l −| β l | ≥ δ m l Fact 4 & 5 ⇒ p � δ m l 1 − 2 √ p ≥ l ∈ S � 1 /s �� δ m l ≥ s l ∈ S which leads to s 1 − 2 √ p s � � � m l ≥ log(1 /δ ) log p l ∈ S lower bound on the total number of “sensitive wires” 15 / 33

  17. Lower bound on number of gates Let N C be the total number of gates in C : � n (Φ C ) N C ≥ n g g � ≥ m l l ∈ S s 1 − 2 √ p s � � ≥ log(1 /δ ) log p Comments: The above proof is for p ∈ (0 , 1 / 4) The case p ∈ (1 / 4 , 1 / 2) can be shown similarly. 16 / 33

  18. Block Sensitivity Let x S be a binary vector that differs from x in the S subset of indicies, i.e., � x i i / ∈ S x S i = i ∈ S . ¬ x i f is (block) sensitive to S on x if f ( x S ) � = f ( x ) . Block sensitivity of f on x : the largest number b such that ◮ there exists b disjoint sets S 1 , S 2 , · · · , S b ◮ for all 1 ≤ i ≤ b , f is sensitive to S i on x Block sensitivity of f : maximum over all x . ◮ block sensitivity ≥ sensitivity Theorem based on block sensitivity ◮ Let ε and p be any constants such that ε ∈ (0 , 1 / 2) , p ∈ (0 , 1 / 2) . ◮ Let f be any Boolean function with block sensitivity b . Under the error model ( ε, p ) , the number of gates of the curcuit is Ω ( b log b ) . 17 / 33

  19. Discussions Lower bound for specific functions Given an explicit function f of n variables, is there a lower boudn that is stronger than Ω ( n log n ) ? Open problem for unrestricted circuit C with complete basis function f that have Ω ( n log n ) noiseless complexity for circuit C with some incomplete basis Φ 18 / 33

  20. Discussions Computation model Exponential blowup A noisy circuit with multiple levels The output of gates at level l goes to a gate at level l + 1 Level 0 has n inputs ◮ Level 0 has N 0 = n log n output gates ◮ Level 1 has N 0 inputs ◮ Level 1 has N 1 = N 0 log N 0 output gates, . . . Why? “The theorem is generally applicable only to the very first step of such a fault tolerant computation” If the input is not the original ones, we can choose them to make the sensitivity of a Boolean function to be 0. ◮ f ( x 1 , x 2 , x 3 , x 4 , x 1 ⊕ x 2 ⊕ x 4 , x 1 ⊕ x 3 ⊕ x 4 , x 2 ⊕ x 3 ⊕ x 4 ) ◮ Lower bound does not apply: sensitivity is 0. How about block sensitivity? Problem formulation issue on the lower bound for coded input ◮ coding is also computation! 19 / 33

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend