a space optimal streaming algorithm for sketching small
play

A Space Optimal Streaming Algorithm for Sketching Small Moments - PowerPoint PPT Presentation

Introduction F p Algorithm Lower Bounds Conclusion A Space Optimal Streaming Algorithm for Sketching Small Moments Daniel M. Kane Jelani Nelson David P. Woodruff Harvard MIT IBM Almaden December 18, 2009 Introduction F p Algorithm Lower


  1. Introduction F p Algorithm Lower Bounds Conclusion A Space Optimal Streaming Algorithm for Sketching Small Moments Daniel M. Kane Jelani Nelson David P. Woodruff Harvard MIT IBM Almaden December 18, 2009

  2. Introduction F p Algorithm Lower Bounds Conclusion Streaming moments: problem formulation Model • x = ( x 1 , x 2 , . . . , x n ) starts off as � 0 • m updates ( i 1 , v 1 ) , ( i 2 , v 2 ) , . . . , ( i m , v m ) • Update ( i , v ) causes change x i ← x i + v • v ∈ {− M , . . . , M }

  3. Introduction F p Algorithm Lower Bounds Conclusion Streaming moments: problem formulation Model • x = ( x 1 , x 2 , . . . , x n ) starts off as � 0 • m updates ( i 1 , v 1 ) , ( i 2 , v 2 ) , . . . , ( i m , v m ) • Update ( i , v ) causes change x i ← x i + v • v ∈ {− M , . . . , M } n | x i | p = � x � p def � Goal: Output F p = p i =1

  4. Introduction F p Algorithm Lower Bounds Conclusion Streaming moments: objectives Objectives • Minimize space usage • Minimize update time Trivial solutions • Keep x in memory: O ( n log( mM )) space / O (1) time • Keep stream in memory: O ( m log( nM )) space / O (1) time Goal: Get polylogarithmic dependence on n , m

  5. Introduction F p Algorithm Lower Bounds Conclusion Streaming moments: bad news Alon, Matias, Szegedy ’99: No sublinear space algorithms without • Approximation (allow output to be (1 ± ε ) F p ) • Randomization (allow 1% failure probability) New goal: Output (1 ± ε ) F p with probability 99%

  6. Introduction F p Algorithm Lower Bounds Conclusion Streaming moments: bad news Alon, Matias, Szegedy ’99: No sublinear space algorithms without • Approximation (allow output to be (1 ± ε ) F p ) • Randomization (allow 1% failure probability) New goal: Output (1 ± ε ) F p with probability 99% More bad news: Polynomial space required for p > 2 ([BJKS ’02] and [CKS ’03])

  7. Introduction F p Algorithm Lower Bounds Conclusion Streaming moments: bad news Alon, Matias, Szegedy ’99: No sublinear space algorithms without • Approximation (allow output to be (1 ± ε ) F p ) • Randomization (allow 1% failure probability) New goal: Output (1 ± ε ) F p with probability 99% More bad news: Polynomial space required for p > 2 ([BJKS ’02] and [CKS ’03]) Newer goal: Output (1 ± ε ) F p with probability 99% for 0 ≤ p ≤ 2

  8. Introduction F p Algorithm Lower Bounds Conclusion Contributions (0 < p ≤ 2) ( Notation: N = min { n , m } ) Ref Upper bound Lower bound Update time O ( ε − 2 log( mM )) (p=2) AMS’99 Ω(log N ) O (1) (*) “ log( NM ) O ( ε − 2 log( mM )) (p=1) ” FKSV’99 (**) ———— O ε 2 O ( ε − 2 log( mM ) log N ) O ( ε − 2 ) Indyk’06, Li’08 ———— O ( ε − (2+ p ) log 2 ( N ) log( mM )) GC’07 ———— polylog ( mM ) Ω( ε − 2 ) Woodruff’04 ———— ———— O ( ε − 2 log( mM )) Ω( ε − 2 log( mM )) O ( ε − 2 ) ˜ This work (*) achieved by CCF’02, TZ’04 , (**) L 1 -difference only

  9. Introduction F p Algorithm Lower Bounds Conclusion F p (0 < p < 2) p -stable distributions Definition (Zolotarev ’86) For 0 < p ≤ 2, there exists a probability distribution D p called the p-stable distribution such that if Q 1 , . . . , Q n ∼ D p are independent, then � n i =1 Q i x i ∼ � x � p D p . (In short: D p carries information about L p norms)

  10. Introduction F p Algorithm Lower Bounds Conclusion F p (0 < p < 2) p -stable distributions Definition (Zolotarev ’86) For 0 < p ≤ 2, there exists a probability distribution D p called the p-stable distribution such that if Q 1 , . . . , Q n ∼ D p are independent, then � n i =1 Q i x i ∼ � x � p D p . (In short: D p carries information about L p norms) • p = 2: Gaussian • p = 1: Cauchy • p = 1 / 2: L´ evy

  11. Introduction F p Algorithm Lower Bounds Conclusion Algorithms based on p -stable sketch matrices   · · · A 1 , 1 A 1 , n . . ... . . A =  , the A i , j are i.i.d. from D p ,   . .  A r , 1 · · · A r , n Maintain Ax = y

  12. Introduction F p Algorithm Lower Bounds Conclusion Algorithms based on p -stable sketch matrices   · · · A 1 , 1 A 1 , n . . ... . . A =  , the A i , j are i.i.d. from D p ,   . .  A r , 1 · · · A r , n Maintain Ax = y • Idea introduced by Indyk ’06 • Indyk ’06: Estimate F p as median {| y j | p } r j =1 Q r j =1 | y j | p / r • Li ’08: Estimate F p as r π Γ ( p 2 · p [ 2 r ) Γ ( 1 − 1 r ) sin ( π r )] • Both cases: r = Θ(1 /ε 2 )

  13. Introduction F p Algorithm Lower Bounds Conclusion Too much randomness • In Indyk’06 and Li’08, Ω( n /ε 2 ) bits needed to store matrix A

  14. Introduction F p Algorithm Lower Bounds Conclusion Too much randomness • In Indyk’06 and Li’08, Ω( n /ε 2 ) bits needed to store matrix A • Indyk derandomized using Nisan’s pseudorandom generator (but blowed up space)

  15. Introduction F p Algorithm Lower Bounds Conclusion Too much randomness • In Indyk’06 and Li’08, Ω( n /ε 2 ) bits needed to store matrix A • Indyk derandomized using Nisan’s pseudorandom generator (but blowed up space) Is there a more efficient derandomization?

  16. Introduction F p Algorithm Lower Bounds Conclusion Our Contributions Yes, via k -wise independence! • For fixed i , make the A i , j k -wise independent • Make the seeds used to generate rows of A pairwise independent

  17. Introduction F p Algorithm Lower Bounds Conclusion Our Contributions Yes, via k -wise independence! • For fixed i , make the A i , j k -wise independent • Make the seeds used to generate rows of A pairwise independent • k = ˜ Θ(1 /ε p ) fools Indyk’s estimator • A different estimator works with k = Θ(log(1 /ε ) / log log(1 /ε )).

  18. Introduction F p Algorithm Lower Bounds Conclusion Our Contributions A different estimator (works with k = O (log(1 /ε ) / log log(1 /ε ))) 1. Maintain Ax = y and A ′ x = y ′ . 2. A has k = Θ(log(1 /ε ) / log log(1 /ε )), r = Θ(1 /ε 2 ). 3. A ′ has k ′ , r ′ = Θ(1). j |} r ′ 4. y ′ med ← median {| y ′ j =1 . � � �� y j 5. Output − y ′ p � r 1 med · ln j =1 cos . r y ′ med

  19. Introduction F p Algorithm Lower Bounds Conclusion Analyzing median F p algorithm (full independence) An argument for the median: Define � 1 , if x ∈ [ a , b ], I [ a , b ] ( x ) = 0 , otherwise • Q = � i Q i x i .

  20. Introduction F p Algorithm Lower Bounds Conclusion Analyzing median F p algorithm (full independence) An argument for the median: Define � 1 , if x ∈ [ a , b ], I [ a , b ] ( x ) = 0 , otherwise • Q = � i Q i x i . • “ median ( | Q | / � x � p ) = 1” means E [ I [ − 1 , 1] ( Q / � x � p )] = 1 / 2.

  21. Introduction F p Algorithm Lower Bounds Conclusion Analyzing median F p algorithm (full independence) An argument for the median: Define � 1 , if x ∈ [ a , b ], I [ a , b ] ( x ) = 0 , otherwise • Q = � i Q i x i . • “ median ( | Q | / � x � p ) = 1” means E [ I [ − 1 , 1] ( Q / � x � p )] = 1 / 2. • E [ I [ − 1+ ε, 1 − ε ] ( Q / � x � p )] = 1 / 2 − Θ( ε ) • E [ I [ − 1 − ε, 1+ ε ] ( Q / � x � p )] = 1 / 2 + Θ( ε ) • Take r = Θ(1 /ε 2 ) trials Q 1 , . . . , Q r . Number of counters inside interval is concentrated by Chebyshev. ⇒ median of the | Q j | is (1 ± Θ( ε )) � x � p with probability 2 / 3

  22. Introduction F p Algorithm Lower Bounds Conclusion Analyzing median F p algorithm ( k -wise independence) One possible path • Replace I [ a , b ] with a well-approximating low-degree polynomial. • k -wise independence fools polynomials.

  23. Introduction F p Algorithm Lower Bounds Conclusion Analyzing median F p algorithm ( k -wise independence) One possible path • Replace I [ a , b ] with a well-approximating low-degree polynomial. • k -wise independence fools polynomials. What we actually do (for good reason) • Replace I [ a , b ] with a well-approximating smooth function ˜ I [ a , b ] . • Show ˜ I [ a , b ] is fooled by k -wise independence via Taylor’s theorem.

  24. Introduction F p Algorithm Lower Bounds Conclusion Defining ˜ I [ a , b ] FT-mollification Define  x 2 e − 1 − x 2 for | x | < 1  b ( x ) = 0 otherwise  and [ a , b ] ( x ) = 1 ˜ I c 2 π ( c · ˆ b ( ct ) ∗ I [ a , b ] ( t ))( x )

  25. Introduction F p Algorithm Lower Bounds Conclusion Defining ˜ I [ a , b ] FT-mollification Define  x 2 e − 1 − x 2 for | x | < 1  b ( x ) = 0 otherwise  and [ a , b ] ( x ) = 1 ˜ I c 2 π ( c · ˆ b ( ct ) ∗ I [ a , b ] ( t ))( x ) Then, for c > 1, i. � (˜ I c [ a , b ] ) ( ℓ ) � ∞ = O ( c ℓ ) for ℓ ≥ 0. ii. For c = ˜ O (1 /ε ), | ˜ I c [ a , b ] − I [ a , b ] | < ε except potentially at a ± ε and b ± ε .

  26. Introduction F p Algorithm Lower Bounds Conclusion Defining ˜ I [ a , b ] FT-mollification Define  x 2 e − 1 − x 2 for | x | < 1  b ( x ) = 0 otherwise  and [ a , b ] ( x ) = 1 ˜ I c 2 π ( c · ˆ b ( ct ) ∗ I [ a , b ] ( t ))( x ) Then, for c > 1, i. � (˜ I c [ a , b ] ) ( ℓ ) � ∞ = O ( c ℓ ) for ℓ ≥ 0. ii. For c = ˜ O (1 /ε ), | ˜ I c [ a , b ] − I [ a , b ] | < ε except potentially at a ± ε and b ± ε . For c large, ˜ I c [ a , b ] looks like I [ a , b ] .

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend