algorithms for big data xiv

Algorithms for Big Data (XIV) Chihao Zhang Shanghai Jiao Tong - PowerPoint PPT Presentation

Algorithms for Big Data (XIV) Chihao Zhang Shanghai Jiao Tong University Dec. 20, 2019 Algorithms for Big Data (XIV) 1/12 We defined the graph Laplacian Review Last week we studied electrical networks using matrices. : We also defined the


  1. Algorithms for Big Data (XIV) Chihao Zhang Shanghai Jiao Tong University Dec. 20, 2019 Algorithms for Big Data (XIV) 1/12

  2. We defined the graph Laplacian Review Last week we studied electrical networks using matrices. : We also defined the notion of efgective resistance between two vertices in terms of : eff e e e e Algorithms for Big Data (XIV) 2/12

  3. We defined the graph Laplacian Review Last week we studied electrical networks using matrices. : We also defined the notion of efgective resistance between two vertices in terms of : eff e e e e Algorithms for Big Data (XIV) 2/12

  4. We also defined the notion of efgective resistance between two vertices in terms of Review Last week we studied electrical networks using matrices. : eff e e e e Algorithms for Big Data (XIV) 2/12 We defined the graph Laplacian L : L = U T WU.

  5. Review Last week we studied electrical networks using matrices. Algorithms for Big Data (XIV) 2/12 We defined the graph Laplacian L : L = U T WU. We also defined the notion of efgective resistance between two vertices in terms of L : R eff ( u, v ) ≜ ( e u − e v ) T L + ( e u − e v ) .

  6. Sparsification Given a graph , the goal of sparsification is to construct a sparse graph such that Similar Laplacian implies similar spectrum; similar efgective resistance between any two vertices; similar clustering; … Algorithms for Big Data (XIV) 3/12

  7. Sparsification Similar Laplacian implies similar spectrum; similar efgective resistance between any two vertices; similar clustering; … Algorithms for Big Data (XIV) 3/12 Given a graph G , the goal of sparsification is to construct a sparse graph H such that ( 1 − ε ) L G ≼ L H ≼ ( 1 + ε ) L G .

  8. Sparsification Similar Laplacian implies Algorithms for Big Data (XIV) 3/12 Given a graph G , the goal of sparsification is to construct a sparse graph H such that ( 1 − ε ) L G ≼ L H ≼ ( 1 + ε ) L G . ▶ similar spectrum; ▶ similar efgective resistance between any two vertices; ▶ similar clustering; ▶ …

  9. The Construction We use to denote the Laplacian of the unweighted graph containing a single edge . For a graph , we have where is the weight on the edge . Let be a collection of probabilities on each pair of vertices. Algorithms for Big Data (XIV) 4/12

  10. The Construction For a graph , we have where is the weight on the edge . Let be a collection of probabilities on each pair of vertices. Algorithms for Big Data (XIV) 4/12 We use L u,v to denote the Laplacian of the unweighted graph containing a single edge { u, v } .

  11. The Construction Let be a collection of probabilities on each pair of vertices. Algorithms for Big Data (XIV) 4/12 We use L u,v to denote the Laplacian of the unweighted graph containing a single edge { u, v } . For a graph G = ( V, E ) , we have ∑ L G = w u,v · L u,v , { u,v } ∈ E where w u,v is the weight on the edge { u, v } ∈ E .

  12. Algorithms for Big Data (XIV) The Construction 4/12 We use L u,v to denote the Laplacian of the unweighted graph containing a single edge { u, v } . For a graph G = ( V, E ) , we have ∑ L G = w u,v · L u,v , { u,v } ∈ E where w u,v is the weight on the edge { u, v } ∈ E . Let { p u,v } { u,v } ∈ E be a collection of probabilities on each pair of vertices.

  13. contains the edge with probability for every pair independently. If an edge , we assign it with weight . It is easy to verify that E We will carefully choose to guarantee that is sparse with high probability; is well-concentrated to its expectation. Algorithms for Big Data (XIV) 5/12 Let H = ( V, E H ) be the sparse graph we are going to construct…

  14. If an edge , we assign it with weight . It is easy to verify that E We will carefully choose to guarantee that is sparse with high probability; is well-concentrated to its expectation. Algorithms for Big Data (XIV) 5/12 Let H = ( V, E H ) be the sparse graph we are going to construct… H contains the edge { u, v } with probability p u,v for every pair { u, v } independently.

  15. It is easy to verify that E We will carefully choose to guarantee that is sparse with high probability; is well-concentrated to its expectation. Algorithms for Big Data (XIV) 5/12 Let H = ( V, E H ) be the sparse graph we are going to construct… H contains the edge { u, v } with probability p u,v for every pair { u, v } independently. If an edge { u, v } ∈ E H , we assign it with weight w u,v /p u,v .

  16. It is easy to verify that We will carefully choose to guarantee that is sparse with high probability; is well-concentrated to its expectation. Algorithms for Big Data (XIV) 5/12 Let H = ( V, E H ) be the sparse graph we are going to construct… H contains the edge { u, v } with probability p u,v for every pair { u, v } independently. If an edge { u, v } ∈ E H , we assign it with weight w u,v /p u,v . E [ L H ] = L G .

  17. is sparse with high probability; It is easy to verify that is well-concentrated to its expectation. Algorithms for Big Data (XIV) 5/12 Let H = ( V, E H ) be the sparse graph we are going to construct… H contains the edge { u, v } with probability p u,v for every pair { u, v } independently. If an edge { u, v } ∈ E H , we assign it with weight w u,v /p u,v . E [ L H ] = L G . We will carefully choose { p u,v } to guarantee that

  18. Algorithms for Big Data (XIV) It is easy to verify that 5/12 Let H = ( V, E H ) be the sparse graph we are going to construct… H contains the edge { u, v } with probability p u,v for every pair { u, v } independently. If an edge { u, v } ∈ E H , we assign it with weight w u,v /p u,v . E [ L H ] = L G . We will carefully choose { p u,v } to guarantee that ▶ H is sparse with high probability; ▶ L H is well-concentrated to its expectation.

  19. is the projection onto the column space of A Transformation Sometimes it is more convenient to work with , the pseudo-inverse of . Note that The matrix . We will now study . Algorithms for Big Data (XIV) 6/12

  20. is the projection onto the column space of A Transformation Note that The matrix . We will now study . Algorithms for Big Data (XIV) 6/12 Sometimes it is more convenient to work with L + G , the pseudo-inverse of L G .

  21. is the projection onto the column space of A Transformation The matrix Algorithms for Big Data (XIV) . We will now study . 6/12 Note that Sometimes it is more convenient to work with L + G , the pseudo-inverse of L G . ⇒ L + /2 G L H L + /2 ≼ ( 1 + ε ) L + /2 G L G L + /2 L H ≼ ( 1 + ε ) L G ⇐ G . G

  22. A Transformation We will now study Algorithms for Big Data (XIV) . 6/12 Note that Sometimes it is more convenient to work with L + G , the pseudo-inverse of L G . ⇒ L + /2 G L H L + /2 ≼ ( 1 + ε ) L + /2 G L G L + /2 L H ≼ ( 1 + ε ) L G ⇐ G . G The matrix L + /2 G L G L + /2 is the projection onto the column space of L G . G

  23. A Transformation Algorithms for Big Data (XIV) 6/12 Note that Sometimes it is more convenient to work with L + G , the pseudo-inverse of L G . ⇒ L + /2 G L H L + /2 ≼ ( 1 + ε ) L + /2 G L G L + /2 L H ≼ ( 1 + ε ) L G ⇐ G . G The matrix L + /2 G L G L + /2 is the projection onto the column space of L G . G We will now study L + /2 G L H L + /2 G .

  24. Chernoff Bound for Matrices and maximum eigenvalues of E Algorithms for Big Data (XIV) . , for Pr , and , for Pr respectively. Then be the minimum The main tool to establish concentration is the following analogue of Chernofg bound for and . Let almost surely. Let that be independent random positive semi-definite matrices such Let Theorem matrices. 7/12

  25. Chernoff Bound for Matrices and maximum eigenvalues of E Algorithms for Big Data (XIV) . , for Pr , and , for Pr respectively. Then be the minimum The main tool to establish concentration is the following analogue of Chernofg bound for and . Let almost surely. Let that be independent random positive semi-definite matrices such Let Theorem matrices. 7/12

  26. Chernoff Bound for Matrices The main tool to establish concentration is the following analogue of Chernofg bound for Algorithms for Big Data (XIV) 7/12 matrices. Theorem Let X 1 , . . . , X n ∈ R n × n be independent random positive semi-definite matrices such that λ max ( X i ) ≤ R almost surely. Let X = ∑ n i = 1 X i . Let µ min and µ max be the minimum and maximum eigenvalues of E [ X ] respectively. Then ) µ min /R ( ▶ Pr [ λ min ( X ) ≤ ( 1 − ε ) µ min ] ≤ n e − ε , for 0 < ε < 1 , and ( 1 − ε ) 1 − ε ) µ max /R ( e ε ▶ Pr [ λ max ( X ) ≥ ( 1 + ε ) µ max ] ≤ n , for ε > 0 . ( 1 + ε ) 1 + ε

  27. For every pair of vertices and , we define Following our construction of , for every , define a random variable w.p. otherwise. Then and max Algorithms for Big Data (XIV) 8/12 Setting p u,v

  28. Following our construction of , for every , define a random variable w.p. otherwise. Then and max Algorithms for Big Data (XIV) 8/12 Setting p u,v For every pair of vertices u and v , we define p u,v ≜ 1 Rw u,v ∥ L + /2 G L u,v L + /2 G ∥ .

  29. 8/12 otherwise. Algorithms for Big Data (XIV) max and Then Setting p u,v For every pair of vertices u and v , we define p u,v ≜ 1 Rw u,v ∥ L + /2 G L u,v L + /2 G ∥ . Following our construction of H , for every { u, v } , define a random variable { ( w u,v /p u,v ) L + /2 G L u,v L + /2 G , w.p. p u,v X u,v = 0,

Recommend


More recommend