adaptive sparse recovery
play

Adaptive Sparse Recovery Eric Price MIT 2012-04-26 Joint work - PowerPoint PPT Presentation

Adaptive Sparse Recovery Eric Price MIT 2012-04-26 Joint work with Piotr Indyk and David Woodruff, 2011-2012 Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 1 / 29 Outline Motivating Example 1 Eric Price (MIT) Adaptive Sparse


  1. General Compressive Sensing Want to observe n -dimensional vector x ◮ Which of n people have a genetic mutation. ◮ Image ◮ Traffic pattern of packets on a network. Normally takes n observations to find. But we know some structure on the input: ◮ Genetics: most people don’t have the mutation. ◮ Images: mostly smooth with some edges. ◮ Traffic: Zipf distribution. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 8 / 29

  2. General Compressive Sensing Want to observe n -dimensional vector x ◮ Which of n people have a genetic mutation. ◮ Image ◮ Traffic pattern of packets on a network. Normally takes n observations to find. But we know some structure on the input: ◮ Genetics: most people don’t have the mutation. ◮ Images: mostly smooth with some edges. ◮ Traffic: Zipf distribution. We use this structure to compress space (e.g. JPEG). Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 8 / 29

  3. General Compressive Sensing Want to observe n -dimensional vector x ◮ Which of n people have a genetic mutation. ◮ Image ◮ Traffic pattern of packets on a network. Normally takes n observations to find. But we know some structure on the input: ◮ Genetics: most people don’t have the mutation. ◮ Images: mostly smooth with some edges. ◮ Traffic: Zipf distribution. We use this structure to compress space (e.g. JPEG). Can we use structure to save on observations ? Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 8 / 29

  4. Cameras 5 megapixel camera takes 15 million byte-size observations. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 9 / 29

  5. Cameras 5 megapixel camera takes 15 million byte-size observations. Compresses it (JPEG) down to a million bytes. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 9 / 29

  6. Cameras 5 megapixel camera takes 15 million byte-size observations. Compresses it (JPEG) down to a million bytes. Why do we need to bother with so many observations? [Donoho,Cand` es-Tao] Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 9 / 29

  7. Cameras 5 megapixel camera takes 15 million byte-size observations. Compresses it (JPEG) down to a million bytes. Why do we need to bother with so many observations? [Donoho,Cand` es-Tao] Cheap in visible light (silicon), very expensive in infrared. ◮ $30,000 for 256x256 IR sensor. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 9 / 29

  8. Cameras 5 megapixel camera takes 15 million byte-size observations. Compresses it (JPEG) down to a million bytes. Why do we need to bother with so many observations? [Donoho,Cand` es-Tao] Cheap in visible light (silicon), very expensive in infrared. ◮ $30,000 for 256x256 IR sensor. Use structure to take only a few million observations. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 9 / 29

  9. Cameras 5 megapixel camera takes 15 million byte-size observations. Compresses it (JPEG) down to a million bytes. Why do we need to bother with so many observations? [Donoho,Cand` es-Tao] Cheap in visible light (silicon), very expensive in infrared. ◮ $30,000 for 256x256 IR sensor. Use structure to take only a few million observations. ◮ What structure? Sparsity. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 9 / 29

  10. Sparsity A vector is k -sparse if k components are non-zero. Images are almost sparse in the wavelet basis: Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 10 / 29

  11. Sparsity A vector is k -sparse if k components are non-zero. Images are almost sparse in the wavelet basis: Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 10 / 29

  12. Sparsity A vector is k -sparse if k components are non-zero. Images are almost sparse in the wavelet basis: Same kind of structure as in genetic testing! Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 10 / 29

  13. Linear sketching/Compressive sensing Suppose an n -dimensional vector x is k -sparse in known basis. Given Ax , a set of m << n linear products. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 11 / 29

  14. Linear sketching/Compressive sensing Suppose an n -dimensional vector x is k -sparse in known basis. Given Ax , a set of m << n linear products. Why linear? Many applications: ◮ Genetic testing: mixing blood samples. ◮ Streaming updates: A ( x + ∆ ) = Ax + A ∆ . ◮ Camera optics: filter in front of lens. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 11 / 29

  15. Linear sketching/Compressive sensing Suppose an n -dimensional vector x is k -sparse in known basis. Given Ax , a set of m << n linear products. Why linear? Many applications: ◮ Genetic testing: mixing blood samples. ◮ Streaming updates: A ( x + ∆ ) = Ax + A ∆ . ◮ Camera optics: filter in front of lens. Then it is possible to recover x from Ax . ◮ Quickly ◮ Robustly: get close to x if x is close to k -sparse. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 11 / 29

  16. Linear sketching/Compressive sensing Suppose an n -dimensional vector x is k -sparse in known basis. Given Ax , a set of m << n linear products. Why linear? Many applications: ◮ Genetic testing: mixing blood samples. ◮ Streaming updates: A ( x + ∆ ) = Ax + A ∆ . ◮ Camera optics: filter in front of lens. Then it is possible to recover x from Ax . ◮ Quickly ◮ Robustly: get close to x if x is close to k -sparse. Note: impossible without using sparsity ( A is underdetermined). Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 11 / 29

  17. Standard Sparse Recovery Framework Specify distribution on m × n matrices A (independent of x ). Given linear sketch Ax , recover ˆ x . Satisfying the recovery guarantee: � ˆ x − x � 2 � ( 1 + ǫ ) min � x − x k � 2 k -sparse x k with probability 2 / 3. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 12 / 29

  18. Standard Sparse Recovery Framework Specify distribution on m × n matrices A (independent of x ). Given linear sketch Ax , recover ˆ x . Satisfying the recovery guarantee: � ˆ x − x � 2 � ( 1 + ǫ ) min � x − x k � 2 k -sparse x k with probability 2 / 3. Solvable with O ( 1 ǫ k log n k ) measurements [Cand` es-Romberg-Tao ’06, Gilbert-Li-Porat-Strauss ’10] Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 12 / 29

  19. Standard Sparse Recovery Framework Specify distribution on m × n matrices A (independent of x ). Given linear sketch Ax , recover ˆ x . Satisfying the recovery guarantee: � ˆ x − x � 2 � ( 1 + ǫ ) min � x − x k � 2 k -sparse x k with probability 2 / 3. Solvable with O ( 1 ǫ k log n k ) measurements [Cand` es-Romberg-Tao ’06, Gilbert-Li-Porat-Strauss ’10] Matching lower bound. [Do Ba-Indyk-P-Woodruff ’10, P-Woodruff ’11] Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 12 / 29

  20. Adaptive Sparse Recovery Framework For i = 1 . . . r : ◮ Choose matrix A i based on previous observations (possibly randomized). ◮ Observe A i x . ◮ Number of measurements m is total number of rows in all A i . ◮ Number of rounds is r . Given linear sketch Ax , recover ˆ x . Satisfying the recovery guarantee: � ˆ x − x � 2 � ( 1 + ǫ ) min � x − x k � 2 k -sparse x k with probability 2 / 3. Solvable with O ( 1 ǫ k log n k ) measurements [Cand` es-Romberg-Tao ’06, Gilbert-Li-Porat-Strauss ’10] Matching lower bound. [Do Ba-Indyk-P-Woodruff ’10, P-Woodruff ’11] Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 12 / 29

  21. Adaptive result in comparison to previous work Nonadaptive: Θ ( 1 ǫ k log n k ) . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 13 / 29

  22. Adaptive result in comparison to previous work Nonadaptive: Θ ( 1 ǫ k log n k ) . Adaptive: O ( k log n k ) with ǫ = o ( 1 ) ([Haupt-Baraniuk-Castro-Nowak ’09], in a slightly different setting) Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 13 / 29

  23. Adaptive result in comparison to previous work Nonadaptive: Θ ( 1 ǫ k log n k ) . Adaptive: O ( k log n k ) with ǫ = o ( 1 ) ([Haupt-Baraniuk-Castro-Nowak ’09], in a slightly different setting) This talk: O ( 1 ǫ k log log n k ) . [Indyk-P-Woodruff ’11] Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 13 / 29

  24. Adaptive result in comparison to previous work Nonadaptive: Θ ( 1 ǫ k log n k ) . Adaptive: O ( k log n k ) with ǫ = o ( 1 ) ([Haupt-Baraniuk-Castro-Nowak ’09], in a slightly different setting) This talk: O ( 1 ǫ k log log n k ) . [Indyk-P-Woodruff ’11] ◮ Using r = O ( log log n log ∗ k ) rounds. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 13 / 29

  25. Adaptive result in comparison to previous work Nonadaptive: Θ ( 1 ǫ k log n k ) . Adaptive: O ( k log n k ) with ǫ = o ( 1 ) ([Haupt-Baraniuk-Castro-Nowak ’09], in a slightly different setting) This talk: O ( 1 ǫ k log log n k ) . [Indyk-P-Woodruff ’11] ◮ Using r = O ( log log n log ∗ k ) rounds. Even when r = 2, can get O ( k log n + 1 ǫ k log ( k /ǫ )) Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 13 / 29

  26. Adaptive result in comparison to previous work Nonadaptive: Θ ( 1 ǫ k log n k ) . Adaptive: O ( k log n k ) with ǫ = o ( 1 ) ([Haupt-Baraniuk-Castro-Nowak ’09], in a slightly different setting) This talk: O ( 1 ǫ k log log n k ) . [Indyk-P-Woodruff ’11] ◮ Using r = O ( log log n log ∗ k ) rounds. Even when r = 2, can get O ( k log n + 1 ǫ k log ( k /ǫ )) ◮ Separating dependence on n and ǫ . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 13 / 29

  27. Applications of Adaptivity When does adaptivity make sense? Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  28. Applications of Adaptivity When does adaptivity make sense? Genetic testing: Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  29. Applications of Adaptivity When does adaptivity make sense? Genetic testing: ◮ Yes , but multiple rounds can be costly. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  30. Applications of Adaptivity When does adaptivity make sense? Genetic testing: ◮ Yes , but multiple rounds can be costly. Cameras: Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  31. Applications of Adaptivity When does adaptivity make sense? Genetic testing: ◮ Yes , but multiple rounds can be costly. Cameras: ◮ Programmable pixels (mirrors or LCD display): Yes . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  32. Applications of Adaptivity When does adaptivity make sense? Genetic testing: ◮ Yes , but multiple rounds can be costly. Cameras: ◮ Programmable pixels (mirrors or LCD display): Yes . ◮ Hardwired lens: No . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  33. Applications of Adaptivity When does adaptivity make sense? Genetic testing: ◮ Yes , but multiple rounds can be costly. Cameras: ◮ Programmable pixels (mirrors or LCD display): Yes . ◮ Hardwired lens: No . Streaming algorithms: Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  34. Applications of Adaptivity When does adaptivity make sense? Genetic testing: ◮ Yes , but multiple rounds can be costly. Cameras: ◮ Programmable pixels (mirrors or LCD display): Yes . ◮ Hardwired lens: No . Streaming algorithms: ◮ Adaptivity corresponds to multiple passes. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  35. Applications of Adaptivity When does adaptivity make sense? Genetic testing: ◮ Yes , but multiple rounds can be costly. Cameras: ◮ Programmable pixels (mirrors or LCD display): Yes . ◮ Hardwired lens: No . Streaming algorithms: ◮ Adaptivity corresponds to multiple passes. ◮ Router finding most common connections: No . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  36. Applications of Adaptivity When does adaptivity make sense? Genetic testing: ◮ Yes , but multiple rounds can be costly. Cameras: ◮ Programmable pixels (mirrors or LCD display): Yes . ◮ Hardwired lens: No . Streaming algorithms: ◮ Adaptivity corresponds to multiple passes. ◮ Router finding most common connections: No . ◮ Mapreduce finding most frequent URLs: Yes . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 14 / 29

  37. Outline Motivating Example 1 Formal Introduction to Sparse Recovery/Compressive Sensing 2 Algorithm 3 Conclusion 4 Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 15 / 29

  38. Outline of Algorithm Theorem Adaptive 1 + ǫ -approximate k-sparse recovery is possible with O ( 1 ǫ k log log ( n / k )) measurements. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 16 / 29

  39. Outline of Algorithm Theorem Adaptive 1 + ǫ -approximate k-sparse recovery is possible with O ( 1 ǫ k log log ( n / k )) measurements. Lemma Adaptive O ( 1 ) -approximate 1 -sparse recovery is possible with O ( log log n ) measurements. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 16 / 29

  40. Outline of Algorithm Theorem Adaptive 1 + ǫ -approximate k-sparse recovery is possible with O ( 1 ǫ k log log ( n / k )) measurements. Lemma Adaptive O ( 1 ) -approximate 1 -sparse recovery is possible with O ( log log n ) measurements. Lemma implies theorem using standard tricks ([GLPS10]). Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 16 / 29

  41. 1-sparse recovery: non-adaptive lower bound Lemma Adaptive C-approximate 1 -sparse recovery is possible with O ( log log n ) measurements for some C = O ( 1 ) . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 17 / 29

  42. 1-sparse recovery: non-adaptive lower bound Lemma Adaptive C-approximate 1 -sparse recovery is possible with O ( log log n ) measurements for some C = O ( 1 ) . Non-adaptive lower bound: why is this hard? Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 17 / 29

  43. 1-sparse recovery: non-adaptive lower bound Lemma Adaptive C-approximate 1 -sparse recovery is possible with O ( log log n ) measurements for some C = O ( 1 ) . Non-adaptive lower bound: why is this hard? Hard case: x is random e i plus Gaussian noise w . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 17 / 29

  44. 1-sparse recovery: non-adaptive lower bound Lemma Adaptive C-approximate 1 -sparse recovery is possible with O ( log log n ) measurements for some C = O ( 1 ) . Non-adaptive lower bound: why is this hard? Hard case: x is random e i plus Gaussian noise w . Noise � w � 2 2 = Θ ( 1 ) so C -approximate recovery requires finding i . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 17 / 29

  45. 1-sparse recovery: non-adaptive lower bound Lemma Adaptive C-approximate 1 -sparse recovery is possible with O ( log log n ) measurements for some C = O ( 1 ) . Non-adaptive lower bound: why is this hard? Hard case: x is random e i plus Gaussian noise w . Noise � w � 2 2 = Θ ( 1 ) so C -approximate recovery requires finding i . Observations � v , x � = v i + � v , w � = v i + � v � 2 √ n z , for z ∼ N ( 0 , Θ ( 1 )) . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 17 / 29

  46. 1-sparse recovery: non-adaptive lower bound Observe � v , x � = v i + � v � 2 √ n z , where z ∼ N ( 0 , Θ ( 1 )) Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 18 / 29

  47. 1-sparse recovery: non-adaptive lower bound Observe � v , x � = v i + � v � 2 √ n z , where z ∼ N ( 0 , Θ ( 1 )) Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 18 / 29

  48. 1-sparse recovery: non-adaptive lower bound Observe � v , x � = v i + � v � 2 √ n z , where z ∼ N ( 0 , Θ ( 1 )) Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 18 / 29

  49. 1-sparse recovery: non-adaptive lower bound Observe � v , x � = v i + � v � 2 √ n z , where z ∼ N ( 0 , Θ ( 1 )) Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 18 / 29

  50. 1-sparse recovery: non-adaptive lower bound Observe � v , x � = v i + � v � 2 √ n z , where z ∼ N ( 0 , Θ ( 1 )) Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 18 / 29

  51. 1-sparse recovery: non-adaptive lower bound Observe � v , x � = v i + � v � 2 √ n z , where z ∼ N ( 0 , Θ ( 1 )) Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 18 / 29

  52. 1-sparse recovery: non-adaptive lower bound Observe � v , x � = v i + � v � 2 √ n z , where z ∼ N ( 0 , Θ ( 1 )) Information capacity I ( i , � v , x � ) � 1 2 log ( 1 + SNR ) where SNR denotes the “signal-to-noise ratio,” SNR = E [ signal 2 ] E [ v 2 i ] � 2 / n = 1 E [ noise 2 ] � v � 2 Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 18 / 29

  53. 1-sparse recovery: non-adaptive lower bound Observe � v , x � = v i + � v � 2 √ n z , where z ∼ N ( 0 , Θ ( 1 )) Information capacity I ( i , � v , x � ) � 1 2 log ( 1 + SNR ) where SNR denotes the “signal-to-noise ratio,” SNR = E [ signal 2 ] E [ v 2 i ] � 2 / n = 1 E [ noise 2 ] � v � 2 Finding i needs Ω ( log n ) non-adaptive measurements. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 18 / 29

  54. 1-sparse recovery: changes in adaptive setting Information capacity I ( i , � v , x � ) � 1 2 log ( 1 + SNR ) . where SNR denotes the “signal-to-noise ratio,” � � E [ v 2 i ] SNR = Θ . � v � 2 2 / n Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 19 / 29

  55. 1-sparse recovery: changes in adaptive setting Information capacity I ( i , � v , x � ) � 1 2 log ( 1 + SNR ) . where SNR denotes the “signal-to-noise ratio,” � � E [ v 2 i ] SNR = Θ . � v � 2 2 / n If i is independent of v , this is O ( 1 ) . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 19 / 29

  56. 1-sparse recovery: changes in adaptive setting Information capacity I ( i , � v , x � ) � 1 2 log ( 1 + SNR ) . where SNR denotes the “signal-to-noise ratio,” � � E [ v 2 i ] SNR = Θ . � v � 2 2 / n If i is independent of v , this is O ( 1 ) . As we learn about i , we can increase E [ v 2 i ] for constant � v � 2 . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 19 / 29

  57. 1-sparse recovery: changes in adaptive setting Information capacity I ( i , � v , x � ) � 1 2 log ( 1 + SNR ) . where SNR denotes the “signal-to-noise ratio,” � � E [ v 2 i ] SNR = Θ . � v � 2 2 / n If i is independent of v , this is O ( 1 ) . As we learn about i , we can increase E [ v 2 i ] for constant � v � 2 . ◮ Equivalently, for constant E [ v 2 i ] we can decrease � v � 2 . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 19 / 29

  58. 1-sparse recovery: idea x = e i + w Candidate set Signal 0 bits v SNR = 2 I ( i , � v , x � ) � log SNR = 1 � v , x � = v i + � v , w � Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 20 / 29

  59. 1-sparse recovery: idea x = e i + w Candidate set Signal 0 bits 1 bit v SNR = 2 2 I ( i , � v , x � ) � log SNR = 2 � v , x � = v i + � v , w � Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 20 / 29

  60. 1-sparse recovery: idea x = e i + w Candidate set Signal 0 bits 1 bit 2 bits v SNR = 2 4 I ( i , � v , x � ) � log SNR = 4 � v , x � = v i + � v , w � Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 20 / 29

  61. 1-sparse recovery: idea x = e i + w Candidate set Signal 0 bits 1 bit 2 bits 4 bits v SNR = 2 8 I ( i , � v , x � ) � log SNR = 8 � v , x � = v i + � v , w � Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 20 / 29

  62. 1-sparse recovery: idea x = e i + w Candidate set Signal 0 bits 1 bit 2 bits 4 bits 8 bits v SNR = 2 16 I ( i , � v , x � ) � log SNR = 16 � v , x � = v i + � v , w � Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 20 / 29

  63. Goal Shown intuition for specific distribution on x Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 21 / 29

  64. Goal Shown intuition for specific distribution on x Match previous convergence for arbitrary x = α e i + w ? Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 21 / 29

  65. Goal Shown intuition for specific distribution on x Match previous convergence for arbitrary x = α e i + w ? ◮ α may not be 1. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 21 / 29

  66. Goal Shown intuition for specific distribution on x Match previous convergence for arbitrary x = α e i + w ? ◮ α may not be 1. ◮ Work for a specific x with 3 / 4 probability. Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 21 / 29

  67. Goal Shown intuition for specific distribution on x Match previous convergence for arbitrary x = α e i + w ? ◮ α may not be 1. ◮ Work for a specific x with 3 / 4 probability. ◮ Distribution over A , for fixed w . Eric Price (MIT) Adaptive Sparse Recovery 2012-04-26 21 / 29

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend