1 5 the curse of dimensionality many applications require
play

. . . 1 / 5 The curse of dimensionality . many applications - PowerPoint PPT Presentation

The curse of dimensionality . . . 1 / 5 The curse of dimensionality . many applications require high dimensional data . . 1 / 5 The curse of dimensionality . many applications require high dimensional data many algorithms become


  1. The curse of dimensionality . . . 1 / 5

  2. The curse of dimensionality . many applications require high dimensional data . . 1 / 5

  3. The curse of dimensionality . many applications require high dimensional data many algorithms become inefficient with high dimensional . . 1 / 5

  4. The curse of dimensionality . many applications require high dimensional data many algorithms become inefficient with high dimensional like to replace high dimensional data by smaller dimensional data without losing too much information . . 1 / 5

  5. The curse of dimensionality . many applications require high dimensional data many algorithms become inefficient with high dimensional like to replace high dimensional data by smaller dimensional data without losing too much information see two techniques for this task . . 1 / 5

  6. The curse of dimensionality . many applications require high dimensional data many algorithms become inefficient with high dimensional like to replace high dimensional data by smaller dimensional data without losing too much information see two techniques for this task 1 Johnson-Lindenstrauss lemma . . 1 / 5

  7. The curse of dimensionality . many applications require high dimensional data many algorithms become inefficient with high dimensional like to replace high dimensional data by smaller dimensional data without losing too much information see two techniques for this task 1 Johnson-Lindenstrauss lemma 2 singular value decomposition / principal component analysis another technique is feature selection . . 1 / 5

  8. The Johnson-Lindenstrauss lemma . . Theorem 5.1 . Let P be a set of n points in R d and 0 < ϵ < 1 . Then, for c large enough, there is an embedding π : P → R c log( n ) /ϵ 2 , such that for all p , q ∈ P (1 − ϵ ) · D l 2 ( p , q ) ≤ D l 2 ( π ( p ) , π ( q )) ≤ (1 + ϵ ) · D l 2 ( p , q ) . . . . 2 / 5

  9. The Johnson-Lindenstrauss lemma - the construction . . Gaussian distribution . µ ∈ R , σ ∈ R > 0 density function N ( ·| µ, σ 2 ) : R → R > 0 2 πσ 2 · exp( − ( x − µ ) 2 1 N ( x | µ, σ 2 ) �→ √ ) 2 σ 2 distribution with density function N ( · | µ, σ 2 ) called Gaussian or normal distribution N ( µ, σ 2 ) with mean µ and standard deviation σ ,i.e. ∫ l N ( x | µ, σ 2 ) d x . ∀ l ∈ R : Pr[ x ≤ l ] = −∞ . . . 3 / 5

  10. The Johnson-Lindenstrauss lemma - the construction . . Density function of Gaussian distribution . . . . 4 / 5

  11. . . . The Johnson-Lindenstrauss lemma - the construction . . Random mapping . A = ( r ij ) 1 ≤ i ≤ k , 1 ≤ j ≤ d ∈ R k × d , where each r ij is chosen according to N (0 , 1). ∀ x ∈ R d : π A ( x ) = 1 k · A · x . √ . . . 5 / 5

  12. The Johnson-Lindenstrauss lemma - the construction . . Random mapping . A = ( r ij ) 1 ≤ i ≤ k , 1 ≤ j ≤ d ∈ R k × d , where each r ij is chosen according to N (0 , 1). ∀ x ∈ R d : π A ( x ) = 1 k · A · x . √ . . Lemma 5.2 . Let π A : R d → R k be a chosen as above, let u ∈ R d be a vector, and let 0 < ϵ < 1 . Then, for c large enough and k = c · log( n ) /ϵ 2 : [ (1 − ϵ ) ≤ ∥ π A ( u ) ∥ 2 ] 1 Pr ≤ (1 + ϵ ) ≥ 1 − 3 n 2 . ∥ u ∥ 2 . . . 5 / 5

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend