image masking schemes for local manifold learning methods
play

Image Masking Schemes for Local Manifold Learning Methods Marco F. - PowerPoint PPT Presentation

Image Masking Schemes for Local Manifold Learning Methods Marco F. Duarte Joint work with Hamid Dadkhahi IEEE ICASSP - April 23 2015 Manifold Learning Given training points in , learn the mapping to the underlying K -dimensional


  1. Image Masking Schemes for 
 Local Manifold Learning Methods Marco F. Duarte Joint work with Hamid Dadkhahi IEEE ICASSP - April 23 2015

  2. Manifold Learning • Given training points in , learn the mapping � to the underlying K -dimensional articulation manifold • Exploit local geometry 
 to capture parameter 
 differences by embedding 
 distances • ISOMAP, LLE, HLLE, … • Ex: � images of � rotating teapot � � � articulation space � = circle �

  3. Compressive Manifold Learning • Given training points in , learn the mapping � to the underlying K -dimensional articulation manifold • Isomap algorithm approximates geodesic distances using distances between neighboring points • Random measurements preserve these distances • Theorem : If , then the Isomap 
 residual variance in the projected [Hegde, Wakin, 
 Baraniuk 2008] domain is bounded by the additive � error factor translating 
 disk manifold 
 N = 4096 (Full Data) M = 50 M = 25 M = 100 ( K =2)

  4. Custom Projection Operators • Goal of Dimensionality Reduction: To preserve distances between points in the manifold, i.e., for • Collect pairwise differences into set of 
 secant vectors • Search for projection that preserves norms of secants: [Hegde, Sankaranarayanan, Baraniuk 2012]

  5. Custom Projection Operators • Usual approach: Principal Component Analysis (PCA) • Collect all secants into a matrix: • Perform eigenvalue decomposition on S : • Select top eigenvectors as projections • PCA minimizes the average squared distortion over secants, but can distort individual secants arbitrarily and therefore warp manifold structure [Hegde, Sankaranarayanan, Baraniuk 2012]

  6. Custom Projection Operators: NuMax • For target distortion , find matrix featuring the smallest number of rows that yields • This is equivalent to minimizing the rank 
 of the matrix such that • Use nuclear norm as proxy for rank to 
 obtain computationally efficient approach • Improves over random projections since matrix is 
 specifically tailored to manifold observed • May be difficult to link target distortion to matrix rank/ 
 number of rows [Hegde, Sankaranarayanan, Baraniuk 2012]

  7. Issues with Randomness and NuMax • Projections matrices have entries 
 with arbitrary values • Physics of sensing process, 
 hardware devices restrict types 
 of projections we can obtain • Example: Low-power imaging 
 for computational eyeglasses • Low-power imaging sensor 
 allows for individual selection of 
 pixels to record • Power consumption proportional 
 to number of pixels sampled • Random projections/NuMax involve 
 half/all pixels and do not enable 
 power savings • How to derive constrained 
 projection matrices that involve 
 [Mayberry, Hu, Marlin, 
 only few pixels? Salthouse, Ganesan 2014]

  8. Masking Strategies for Manifold Data • Select only a subset of the pixels of size M that minimizes distortion to manifold structure • Emulate strategies for projection design into mask design • Random Masking : 
 Pick M pixels uniformly at random across image M = 100 
 pixels

  9. Masking Strategies for Manifold Data • Select only a subset of the pixels of size M that minimizes distortion to manifold structure • Emulate strategies for projection design into mask design • Principal coordinate analysis : 
 Pick M coordinates that maximize variance among secants M = 100 
 pixels [Dadkhahi and Duarte 2014]

  10. Masking Strategies for Manifold Data • Select only a subset of the pixels of size M that minimizes distortion to manifold structure • Emulate strategies for projection design into mask design • Adaptation of NuMax : • Define secants from k -nearest neighbor graph: 
 • Pick masking matrix (row submatrix of I ) to minimize secant norm distortion after scaling: 
 • Combinatorial integer program M = 100 
 replaced by greedy approximation pixels [Dadkhahi and Duarte 2014]

  11. Isomap vs. Locally Linear Embedding • While Isomap employs distances between neighbors when designing the embedding, Locally Linear Embedding (LLE) employs additional local geometry , 
 representing each vector as a weighted 
 linear combination of its neighbors 
 • In particular, Isomap embedding is 
 sensitive to scaling of the point clouds, 
 while LLE isn’t

  12. Isomap vs. Locally Linear Embedding • While Isomap employs distances between neighbors when designing the embedding, Locally Linear Embedding (LLE) employs additional local geometry , 
 representing each vector as a weighted 
 linear combination of its neighbors 
 • In particular, Isomap embedding is 
 sensitive to scaling of the point clouds, 
 while LLE isn’t

  13. Isomap vs. Locally Linear Embedding • While Isomap employs distances between neighbors when designing the embedding, Locally Linear Embedding (LLE) employs additional local geometry , 
 representing each vector as a weighted 
 linear combination of its neighbors 
 • In particular, Isomap embedding is 
 sensitive to scaling of the point clouds, 
 while LLE isn’t • To preserve this additional local 
 information, we expand the set of secants 
 to include distances between neighbors of 
 each point

  14. Isomap vs. Locally Linear Embedding • While Isomap employs distances between neighbors when designing the embedding, Locally Linear Embedding (LLE) employs additional local geometry , 
 representing each vector as a weighted 
 linear combination of its neighbors 
 • In particular, Isomap embedding is 
 sensitive to scaling of the point clouds, 
 while LLE isn’t • To preserve this additional local 
 information, we expand the set of secants 
 to include distances between neighbors of 
 each point

  15. 
 
 
 Manifold-Aware Pixel Selection for LLE • Expand the set of secants considered: 
 • Compute squared norms of secants 
 in obtained from the original 
 images and from masked images; 
 collect into “norm” vectors 
 • Choose mask that maximizes sum of 
 cosine similarities between original and 
 masked “norm” vectors: 
 • Replace combinatorial optimization by 
 greedy forward selection algorithm • Cosine similarity is invariant to (local) scaling of point cloud

  16. Manifold-Aware Pixel Selection for LLE M = 100 pixels M = 100 pixels

  17. Performance Analysis: 
 LLE Embedding Error Full Data Full Data 10 PCA PCA SPCA SPCA Embedding Error 8 PCoA PCoA MAPS − LLE MAPS − LLE 6 MAPS − Isomap MAPS − Isomap Random Random 4 2 50 100 150 200 250 300 Embedding Dim./Masking Size m M

  18. Performance Analysis: 
 LLE Embedding Error Full Data PCA 2500 SPCA Embedding Error PCoA 2000 MAPS − LLE MAPS − Isomap 1500 Random 1000 500 50 100 150 200 250 300 Embedding Dim./Masking Size m M

  19. Performance Analysis: 
 LLE Embedding Error Full Data 50 SPCA PCA PCoA SPCA Embedding Error 40 MAPS − LLE PCoA MAPS − Isomap MAPS − LLE Random 30 20 10 50 100 150 200 250 300 Embedding Dim./Masking Size m M

  20. Performance Analysis: 2-D LLE M = 100 pixels

  21. Computational Eyeglasses: Eye Gaze Tracking Average Gaze Estimation Error 45 Full Data PCA SPCA 40 PCoA MAPS − LLE 35 MAPS − Isomap Random 30 Error measured 
 25 in “target” pixels 50 100 150 200 250 300 Embedding Dim./Masking Size m M

  22. Conclusions • Compressive sensing (CS) for manifold-modeled images via random or customized projections (NuMax) • New sensors enable CS by masking images, 
 i.e., restricting the type of projections • Our MAPS algorithms find image masks that best preserve geometric structure used during manifold learning for image datasets • Greedy algorithms provide good preservation of learned manifold embeddings, suitable for parameter estimation • While Isomap relies on distances between neighbors, LLE also leverages local geometric structure; different algorithms are optimal for these cases • Concept of subsampling as feature selection - supervised and unsupervised learning? http://www.ecs.umass.edu/~mduarte mduarte@ecs.umass.edu

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend