tensor clustering and error bounds
play

Tensor Clustering and Error Bounds Chris Ding Department of - PowerPoint PPT Presentation

Tensor Clustering and Error Bounds Chris Ding Department of Computer Science and Engineering University of Texas, Arlington Joint work with Heng Huang and Dijun Luo Work Supported by NSF CISE/DMS C. Ding, Matrix-model Machine Learning 1


  1. Tensor Clustering and Error Bounds Chris Ding Department of Computer Science and Engineering University of Texas, Arlington Joint work with Heng Huang and Dijun Luo Work Supported by NSF CISE/DMS C. Ding, Matrix-model Machine Learning 1

  2. Tensors • The word tensor is used in 1900 (time of A. Einstein) in physics � General relativity is entirely written in tensor format – Physicists see tensor and think of coordinate transformation properties – Computer scientists see tensor and wants to compute them faster C. Ding, Tensor Clustering 2

  3. Tensor Decompositions: Main new results • Two main tensor decompositions – ParaFac (CanDecomp, rank-1) – HOSVD (Tucker-3) • Data clustering – ParaFac does simultaneous compression and K-means clustering • Cluster centrods are rank-1 matrices: – HOSVD does simultaneous compression and K-means clustering • Cluster centroids are of the type: • Eckart-Young type lower and upper error bounds – ParaFac – HOSVD • Extensive experiments C. Ding, Matrix-model Machine Learning 3

  4. ParaFac Objective Function • ParaFac is the simplest and most widely used model C. Ding, Matrix-model Machine Learning 4

  5. Bounds on ParaFac Reconstruction Error Eckart-Young type Error bounds: = C. Ding, Matrix-model Machine Learning 5

  6. Outline of the Upper Error Bounds • In standard ParaFac, columns of W is only required to be linearly independent – We study W-orthogonal ParaFac where W is required to be orthogonal . – Upper bound is obtained because the domain is further restricted. – Any feasible solution of W-orthogonal ParaFac gives an upper bound. • W-orthogonal ParaFac can be reduction [(U,V,W) to W-only] – C. Ding, Matrix-model Machine Learning 6

  7. Outline of the Lower Error Bound Increasing the domain of variables � more • accurate approximation � lower bound In ParaFac decomposition : – We replace Lower bound : C. Ding, Matrix-model Machine Learning 7

  8. Experiments on ParaFac Error Bounds C. Ding, Matrix-model Machine Learning 8

  9. High Order SVD (HOSVD) • Initially called Tucker-3 Decomposition • HOSVD uses 3 factors and a core tensor S: U, V, W, S are obtained by minimizing the reconstruction error C. Ding, Tensor Clustering 9

  10. HOSVD Error Bounds HOSVD U = eigenvectors(F), V=eigenvectors(G ) C. Ding, Matrix-model Machine Learning 10

  11. Outline of the Upper Error Bound We need to find a feasible solution, which gives an upper bound C. Ding, Matrix-model Machine Learning 11

  12. Outline of the Upper Error Bound All these are T1 decompositions and trivially solved. C. Ding, Matrix-model Machine Learning 12

  13. Outline of the Upper Error Bound C. Ding, Matrix-model Machine Learning 13

  14. Compute eigenvalues and use the error bounds to determine HOSVD/ParaFac parameters C. Ding, Matrix-model Machine Learning 14

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend