a minimization algorithm
play

A Minimization Algorithm Consider the minimization problem: * M - PowerPoint PPT Presentation

A Minimization Algorithm Consider the minimization problem: * M min M M * subject to (i, 2 (M(i, j) j) ) (i, j) There are many techniques to solve this problem


  1. A Minimization Algorithm • Consider the minimization problem:  * M min M M * subject to     Γ(i, 2 (M(i, j) j) )   (i, j) • There are many techniques to solve this problem (http://perception.csl.illinois.edu/matrix- rank/sample_code.html) • Out of these, we will study one method called “singular value thresholding ”.

  2. Ref: Cai et al, A singular value thresholding algorithm for matrix completion , SIAM Journal on Optimization, 2010. Singular Value Thresholding (SVT) ˆ           * n n SVT ( , 0 ) Y soft threshold ( Y R ; ) 1 2 { {     ( 0 ) n n T Y 0 R Y USV ( using svd ) 1 2   for ( k 1 : rank ( Y )) k 1 while(conv ergence criterion not met) {    { S ( k , k ) max( 0 , S ( k , k ) );      The soft- } ( k ) ( k 1 ) soft threshold ( Y ; ) thresholding  rank ( Y )         ( k ) ( k 1 ) ( k ) procedure obeys the  Y Y P ( ); k k 1 ; ˆ  t  Y S ( k , k ) u v k following property k k } (which we state w/o  i 1 proof). }    * ( k ) ; }    soft threshold ( Y ; ) 1 2    arg min X Y X X F * 2

  3. Properties of SVT (stated w/o proof) • The sequence {  k } converges to the true solution of the problem below provided the step-sizes {  k } all lie between 0 and 2. 2    * M min M 0 . 5 M M * F subject to     Γ ( i , j ) , M ( i , j ) ( i , j ) • For large values of  , this converges to the solution of the original problem (i.e. without the Frobenius norm term).

  4. Properties of SVT (stated w/o proof) • The matrices {  k } turn out to have low rank (empirical observation – proof not established). • The matrices { Y k } also turn out to be sparse (empirical observation – rigorous proof not established). • The SVT step does not require computation of full SVD – we need only those singular vectors whose singular values exceed τ . There are special iterative methods for that.

  5. Results • The SVT algorithm works very efficiently and is easily implementable in MATLAB. • The authors report reconstruction of a 30,000 by 30,000 matrix in just 17 minutes on a 1.86 GHz dual-core desktop with 3 GB RAM and with MATLAB’s multithreading option enabled.

  6. Results (Data without noise) https://arxiv.org/abs/0810.3286

  7. Results (Noisy Data) https://arxiv.org/abs/0810.3286

  8. Results on real data • Dataset consists of a matrix M of geodesic distances between 312 cities in the USA/Canada. • This matrix is of approximately low-rank (in fact, the relative Frobenius error between M and its rank-3 approximation is 0.1159). • 70% of the entries of this matrix (chosen uniformly at random) were blanked out.

  9. Results on real data https://arxiv.org/abs/0810.3286

  10. Algorithm for Robust PCA • The algorithm uses the augmented Lagrangian technique. • See https://en.wikipedia.org/wiki/Augmented_Lag rangian_method and https://www.him.uni- bonn.de/fileadmin/him/Section6_HIM_v1.pdf • Suppose you want to solve: min f ( x ) w.r.t. x    s.t. i I , c ( x ) 0 i

  11. Algorithm for Robust PCA • Suppose you want to solve: min f ( x ) w.r.t. x    s.t. i I , c ( x ) 0 i • The augmented Lagrangian method (ALM) adopts the following iterative updates:        2 x arg min f ( x ) c ( x ) c ( x ) k x k i i i   i I i I      c ( x ) i i k i k Augmentation term Lagrangian term

  12. ALM: Some intuition • What is the intuition behind the update of the Lagrange parameters { λ i }? • The problem is:  λ t min max f ( x ) ( x ) c min f ( x ) λ x     ( x ) ( c ( x ), c ( x ),..., c ( x )) s.t. i I, c ( x ) 0 c 1 2 | I | i The maximum w.r.t. λ will be ∞ unless the constraint is satisfied. Hence these problems are equivalent.

  13. ALM: Some intuition • The problem is:  λ t min f ( x ) min max f ( x ) ( x ) c λ x     s.t. i I, c ( x ) 0 ( x ) ( c ( x ), c ( x ),..., c ( x )) c i 1 2 | I | Due to non-smoothness of the max function, the equivalence has little computational benefit. We smooth it by adding another term that penalizes deviations from a prior estimate of the λ parameters. 2 λ λ -    λ λ ( x ) c   λ t min max f ( x ) ( x ) c λ  x 2 Maximization w.r.t. λ is now easy

  14. ALM: Some inutuion – inequality constraints  min f ( x ) λ t min max f ( x ) ( x ) c  λ x 0     s.t. i I, c ( x ) 0 ( x ) ( c ( x ), c ( x ),..., c ( x )) c i 1 2 | I | 2 λ λ -    λ λ max( ( x ), 0 ) c   λ t min max f ( x ) ( x ) c λ  x 2 Maximization w.r.t. λ is now easy

  15. Theorem 1 (Informal Statement) • Consider a matrix M of size n 1 by n 2 which is the sum of a “sufficiently low - rank” component L and a “sufficiently sparse” component S whose support is uniformly randomly distributed in the entries of M . • Then the solution of the following optimization problem (known as principal component pursuit ) yields exact estimates of L and S with “very high” probability: 1   E ( L ' , S ' ) min L S ( L , S ) * 1 max( n , n ) 1 2 This is a convex   subject to L S M . optimization problem. n n  1 2  Note : S | S | ij 1   i 1 j 1

  16. Algorithm for Robust PCA • In our case, we seek to optimize: • Basic algorithm: Lagrange matrix       ( L , S ) arg min l ( L , S , Y ), Y Y ( M L S )  k k ( L , S ) k k 1 k k k Update of S using soft-thresholding Update of L using singular-value soft-thresholding

  17. Alternating Minimization Algorithm for Robust PCA

  18. https://statweb.stanford.edu/~candes/papers/RobustPCA.pdf Results

  19. (Compressive) Low Rank Matrix Recovery

  20. Compressive RPCA: Algorithm and an Application Primarily based on the paper: Waters et al, “ SpaRCS: Recovering Low-Rank and Sparse Matrices from Compressive Measurements ”, NIPS 2011

  21. Problem statement • Let M be a matrix which is the sum of low rank matrix L and sparse matrix S . • We observed compressive measurements of M in the following form: A         n n n n m ( ), R , R R , m n n y L S L S , y 1 2 1 2 1 2 A  linear operator acting/map on M A Retrieve , given , L S y

  22. Scenarios • M could be a matrix representing a video – each column of M is a vectorized frame from the video. • M could also be a matrix representing a hyperspectral image – each column is the vectorized form of a slice at a given wavelength. • Robust Matrix completion – a special form of a compressive L + S recovery problem.

  23. Objective function: SpaRCS Free parameters SpaRCS = sparse and low rank decomposition via compressive sampling

  24. SparCS Algorithm https://papers.nips.cc/pap er/4438-sparcs-recovering- low-rank-and-sparse- matrices-from- compressive- measurements.pdf Very simple to implement; but requires tuning of K , r parameters; convergence guarantees not established.

  25. Results: Phase transition https://papers.nips.cc/paper/4438-sparcs- recovering-low-rank-and-sparse-matrices- from-compressive-measurements.pdf Code: https://www.ece.rice.edu/~aew2/sparcs.html

  26. Results: Video CS Follows Rice SPC model, independent compressive measurements on each frame of the matrix M representing the video. https://papers.nips.cc/paper/4438-sparcs- recovering-low-rank-and-sparse-matrices- from-compressive-measurements.pdf

  27. Results: Video CS Follows Rice SPC model, independent compressive measurements on each frame of the matrix M representing the video. https://papers.nips.cc/paper/4438-sparcs- recovering-low-rank-and-sparse-matrices- from-compressive-measurements.pdf

  28. Results: Hyperspectral CS Rice SPC model of CS measurements on https://papers.nips.cc/paper/4438-sparcs-recovering-low-rank-and-sparse-matrices- from-compressive-measurements.pdf every spectral band

  29. Results: Robust matrix completion https://papers.nips.cc/paper/4438-sparcs-recovering-low-rank-and-sparse- matrices-from-compressive-measurements.pdf

  30. Theorem for Compressive PCP Q is obtained from the linear span of different independent N(0,1) matrices with iid entries Wright et al, “Compressive Principal Component Pursuit” http://yima.csl.illinois.edu/psfile/CPCP.pdf

  31. Summary • Low rank matrix completion: motivation, key theorems, numerical results • Algorithm for low rank matrix completion • Robust PCA • (Compressive) low rank matrix recovery • Compressive RPCA • Several papers linked on moodle

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend