matrix completion and matrix
play

Matrix Completion and Matrix Concentration Lester Mackey, Ameet - PowerPoint PPT Presentation

Matrix Completion and Matrix Concentration Lester Mackey, Ameet Talwalkar, Michael I. Jordan University of California, Berkeley Richard Chen, Brendan Farrell, Joel Tropp Caltech October 8, 2012 Part I Divide-Factor-Combine Jordan (UC


  1. Matrix Completion and Matrix Concentration Lester Mackey, Ameet Talwalkar, Michael I. Jordan University of California, Berkeley Richard Chen, Brendan Farrell, Joel Tropp Caltech October 8, 2012

  2. Part I Divide-Factor-Combine Jordan (UC Berkeley) Matrix Completion and Concentration February 21, 2012 2 / 39

  3. Introduction Motivation: Large-scale Matrix Completion Goal: Estimate a matrix L 0 ∈ R m × n given a subset of its entries     ? ? 1 . . . 4 2 3 1 . . . 4  → 3 ? ? . . . ? 3 4 5 . . . 1    ? 5 ? . . . 5 2 5 3 . . . 5 Examples Collaborative filtering: How will user i rate movie j ? Netflix: 10 million users, 100K DVD titles Ranking on the web: Is URL j relevant to user i ? Google News: millions of articles, millions of users Link prediction: Is user i friends with user j ? Facebook: 500 million users Jordan (UC Berkeley) Matrix Completion and Concentration February 21, 2012 3 / 39

  4. Introduction Motivation: Large-scale Matrix Completion Goal: Estimate a matrix L 0 ∈ R m × n given a subset of its entries     ? ? 1 . . . 4 2 3 1 . . . 4 3 ? ? . . . ?  → 3 4 5 . . . 1    ? 5 ? . . . 5 2 5 3 . . . 5 State of the art MC algorithms Strong estimation guarantees Plagued by expensive subroutines (e.g., truncated SVD) This talk Present divide and conquer approaches for scaling up any MC algorithm while maintaining strong estimation guarantees Jordan (UC Berkeley) Matrix Completion and Concentration February 21, 2012 4 / 39

  5. Matrix Completion Background Exact Matrix Completion Goal: Estimate a matrix L 0 ∈ R m × n given a subset of its entries Jordan (UC Berkeley) Matrix Completion and Concentration February 21, 2012 5 / 39

  6. Matrix Completion Background Noisy Matrix Completion Goal: Given entries from a matrix M = L 0 + Z ∈ R m × n where Z is entrywise noise and L 0 has rank r ≪ m, n , estimate L 0 Good news: L 0 has ∼ ( m + n ) r ≪ mn degrees of freedom B ⊤ = L 0 A Factored form: AB ⊤ for A ∈ R m × r and B ∈ R n × r Bad news: Not all low-rank matrices can be recovered Question: What can go wrong? Jordan (UC Berkeley) Matrix Completion and Concentration February 21, 2012 6 / 39

  7. Matrix Completion Background What can go wrong? Entire column missing   1 2 ? 3 . . . 4 3 5 ? 4 . . . 1   2 5 ? 2 . . . 5 No hope of recovery! Solution: Uniform observation model Assume that the set of s observed entries Ω is drawn uniformly at random: Ω ∼ Unif ( m, n, s ) Jordan (UC Berkeley) Matrix Completion and Concentration February 21, 2012 7 / 39

  8. Matrix Completion Background What can go wrong? Bad spread of information     1 1 0 0 � �� � 0 1 1 0 0 0 0 0 L = =    0 0 0 0 Can only recover L if L 11 is observed Solution: Incoherence with standard basis (Cand` es and Recht, 2009) A matrix L = UΣV ⊤ ∈ R m × n with rank( L ) = r is ( µ, r ) -coherent if 2 ≤ µr/m � max i � UU ⊤ e i � Singular vectors are not too sparse: 2 ≤ µr/n max i � VV ⊤ e i � � µr and not too cross-correlated: � UV ⊤ � ∞ ≤ mn Jordan (UC Berkeley) Matrix Completion and Concentration February 21, 2012 8 / 39

  9. Matrix Completion Background How do we estimate L 0 ? First attempt: minimize A rank( A ) ( i,j ) ∈ Ω ( A ij − M ij ) 2 ≤ ∆ 2 . � subject to Problem: Intractable to solve! Solution: Solve convex relaxation (Fazel, Hindi, and Boyd, 2001; Cand` es and Plan, 2010) minimize A � A � ∗ ( i,j ) ∈ Ω ( A ij − M ij ) 2 ≤ ∆ 2 subject to � where � A � ∗ = � k σ k ( A ) is the trace/nuclear norm of A . Questions: Will the nuclear norm heuristic successfully recover L 0 ? Can nuclear norm minimization scale to large MC problems? Jordan (UC Berkeley) Matrix Completion and Concentration February 21, 2012 9 / 39

  10. Matrix Completion Background Noisy Nuclear Norm Heuristic: Does it work? Yes, with high probability. Typical Theorem If L 0 is ( µ, r ) -coherent, s = O ( µrn log 2 ( n )) entries of M ∈ R m × n are observed uniformly at random, and ˆ L solves the noisy nuclear norm heuristic, then � ˆ L − L 0 � F ≤ f ( m, n )∆ with high probability when � M − L 0 � F ≤ ∆ . See Cand` es and Plan (2010); Mackey, Talwalkar, and Jordan (2011); Keshavan, Montanari, and Oh (2010); Negahban and Wainwright (2010) Implies exact recovery in the noiseless setting ( ∆ = 0 ) Jordan (UC Berkeley) Matrix Completion and Concentration February 21, 2012 10 / 39

  11. Matrix Completion Background Noisy Nuclear Norm Heuristic: Does it scale? Not quite... Standard interior point methods (Cand` es and Recht, 2009) : O( | Ω | ( m + n ) 3 + | Ω | 2 ( m + n ) 2 + | Ω | 3 ) More efficient, tailored algorithms: Singular Value Thresholding (SVT) (Cai, Cand` es, and Shen, 2010) Augmented Lagrange Multiplier (ALM) (Lin, Chen, Wu, and Ma, 2009) Accelerated Proximal Gradient (APG) (Toh and Yun, 2010) All require rank- k truncated SVD on every iteration Take away: Provably accurate MC algorithms are still too expensive for large-scale or real-time matrix completion Question: How can we scale up a given matrix completion algorithm and still retain estimation guarantees? Jordan (UC Berkeley) Matrix Completion and Concentration February 21, 2012 11 / 39

  12. Matrix Completion DFC Divide-Factor-Combine ( DFC ) Our Solution: Divide and conquer Divide M into submatrices. 1 Factor each submatrix in parallel . 2 Combine submatrix estimates to estimate L 0 . 3 Advantages Factoring a submatrix is often much cheaper than factoring M Multiple submatrix factorizations can be carried out in parallel DFC works with any base MC algorithm With the right choice of division and recombination, yields estimation guarantees comparable to those of the base algorithm Jordan (UC Berkeley) Matrix Completion and Concentration February 21, 2012 12 / 39

  13. Matrix Completion DFC DFC-Proj : Partition and Project Randomly partition M into n/l column submatrices 1 where each C i ∈ R m × l � � M = C 1 C 2 · · · C n/l Complete the submatrices in parallel to obtain 2 � ˆ ˆ ˆ � · · · C 1 C 2 C n/l Reduced cost: Expect min( n/l, m/d ) speed-up per iteration Parallel computation: Pay cost of one cheaper MC Recover a single factorization for M by projecting each 3 submatrix onto the column space of ˆ C 1 � ˆ L proj = ˆ ˆ C 1 ˆ C + ˆ ˆ � C 1 C 2 · · · C n/l 1 Minimal cost: O( mk 2 + lk 2 ) where k = rank(ˆ L proj ) Ensemble: Project onto column space of each ˆ C j and average 4 Jordan (UC Berkeley) Matrix Completion and Concentration February 21, 2012 13 / 39

  14. Matrix Completion DFC DFC : Does it work? Yes, with high probability. Theorem (Mackey, Talwalkar, and Jordan, 2011) If L 0 is ( µ, r ) -coherent and s entries of M ∈ R m × n are observed uniformly at random, then � µ 2 r 2 n 2 log 2 ( n ) � l = O sǫ 2 random columns suffice to have L proj − L 0 � F ≤ (2 + ǫ ) f ( m, n )∆ � ˆ with high probability when � M − L 0 � F ≤ ∆ and the noisy nuclear norm heuristic is used as a base algorithm. Can sample vanishingly small fraction of columns ( l/n → 0 ) whenever s = ω ( n log 2 ( n )) Implies exact recovery for noiseless ( ∆ = 0 ) setting Jordan (UC Berkeley) Matrix Completion and Concentration February 21, 2012 14 / 39

  15. Matrix Completion DFC DFC : Does it work? Yes, with high probability. Proof Ideas: Uniform column/row sampling yields submatrices with low 1 coherence (high spread of information) w.h.p. Each submatrix has sufficiently many observed entries w.h.p. 2 ⇒ Submatrix completion succeeds Uniform sampling of columns/rows captures the full column/row 3 space of L 0 w.h.p. Noisy analysis builds on randomized ℓ 2 regression work of Drineas, Mahoney, and Muthukrishnan (2008) ⇒ Column projection succeeds Jordan (UC Berkeley) Matrix Completion and Concentration February 21, 2012 15 / 39

  16. Matrix Completion Simulations DFC Noisy Recovery Error MC 0.25 Part−10% Proj−10% 0.2 Nys−10% Proj−Ens−10% Nys−Ens−10% 0.15 RMSE Proj−Ens−25% Base−MC 0.1 0.05 0 0 2 4 6 8 10 % revealed entries Figure: Recovery error of DFC relative to base algorithms with ( m = 10 K, r = 10) . Jordan (UC Berkeley) Matrix Completion and Concentration February 21, 2012 16 / 39

  17. Matrix Completion Simulations DFC Speed-up MC 3000 Part−10% Proj−10% 2500 Nys−10% Proj−Ens−10% 2000 Nys−Ens−10% time (s) Base−MC 1500 1000 500 0 1.5 2 2.5 3 3.5 4 4.5 5 m 4 x 10 Figure: Speed-up over APG for random matrices with r = 0 . 001 m and 4% of entries revealed. Jordan (UC Berkeley) Matrix Completion and Concentration February 21, 2012 17 / 39

  18. Matrix Completion CF Application: Collaborative filtering Task: Given a sparsely observed matrix of user-item ratings, predict the unobserved ratings Issues Full-rank rating matrix Noisy, non-uniform observations The Data Netflix Prize Dataset 1 100 million ratings in { 1 , . . . , 5 } 17,770 movies, 480,189 users 1 http://www.netflixprize.com/ Jordan (UC Berkeley) Matrix Completion and Concentration February 21, 2012 18 / 39

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend