rlt residual loop training in collaborative filtering for
play

RLT: Residual-Loop Training in Collaborative Filtering for Combining - PowerPoint PPT Presentation

RLT: Residual-Loop Training in Collaborative Filtering for Combining Factorization and Global-Local Neighborhood Lei Li 1 , 2 , Weike Pan 1 , Li Chen 2 , and Zhong Ming 1 lilei1995eli@gmail.com, panweike@szu.edu.cn,


  1. RLT: Residual-Loop Training in Collaborative Filtering for Combining Factorization and Global-Local Neighborhood Lei Li 1 , 2 , Weike Pan 1 ∗ , Li Chen 2 , and Zhong Ming 1 ∗ lilei1995eli@gmail.com, panweike@szu.edu.cn, lichen@comp.hkbu.edu.hk, mingz@szu.edu.cn 1 College of Computer Science and Software Engineering Shenzhen University, Shenzhen, China 2 Department of Computer Science Hong Kong Baptist University, Hong Kong, China Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 1 / 24

  2. Introduction Problem Definition Rating Prediction Input: A set of (user, item, rating) triples as training data denoted by R = { ( u , i , r ui ) } , where r ui is the numerical rating assigned by user u to item i . Goal: Estimate the preference of user u to item j , i.e., ˆ r uj , for each record in the test data R te = { ( u , j , r uj ) } . Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 2 / 24

  3. Introduction Limitations of Related Work Traditional pipelined residual training paradigm may not be able to fully exploit the merits of factorization- and neighborhood-based methods. There are two different types of neighborhood, i.e., global 1 neighborhood in FISM and SVD++, and local neighborhood in ICF, but most residual training approaches ignore the global neighborhood. Combining the factorization-based method and 2 neighborhood-based method in a pipelined residual chain may not be the best because the one-time interaction between the two methods may not be sufficient. Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 3 / 24

  4. Introduction Overall of Our Solution Residual-Loop Training (RLT) : a new residual training paradigm, which aims to fully exploit the complementarity of factorization, global neighborhood and local neighborhood in one single algorithm. Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 4 / 24

  5. Introduction Advantages of Our Solution We recognize the difference between global neighborhood and 1 local neighborhood in the context of residual training. We propose to combine factorization-, global neighborhood-, and 2 local neighborhood-based methods by residual training. We propose a new residual training paradigm called residual-loop 3 training (RLT). Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 5 / 24

  6. Introduction Notations Table: Some notations and explanations. u user ID i , i ′ , j item ID r ui rating of user u to item i R = { ( u , i , r ui ) } rating records of training data users who rate item i U i items rated by user u I u nearest neighbors of item i N i µ ∈ R global average rating value b u ∈ R user bias b i ∈ R item bias d ∈ R number of latent dimensions U u · ∈ R 1 × d user-specific latent feature vector V i · , W i · ∈ R 1 × d item-specific latent feature vector R te = { ( u , j , r uj ) } rating records of test data r ui predicted rating of user u to item i ˆ λ tradeoff parameter T iteration number in the algorithm Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 6 / 24

  7. Background Factorization-based Method Probabilistic matrix factorization (PMF) is a factorization-based method for rating prediction in collaborative filtering. Specifically, the prediction rule of the rating assigned by user u to item i is as follows, ui = µ + b u + b i + U u · V T r F ˆ i · , (1) where µ , b u and b i are the global average, the user bias, and the item bias, respectively, and U u · ∈ R 1 × d and V i · ∈ R 1 × d are the user-specific latent feature vector and the item-specific latent feature vector, respectively. Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 7 / 24

  8. Background Local Neighborhood-based Method Item-oriented collaborative filtering (ICF) is a neighborhood-based method for preference estimation in recommendation. The estimated preference of user u to item i can be written as follows, r s i ′ i r ui ′ , � ˆ N ℓ ¯ ui = (2) i ′ ∈I u ∩N i s i ′ i = s i ′ i / � i ′ ∈I u ∩N i s i ′ i is the normalized similarity with where ¯ s i ′ i = |U i ′ ∩ U i | / |U i ′ ∪ U i | as the Jaccard index between item i ′ and item i . N i is a set of locally nearest neighboring items of item i , i.e., their similarities are predefined without global propagation among the users, thus we call it a local neighborhood-based method. Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 8 / 24

  9. Background Global Neighborhood-based Method The similarity in Eq.(2) may also be learned from the data instead of being calculated, e.g., in asymmetric factor model (AFM), the prediction rule of user u to item i is as follows, r N g p i ′ i , � ˆ ¯ ui = (3) i ′ ∈I u \{ i } p i ′ i = W i ′ · V i · / |I u \{ i }| . where ¯ � Two items without common users may still be well connected via 1 the learned latent factors. The prediction rule in Eq.(3) does not restrict to a local 2 neighborhood set N i as that in Eq.(2). We thus call AMF with the prediction rule in Eq.(3) a global neighborhood-based method. Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 9 / 24

  10. Background Factorization- and Global Neighborhood-based Method Matrix factorization with implicit feedback (SVD++) integrates the prediction rules of a factorization-based method and a global neighborhood-based method, µ + b u + b i + U u · V T F - N g r � p i ′ i , ˆ ¯ = i · + ui i ′ ∈I u \{ i } r F r N g ˆ ui + ˆ = ui , (4) from which we can see that SVD++ is a generalized factorization model that inherits the merits of both factorization- and global neighborhood-based methods. Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 10 / 24

  11. Background Residual Training Residual training (RT) is an alternative approach to combining a factorization-based method and a neighborhood-based method. Specifically, a factorization-based model is built using the training data, r F ui for each ( u , i , r ui ) ∈ R can then be obtained, and a predicted rating ˆ based on which a neighborhood-based method is developed using s i ′ i r res ui ′ , where r res ui ′ = r ui ′ − ˆ r F i ′ ∈I u ∩N i ¯ � ui ′ is the residual. The learning procedure can be represented as follows, r F r ˆ ui → ˆ N ℓ ui . (5) r F r N ℓ The final prediction rule is then the summation of ˆ ui and ˆ ui , i.e., r F r ˆ ui + ˆ N ℓ ui . Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 11 / 24

  12. Background Differences between SVD++ and RT The main differences between SVD++ and RT are: SVD++ is an integrative method with one single prediction rule, 1 while RT is a two-step approach with two separate prediction rules. SVD++ exploits factorization and global neighborhood, while RT 2 makes use of factorization and local neighborhood. Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 12 / 24

  13. Method Residual-Loop Training (1/3) In order to fully exploit the complementarity of factorization, global neighborhood and local neighborhood, we propose a new residual training paradigm called residual-loop training (RLT), which is depicted as follows, r F - N g r r F - N g N ℓ ˆ → ˆ ui → ˆ (6) ui ui r F - N g r where ˆ is from Eq.(4) and ˆ N ℓ ui is from Eq.(2). ui Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 13 / 24

  14. Method Residual-Loop Training (2/3) r F - N g For the first ˆ in Eq.(6), we aim to exploit both factorization and 1 ui global neighborhood. The interaction between the factorization-based method and the global neighborhood-based method is richer in such an integrative method than that in two separate steps of RT. r N ℓ For ˆ ui , we aim to boost the performance via local neighborhood, 2 i.e., explicitly combining factorization, global neighborhood and local neighborhood for rating prediction in a residual-training manner. r F - N g For the second ˆ , we aim to further capture the remaining 3 ui effects related to users’ preferences that have not been modeled by the previous two methods yet. Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 14 / 24

  15. Method Residual-Loop Training (3/3) Input : Users’ rating records R = { ( u , i , r ui ) } . r uj , ( u , j ) ∈ R te . Output : Predicted preference of each record in the test data, i.e., ˆ Task 1. Conduct factorization- and global neighborhood-based preference learning (i.e., F-N g r SVD++), and estimate the preference of each record in the training data ˆ and the ui r F-N g preference of each record in the test data ˆ . uj Task 2. Conduct local neighborhood-based preference learning (i.e., ICF) on the residual r ui − ˆ r F-N g r N ℓ , and estimate the preference of each record in the training data ˆ and the ui ui r N ℓ preference of each record in test data ˆ uj . Task 3. Conduct factorization- and global neighborhood-based preference learning again F-N g (i.e., SVD++) on the residual r ui − ˆ r r N ℓ − ˆ ui , and estimate the preference of each record ui F-N g r ′ . Finally, the prediction of each record in the test data is obtained in the test data ˆ uj r uj = ˆ r F-N g r N ℓ r F-N g ˆ + ˆ + ˆ ′ . uj uj uj Figure: The algorithm of residual-loop training (RLT). Li et al., (SZU & HKBU) Residual-Loop Training (RLT) SCF ICWS 2018 15 / 24

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend