On the Effectiveness of Linear Models for One-Class Collaborative Filtering
Suvash Sedhain1,2, Aditya Menon2,1, Scott Sanner3,1, Darius Braziunas4
Australian National University1 NICTA2 Oregon State University3 Rakuten Kobo Inc 4
On the Effectiveness of Linear Models for One-Class Collaborative - - PowerPoint PPT Presentation
On the Effectiveness of Linear Models for One-Class Collaborative Filtering Suvash Sedhain 1,2 , Aditya Menon 2,1 , Scott Sanner 3,1 , Darius Braziunas 4 Australian National University 1 NICTA 2 Oregon State University 3 Rakuten Kobo Inc 4
Suvash Sedhain1,2, Aditya Menon2,1, Scott Sanner3,1, Darius Braziunas4
Australian National University1 NICTA2 Oregon State University3 Rakuten Kobo Inc 4
. . . . . .
1 1 1 1
. . . . . .
. . . . 1 1 ? ? ? ? ? ? ? ? 1 1 1 1 ? ?
. . . . .
. . . .
. . . . . .
=
(Weighted) Matrix Factorization
m × k k × n
. . . . . .
1 1 1 1 . . . . 1 1 1 1 1 1
=
. . . . . .
User Projection Item Projection
10
. . . . . .
1 1 1 1 . . . . 1 1 1 1 1 1
=
user item item
. . . . .
. . . . . . . . . .
1 1 1 1 . . . . 1 1 1 1 1 1 item
Recommendation for
1 1 1
Wu1
Any loss function
Recommendation Learning a model per user
. . .
. . .
1 1 1 1 1 1 1 1 1 1 1
. . .
. . .
instance
learning user-user affinities
from the trivial solution
– L2 loss – Logistic Loss : Liblinear (dual iff #users >> #items)
constraints
constraints; optimization via SGD Truncated Gradient
classification/regression problem – can be interpreted as learning user- user similarities
similarity metrics(eg: Cosine, Jaccard)
classification/regression problem – can be interpreted as learning user- user similarities
Recommendation
Where,
If
communication
. . .
. . .
1 1 1 1 1 1 1 1 1 1 1
. . .
. . .
Genre Actors
. . .
1 1 Item features
Evaluation Metrics
users
– Most Popular – Neighborhood
Ranking (BPR)
– Elastic Net Lrec + Non-Negativity (Lrec + Sq + L1+ NN) – Squared Loss LRec (Lrec + Sq) – Logistic Loss LRec (LRec)
Did not finish
Precision@20 on ML1M and LastFM dataset
Did not finish
Precision@20 on Kobo and LastFM dataset
% improvement over WRMF on ML1M dataset
Users segmented by the number of observation
Recommendation from WRMF vs LRec LRec is more personalized