Combining Factorization Model and Additive Forest for Recommendation
Presenter: Tianqi Chen
Team ACMClass@SJTU
Combining Factorization Model and Additive Forest for Recommendation - - PowerPoint PPT Presentation
Combining Factorization Model and Additive Forest for Recommendation Presenter: Tianqi Chen Team ACMClass@SJTU August 11, 2012 Team ACMClass@SJTU Original team name: undergrads Members are students from ACMClass in SJTU All members
Team ACMClass@SJTU
◮ Original team name: undergrads ◮ Members are students from ACMClass in SJTU ◮ All members are undergraduates, except the presenter:)
1/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
1/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
social network/actjon user age/gender item taxonomy tjmestamp … Factorizatjon Models Additjve Forest Final Solutjon Rank Optjmizatjon Incorporated Informatjon Modeling Approach Combinatjon Focus point of this presentatjon One Joint Model, No Ensemble
2/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
c∈C(u)
c pc
T
c∈C(i)
c qc
c
◮ Θ = {p, q, g}, trained via stochastic gradient descent ◮ α(u) c : user feature of user u: user social network/action,
◮ β(i) c : item feature weight of item(celeberity) i: item
◮ γ(u,i) c
3/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
3/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
S
◮ xu: property feature of user u ◮ fs,root(i,s): function defined by a regression tree ◮ Learning via gradient boosting algorithm
item 1 Forest 1 Forest 2 item 2 item k
item i: Kaifu LEE
Major=IT? no yes no yes no yes no yes no yes no yes Occupation=Student? Age<25? Age>18? Age>12? Occupation=Student? no yes Gender=Female?
Like Dislike
4/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
age<20? Yes No
root:SIGKDD
Major=CS? +1 0.5
Yes No
Has User Tag Data Mining?
+2
Yes No
root:SIGKDD
f1 f2
is female?
age< 18? +1 0.2
Yes No Yes No
root:Barbie
is female?
age< 14? +1
Yes No Yes No
root:Barbie
Forest 1
Individual tree for each item
Forest 2 is learned to complement Forest 1
5/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
◮ Both models have their own advantages on different aspect ◮ Understanding their properties and knowing when to use
6/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
T
◮ F(u) : follow set of u ◮ Linear combination
Follow A? pA
T qi
pB
T qi
Yes Follow B? Yes Score for users who “Follow both A and B” = pA
T qi +pB T qi
◮ Condition composition ◮ Feature selection
Follow B? Follow A? +1 Yes No Yes No root: item i Specific score for conditjon “Follow A and B”
7/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
ui = pT u qi + Wi,ag(u)
◮ ag(u): age segment index ◮ Require predefined partition Wi,1 Wi,2 Wi,3 Wi,4 10 20 30 age partjtjon points age bias parameters
age < 17? age <10? Yes No Yes No root: item i +1
Automatjc find splittjng point
8/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
◮ Both models have their own advantages on different aspect ◮ Understanding their properties and knowing when to use
9/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
9/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
ui(t) = ˆ
◮ binid(t): time bin index of t
ui(t) = ˆ
S
◮ fs,i(t): k-piece step function
Bias Time t1 t2 t3 t4
(a) Item Time Bin
Bias t1 t2 t3 t4 Time
(b) K-piece Step Function
Figure: Comparison of Two Temporal Models
10/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
ui(t) = ˆ
S
◮ time difference between clicks ◮ average click speed of current user
50 100 −4 −3 −2 −1 1 ∆ t (sec) f(∆ t)
(a) ∆t = tnext − tcurr
50 100 −0.4 −0.2 0.2 0.4 ∆ t (sec) f(∆ t)
(b) ∆t = tcurr − tprev
Figure: Single Variable Pattern S
s=1 fs(∆t)
11/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
c∈C(u)
c pc
T
c∈C(i)
c qc
c
S
◮ Combination of all the factorization model and additive forest ◮ Boosting from result of factorization part
12/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
12/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
Table: MAP@3 of different methods
◮ User Modeling and Sequential Patterns contributes the most ◮ Time-aware model is more effective in public data ◮ All of them are important for winning
13/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
◮ Seems Ensemble methods do not work in our experiment ◮ Choose right methods to utilize different kinds of data
◮ Factorization models are powerful, but also have drawbacks ◮ Additive forest can automatic cut the continuous features,
sometimes smarter than human
◮ Use automatic cutting to build robust time-aware model ◮ Fully utilize the available information ◮ Source code: svdfeature.apexlab.org
14/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
◮ The rest parts of the slides are appendix
15/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
◮ Loss function of Pairwise Ranking: AUC optimization
◮ Pseudo loss function of LambdaRank: MAP optimization
◮ ∆ijMAP is MAP change when we swap i and j in current list
◮ C(x) is a surrogate convex loss function
◮ logistic loss(BPR): C(x) = ln(1 + e−x) ◮ hinge loss(maximum margin): C(x) = max(0, 1 − x)
◮ Lu is normalized by number of pairs( |{(i, j)|rui > ruj}| ):
16/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
◮ W: weight matrix ◮ xu: property vector of user u ◮ yi: property vector of item i
|F(u)|
◮ Wc→i: confidence of rule u follows c → u accept i
17/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
◮ Feature-based matrix factorization can be viewed as a
◮ Advantage of W: direct modeling effect of c → i ◮ Advantage of PTQ: less parameter, topic level matching
◮ When W is large and with sparse data support, use
factorization
◮ When W is small and with dense data support, use bilinear 18/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
T
◮ F(u) : set of items user u followed ◮ A(u): set of items user u has action with ◮ αu: weight by action count
19/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation
i = qi + qc1(i) + qc2(i) + qc3(i) + qc4(i)
◮ Taxonomy aware parameter sharing ◮ ck(i): k-th level category of item i belongs to
i = qi +
20/14 ACMClass@SJTU Combining Factorization Model and Additive Forest for Recommendation