NCDREC: A Decomposability Inspired Framework for Top-N - - PowerPoint PPT Presentation

ncdrec a decomposability inspired framework for top n
SMART_READER_LITE
LIVE PREVIEW

NCDREC: A Decomposability Inspired Framework for Top-N - - PowerPoint PPT Presentation

Introduction & Motivation NCDREC Framework Experimental Evaluation Future Work & Conclusion NCDREC: A Decomposability Inspired Framework for Top-N Recommendation Athanasios N. Nikolakopoulos 1 , 2 John D. Garofalakis 1 , 2 1 Computer


slide-1
SLIDE 1

Introduction & Motivation NCDREC Framework Experimental Evaluation Future Work & Conclusion

NCDREC: A Decomposability Inspired Framework for Top-N Recommendation

Athanasios N. Nikolakopoulos1,2 John D. Garofalakis1,2

1Computer Engineering and Informatics Department,

University of Patras, Greece

2Computer Technology Institute & Press “Diophantus”

IEEE/WIC/ACM International Conference on Web Intelligence Warsaw, August 2014

  • A. N. Nikolakopoulos and J. D. Garofalakis

NCDREC

slide-2
SLIDE 2

Introduction & Motivation NCDREC Framework Experimental Evaluation Future Work & Conclusion

Outline

1

Introduction & Motivation Recommender Systems - Collaborative Filtering Challenges of Modern CF Algorithms

2

NCDREC Framework NCDREC Model Criterion for ItemSpace Coverage NCDREC Algorithm: Storage and Computational Issues

3

Experimental Evaluation Evaluation Methodology Quality of Recommendations Long-Tail Recommendations Cold-Start Recommendations

4

Future Work & Conclusion

  • A. N. Nikolakopoulos and J. D. Garofalakis

NCDREC

slide-3
SLIDE 3

Introduction & Motivation NCDREC Framework Experimental Evaluation Future Work & Conclusion Recommender Systems - Collaborative Filtering Challenges of Modern CF Algorithms

Recommender System Algorithms

. RECOMMENDER SYSTEM .

USERS

.

ITEMS

.

RATINGS

. . . . .

Recommendation List

.

Rating Predictions

.

Collaborative Filtering Recommendation Algorithms

Wide deployment in Commercial Environments Significant Research Efforts

  • A. N. Nikolakopoulos and J. D. Garofalakis

NCDREC

slide-4
SLIDE 4

Introduction & Motivation NCDREC Framework Experimental Evaluation Future Work & Conclusion Recommender Systems - Collaborative Filtering Challenges of Modern CF Algorithms

Challenges of Modern CF Algorithms Sparsity

It is an Intrinsic RS Characteristic related to serious problems:

Long-Tail Recommendation Cold start Problem Limited ItemSpace Coverage

Traditional CF techniques, such as neighborhood models, are very susceptible to sparsity. Among the most promising approaches in alleviating sparsity related problems are: Dimensionality-Reduction Models

Build a reduced latent space which is dense.

Graph-Based Models.

Exploit transitive relations in the data, while preserving some of the “locality”.

  • A. N. Nikolakopoulos and J. D. Garofalakis

NCDREC

slide-5
SLIDE 5

Introduction & Motivation NCDREC Framework Experimental Evaluation Future Work & Conclusion Recommender Systems - Collaborative Filtering Challenges of Modern CF Algorithms

Exploiting Decomposability

We attack the problem from a different perspective: Sparsity ← → Hierarchy ← → Decomposability. Nearly Completely Decomposable Systems

Pioneered by Herbert A. Simon. Many Applications in Diverse Disciplines such as economics, cognitive theory and social sciences, to computer systems performance evaluation, data mining and information retrieval

Main Idea: Exploit the innate Hierarchy of the Item Set, and view it as a decomposable space.

Can this enrich the Collaborative Filtering Paradigm in an Efficient and Scalable Way? Does this approach offer any qualitative advantages in alleviating sparsity related problems?

  • A. N. Nikolakopoulos and J. D. Garofalakis

NCDREC

slide-6
SLIDE 6

Introduction & Motivation NCDREC Framework Experimental Evaluation Future Work & Conclusion NCDREC Model Criterion for ItemSpace Coverage NCDREC Algorithm: Storage and Computational Issues

NCDREC Model Overview

Definitions We define a D-decomposition to be an indexed family of sets D {D1, . . . , DK }, that span the ItemSpace V, We define Dv

v∈Dk Dk to be the proximal

set of items of v ∈ V, We also define the associated block coupling graph GD (VD, ED); its vertices correspond to the D-blocks, and an edge between two vertices exists whenever the intersection of these blocks is a non-empty set. Finally, we introduce an aggregation matrix AD ∈ Rm×K , whose jkth element is 1, if vj ∈ Dk and zero

  • therwise.

NCDREC Components Main Component: Recommendation vectors produced by projecting the NCD perturbed data onto an f -dimensional space. ColdStart Component:Recommendation vectors are the stationary distributions of a Discrete Markov Chain Model. W ǫZX⊺ X diag(ADe)−1AD [Z]ik (nk

ui )−1[RAD]ik, when nk ui > 0,

and zero otherwise. G R + ǫW H diag(Ce)−1C, where [C]ij r⊺

i rj for i = j

D XY, X diag(ADe)−1AD, Y diag(A⊺

De)−1A⊺ D,

E eω⊺

S(ω) (1 − α)E + α(βH + (1 − β)D)

  • A. N. Nikolakopoulos and J. D. Garofalakis

NCDREC

slide-7
SLIDE 7

Introduction & Motivation NCDREC Framework Experimental Evaluation Future Work & Conclusion NCDREC Model Criterion for ItemSpace Coverage NCDREC Algorithm: Storage and Computational Issues

Criterion for ItemSpace Coverage

Theorem (ItemSpace Coverage) If the block coupling graph GD is connected, there exists a unique steady state distribution π of the Markov chain corresponding to matrix S that depends on the preference vector ω; however, irrespectively of any particular such vector, the support of this distribution includes every item

  • f the underlying space.

Proof Sketch: When GD is connected, the Markov chain induced by the stochastic matrix S consists of a single irreducible and aperiodic closed set of states, that includes all the items. The above is true for every stochastic vector ω, and for every positive real numbers α, β < 1. Taking into account the fact that the state space is finite, the resulting Markov chain becomes ergodic. So πi > 0, for all i, and the support of the distribution that defines the recommendation vector includes every item

  • f the underlying space.
  • A. N. Nikolakopoulos and J. D. Garofalakis

NCDREC

slide-8
SLIDE 8

Introduction & Motivation NCDREC Framework Experimental Evaluation Future Work & Conclusion NCDREC Model Criterion for ItemSpace Coverage NCDREC Algorithm: Storage and Computational Issues

NCDREC Algorithm: Storage and Computational Issues

Input: Matrices R ∈ Rn×m, H ∈ Rm×m, X ∈ Rm×K , Y ∈ RK×m, Z ∈ Rn×K . Parameters α, β, f , ǫ Output: The matrix with recommendation vectors for every user, Π ∈ Rn×m Step 1: Find the newly added users and collect their preference vectors into matrix Ω. Step 2: Compute Πsparse using the

ColdStart Procedure .

Step 3: Initialize vector p1 to be a random unit length vector. Step 4: Compute the modified Lanczos procedure up to step M, using

NCD PartialLBD

with starting vector p1. Step 5: Compute the SVD of the bidiagonal matrix B and use it to extract f < M approximate singular triplets: {˜ uj, σj, ˜ vj} ← {Qu(B)

j

, σ(B)

j

, Pv(B)

j

} Step 6: Orthogonalize against the approximate singular vectors to get a new starting vector p1. Step 7: Continue the Lanczos procedure for M more steps using the new starting vector. Step 8: Check for convergence tolerance. If met compute matrix: Πfull = ˜ UΣ˜ V⊺ else go to Step 4 Step 9: Update Πfull, replacing the rows that correspond to new users with Πsparse. Return Πfull

  • A. N. Nikolakopoulos and J. D. Garofalakis

NCDREC

slide-9
SLIDE 9

Introduction & Motivation NCDREC Framework Experimental Evaluation Future Work & Conclusion Evaluation Methodology Quality of Recommendations Long-Tail Recommendations Cold-Start Recommendations

Experimental Evaluation Datasets

Yahoo!R2Music MovieLens

Competing Methods

Commute Time (CT) Pseudo-Inverse of the user-item graph Laplacian (L†) Matrix Forest Algorithm (MFA) First Passage Time (FP) Katz Algorithm (Katz)

Metrics

Recall Precision R-Score Normalized Discounted Cumulative Gain (NDCG@k) Mean Reciprocal Rank

  • A. N. Nikolakopoulos and J. D. Garofalakis

NCDREC

slide-10
SLIDE 10

Introduction & Motivation NCDREC Framework Experimental Evaluation Future Work & Conclusion Evaluation Methodology Quality of Recommendations Long-Tail Recommendations Cold-Start Recommendations

Full Dist. Recommendations Methodology

Randomly sample 1.4% of the ratings of the dataset ⇒ probe set P Use each item vj, rated with 5 stars by user ui in P ⇒ test set T Randomly select another 1000 unrated items of the same user for each item in T Form ranked lists by ordering all the 1001 items

TABLE I RECOMMENDATION QUALITY ON MOVIELENS1M AND YAHOO!R2MUSIC DATASETS USING R-SCORE AND MRR METRICS

MovieLens1M Yahoo!R2Music R(5) R(10) MRR R(5) R(10) MRR NCDREC 0.3997 0.5098 0.3008 0.3539 0.4587 0.2647 MFA 0.1217 0.1911 0.0887 0.2017 0.2875 0.1591 L† 0.1216 0.1914 0.0892 0.1965 0.2814 0.1546 FP 0.2054 0.2874 0.1524 0.1446 0.2241 0.0998 Katz 0.2187 0.3020 0.1642 0.1704 0.2529 0.1203 CT 0.2070 0.2896 0.1535 0.1465 0.2293 0.1019 5 10 15 20 0.2 0.4 0.6 MovieLens1M Recall@N 0.2 0.4 0.6 0.8 1 5 · 10−2 0.1 0.15 0.2 Precision/Recall 5 10 15 20 0.2 0.4 0.6 NDCG@N 5 10 15 20 0.2 0.4 0.6 Yahoo!R2Music 0.2 0.4 0.6 0.8 1 5 · 10−2 0.1 0.15 5 10 15 20 0.2 0.4 0.6 NCDREC Katz FP MFA L† CT

  • A. N. Nikolakopoulos and J. D. Garofalakis

NCDREC

slide-11
SLIDE 11

Introduction & Motivation NCDREC Framework Experimental Evaluation Future Work & Conclusion Evaluation Methodology Quality of Recommendations Long-Tail Recommendations Cold-Start Recommendations

Long-Tail Recommendations Methodology (Long Tail)

We order the items according to their popularity (measured in terms of number

  • f ratings)

We further partition the test set T into two subsets, Thead and Ttail We discard the popular items and we evaluate NCDREC and the other algorithms on the Ttail test set.

TABLE II LONG TAIL RECOMMENDATION QUALITY ON MOVIELENS1M AND YAHOO!R2MUSIC DATASETS USING R-SCORE AND MRR METRICS

MovieLens1M Yahoo!R2Music R(5) R(10) MRR R(5) R(10) MRR NCDREC 0.3279 0.4376 0.2395 0.3520 0.4322 0.2834 MFA 0.1660 0.2517 0.1188 0.2556 0.3530 0.1995 L† 0.1654 0.2507 0.1193 0.2492 0.3461 0.1939 FP 0.0183 0.0654 0.0221 0.0195 0.0684 0.0224 Katz 0.0275 0.0822 0.0267 0.0349 0.0939 0.0309 CT 0.0192 0.0675 0.0227 0.0215 0.0747 0.0249 5 10 15 20 0.2 0.4 0.6 MovieLens1M Recall@N 0.2 0.4 0.6 0.8 1 0.1 0.2 Precision/Recall 5 10 15 20 0.2 0.4 0.6 NDCG@N 5 10 15 20 0.2 0.4 0.6 Yahoo!R2Music 0.2 0.4 0.6 0.8 1 0.1 0.2 5 10 15 20 0.2 0.4 0.6 NCDREC Katz FP MFA L† CT

  • A. N. Nikolakopoulos and J. D. Garofalakis

NCDREC

slide-12
SLIDE 12

Introduction & Motivation NCDREC Framework Experimental Evaluation Future Work & Conclusion Evaluation Methodology Quality of Recommendations Long-Tail Recommendations Cold-Start Recommendations

New Users problem Methodology

Randomly select 100 users having rated at least 100 items and delete 96%, 94%, 92% and 90% of each users’ ratings. Compare the rankings induced on the modified data with the complete set of ratings.

Metrics

Spearman’s ρ Kendall’s τ Degree of Agreement (DOA) Normalized Distance-based Performance Measure (NDPM)

4% 6% 8% 10% 0.3 0.32 Percentage of included ratings Spearman’s ρ 4% 6% 8% 10% 0.24 0.25 0.26 Percentage of included ratings Kendall’s τ 4% 6% 8% 10% 0.86 0.88 Percentage of included ratings Degree Of Agreement 4% 6% 8% 10% 0.1 0.12 0.14 Percentage of included ratings NDPM (the smaller the better) NCDREC Katz FP MFA L† CT

  • A. N. Nikolakopoulos and J. D. Garofalakis

NCDREC

slide-13
SLIDE 13

Introduction & Motivation NCDREC Framework Experimental Evaluation Future Work & Conclusion

Future Research Directions & Conclusion Future Work Decomposition Granularity Effect

Coarse Grained ⇒ Sparseness Insensitivity Fine Grained ⇒ Higher Quality of Recommendations

Multiple-Criteria Decompositions

How it affects the theoretical properties of the ColdStart Component?

  • A. N. Nikolakopoulos and J. D. Garofalakis

NCDREC

slide-14
SLIDE 14

Introduction & Motivation NCDREC Framework Experimental Evaluation Future Work & Conclusion

Thanks! Q&A

  • A. N. Nikolakopoulos and J. D. Garofalakis

NCDREC

slide-15
SLIDE 15

Example NCD Proximity Matrix NCDREC Basic SubComponents

Example NCD Proximity Matrix

Back

           D1 D2 D3 Nv Gv v1

− 1 {v1, v2, v4} v2

  • 2

{v1, v2, v4, v5, v6, v7, v8} v3 −

1 {v3, v4, v5, v8} v4

2 {v1, v2, v3, v4, v5, v8} v5 −

  • 2

{v2, v3, v4, v5, v6, v7, v8} v6 − −

  • 1

{v2, v5, v6, v7, v8} v7 − −

  • 1

{v2, v5, v6, v7, v8} v8 −

  • 2

{v2, v3, v4, v5, v6, v7, v8}           

v1 v2 v3 v4 v5 v6 v7 v8 D1 D2 D3 v1 v2 v3 v4 v5 v6 v7 v8

1 1/2 1/2 1 1/2 1/2 1/2 1/2 1 1 1/2 1/2 1/3 1/3 1/3 1/4 1/4 1/4 1/4 1/5 1/5 1/5 1/5 1/5

=⇒ D =                                              

1 3 1 3 1 3 1 6 4 15 1 6 1 10 1 10 1 10 1 10 1 4 1 4 1 4 1 4 1 6 1 6 1 8 7 24 1 8 1 8 1 10 1 8 1 8 9 40 1 10 1 10 9 40 1 5 1 5 1 5 1 5 1 5 1 5 1 5 1 5 1 5 1 5 1 10 1 8 1 8 9 40 1 10 1 10 9 40

                                             

  • A. N. Nikolakopoulos and J. D. Garofalakis

NCDREC

slide-16
SLIDE 16

Example NCD Proximity Matrix NCDREC Basic SubComponents NCD PartialLBD Procedure ColdStart Procedure

NCD PartialLBD

Back

procedure NCD PartialLBD(R, X, Z, p1, ǫ) φ ← X⊺p1; q1 ← Rp1 + ǫZφ; b1,1 ← q12 ; u1 ← q1/b1,1; for j = 1 to M do φ ← Z⊺qj; r ← R⊺qj + ǫXφ − bj,jpj; r ← r − [p1 . . . pj] ([p1 . . . pj]⊺r); if j < M then bj,j+1 ← r; pj+1 ← r/bj,j+1; φ ← X⊺pj+1; qj+1 ← Rpj+1 + ǫZφ − bj,j+1qj; qj+1 ← qj+1 − [q1 . . . qj] ([q1 . . . qj]⊺qj+1); bj+1,j+1 ← qj+1; qj+1 ← qj+1/bj+1,j+1; end if end for end procedure

  • A. N. Nikolakopoulos and J. D. Garofalakis

NCDREC

slide-17
SLIDE 17

Example NCD Proximity Matrix NCDREC Basic SubComponents NCD PartialLBD Procedure ColdStart Procedure

ColdStart Procedure

Back

procedure ColdStart(H, X, Y, Ω, α, β) Π ← Ω; k ← 0; r ← 1; while r > tol and k ≤ maxit do k ← k + 1; ˆ Π ← αβΠH; Φ ← ΠX; ˆ Π ← ˆ Π + α(1 − β)ΦY + (1 − α)Ω; r ← ˆ Π − Π; Π ← ˆ Π; end while return Πsparse ← Π end procedure

  • A. N. Nikolakopoulos and J. D. Garofalakis

NCDREC