Distributed Collaborative Filtering and Adaptive User-to-User - - PowerPoint PPT Presentation

distributed collaborative filtering and adaptive user to
SMART_READER_LITE
LIVE PREVIEW

Distributed Collaborative Filtering and Adaptive User-to-User - - PowerPoint PPT Presentation

Distributed Collaborative Filtering and Adaptive User-to-User Correlation Francesco Ricci Faculty of Computer Science Free University of Bozen-Bolzano, Italy Joint work with Shlomo Berkovsky (CIRSO), Tsvi Kuflik (University of Haifa), and


slide-1
SLIDE 1

Distributed Collaborative Filtering and Adaptive User-to-User Correlation

Francesco Ricci

Faculty of Computer Science Free University of Bozen-Bolzano, Italy

Joint work with Shlomo Berkovsky (CIRSO), Tsvi Kuflik (University of Haifa), and Linas Baltrunas (University of Bozen)

slide-2
SLIDE 2

2

Content

Introduction to recommender systems and

collaborative filtering

Motivations:

Decentralized collaborative filtering Improve accuracy by partitioning ratings and re-

aggregating information

Domain-based rating partitioning Importing user modelling information Computing inter-domain correlations Evaluation Extension: adapting the similarity metric to the

prediction problem

slide-3
SLIDE 3

3

W hat m ovie should I see?

The Internet Movie Database (IMDb) provides information about actors, films, television shows, television stars, video games and production crew personnel. Owned by Amazon.com since 1998 September 15, 2008 IMDb featured 1,039,447 titles and 2,723,306 people More than 57M users per month.

slide-4
SLIDE 4

4

Recom m ender System s

In everyday life w e rely on recom m endations

from other people either by word of mouth, recommendation letters, movie and book reviews printed in newspapers …

In a typical recommender system people

provide recom m endations as inputs, w hich the system then aggregates and directs to appropriate recipients

Aggregation of recommendations Match the recommendations with those

searching for recommendations

[Resnick and Varian, 1997]

slide-5
SLIDE 5

5

Exam ples

  • Am azon.com – looks in the user past buying history,

and recommends product bought by a user with similar buying behavior

  • Tripadvisor.com - Quoting product reviews of a

community of users

  • Myproductadvisor.com – make questions about

searched benefits (product features) to reduce the number of candidate products

  • Yahoo.com – “Today’s Picks” highlight ten destinations

that are highly-relevant to individual users, based on recent online activity and preferences.

  • iTunes Genius – recommend albums similar to those

found in your library

  • Sm arter Kids – self selection of a user profile –

classification of products in user profiles.

slide-6
SLIDE 6

6

Social Filtering

???

slide-7
SLIDE 7

7

Movie Lens

http://movielens.umn.edu

slide-8
SLIDE 8

8

slide-9
SLIDE 9

9

slide-10
SLIDE 10

10

slide-11
SLIDE 11

11

slide-12
SLIDE 12

12

slide-13
SLIDE 13

13

Items Users

Matrix of ratings

slide-14
SLIDE 14

14

Collaborative-Based Filtering

  • A collection of n user ui and a collection of m products pj
  • A n × m matrix of ratings vij , with vij = ? if user i did not rate

product j

  • Prediction for user i and product j is computed as
  • Where, vi is the average rating of user i, K is a normalization

factor such that the sum of uik is 1, and

∑ ∑ ∑

− − − − =

j j k kj i ij j k kj i ij ik

v v v v v v v v u

2 2

) ( ) ( ) )( (

  • Where the sum (and averages) is over j s.t. vij and vkj are

not “?”. Similarity of users i and k

[Breese et al., 1998]

− + =

? *

) (

kj

v k kj ik i ij

v v u K v v

slide-15
SLIDE 15

15

Exam ple

ui u8 u9 u5 vi = 3.2 v8 = 3.5 v9 = 3 v5= 4 pj 4 ? 3 5 Users’ similarities: ui5 = 0.5, ui8 = 0.5, ui9 = 0.8 v*ij = 3.2 + 1/(0.5+0.5+0.8) * [0.5 (4 -4) + 0.5 (3 – 3.5) + 0.8 (5 -3) = 3.2 + 1/1.8 * [0 - 0.25 + 1.6] = 3.2 + 0.75 = 3.95

− + =

? *

) (

kj

v k kj ik i ij

v v u K v v

slide-16
SLIDE 16

16

Distributed Scenario

Target Recommender Sys. Remote Recommender Sys. Remote Recommender Sys. Remote Recommender Sys.

q=<user = i> recommend j reply from a remote system

User identifiers User models User identifiers and their similarities Rating prediction for j

q=<user = i, item = j, target = t>

slide-17
SLIDE 17

17

Related W orks

  • B. N. Miller, J. A. Konstan, J. Riedl, “PocketLens:

Toward a Personal Recommender System”, 2004

  • R. Burke, Hybrid web recommender systems. In

The Adaptive Web, page 377-408. Springer Berlin / Heidelberg, 2007.

  • K. Yu, X. Xu, M. Ester, H. P. Kriegel, Feature

Weighting and Instance Selection for Collaborative Filtering: An Information-Theoretic Approach, in Knowledge Information Systems,

  • vol. 5(2), 2003.
  • J. Freyne, B. Smyth, Communities,

Collaboration and Cooperation in Personalized Web Search, in proc. of the ITWP Workshop, Edinburgh, UK, 2005.

slide-18
SLIDE 18

18

I nform ation processing in CF

1 . Sim ilarity com putation: assessing the similarity of all the users to the active user, i.e., the user for whom a recommendation is searched 2 . Neighborhood form ation: selecting the K most similar users to the active user 3 . Com puting the active user rating prediction: for a target item whose rating is unknown

1. weight the ratings - on the target item - of the K most similar users, found at (2) according to the user-to-user similarity computed at (1) 2. the predicted rating is the weighted average.

slide-19
SLIDE 19

19

W hat inform ation can be exchanged

UMs (rating vectors) stored

by the remote systems

Lists of the neighborhood

candidates computed by the remote systems

Degrees of sim ilarity between the active

user and the other users, computed over the data stored by the remote systems

Complete predictions generated by the

remote systems.

slide-20
SLIDE 20

20

Assum ptions

Users can be identified uniquely in all the

domains

I tem s can be identified uniquely in all domains Target dom ain sends a request to remote

domains specifying q= < i, j, t>

i is the identifier of the active user j is the target item identifier (possibly null) t is the target domain.

Different distributed prediction methods are

characterized by "w hat the rem ote dom ains reply".

slide-21
SLIDE 21

21

Rating Matrix and Dom ains

Given the assumptions: there is a "centralized"

(aggregated) model of the distributed scenario

V is the overall rating matrix Va, Vb, Vc, are rating sub matrices for three

domains Ra, Rb, Rc

v11 v12 v13 … v1m v21 v22 v23 … v2m v31 v32 v33 … v3m vn1 vn2 vn3 … vnm Va Vb Vc V =

slide-22
SLIDE 22

22

Prediction Methods - Rem ote Replies

  • Local Prediction: the remote systems { Rd} d∈D do not return any

data

  • Centralized Prediction: All the ratings managed by { Rd} d∈D are

sent back - we assume that all the domains are related and D is the full set of domains

  • Distributed Peer I dentification: The identifiers of users that all

the remote systems { Rd} d∈D consider as “similar” to the target user i are sent back

  • Distributed Neighborhood Form ation: The identifiers of the

users that all the remote systems { Rd} d∈D consider as “similar” to the target user i, together w ith their sim ilarities to the target user i - similarities are computed by the remote system using only the ratings in Vd

  • Distributed Prediction: The rating predictions for item j

computed by the remote systems { Rd} d∈D using the ratings contained in Vd, d∈D are sent back.

slide-23
SLIDE 23

23

Distributed Peer I dentification

The identifiers of K users that all the remote

systems { Rd} d∈D consider as “similar” to the target user i are sent back

D is the set of all rem ote dom ains In our experiments a domain is identified by a

tag (a genre)

The target domain m erge the received peers and

make a prediction using the local data (ratings

  • nly in the target domain).

Remote systems provide knowledge as an

inform ed selection of users.

slide-24
SLIDE 24

24

Distributed Peer I dentification

Va is the target domain Vb and Vc are remote domains (containing

some ratings of the target user)

v11 v12 v13 … v1m v21 v22 v23 … v2m v31 v32 v33 … v3m vn1 vn2 vn3 … vnm Va Vb Vc V =

target user target domain

Vb peer Vc peer Vb peer

slide-25
SLIDE 25

25

Distributed Neighborhood Form ation

  • The identifiers of some users that the remote systems

{ Rd} d∈D consider as “similar” to the target user i, together w ith their sim ilarities to the target user i - similarities are computed by the remote system using only the ratings in Vd

  • The similarity of a neighbor user l with the target user i is a

weighted average

  • where D is the set of all the domains (including t), t is the

target domain and

cor(d,t) is the I nter-Dom ain Correlation measure

between domains.

∑ ∑

∈ ∈

=

D d D d d

t d cor l i sim t d cor l i sim ) , ( ) , ( ) , ( ) , (

slide-26
SLIDE 26

26

I nter-Dom ain Correlations

Content-based m ethod

Mining the textual descriptions of the items in the

domains from external data sources to obtain tf-idf description (vector v) of each domain

Computing cosine correlation of the domain

representations

Rating-based m ethod

Average correlation of the items in the domains (Jd)

|| || || || ) , ( ) , (

2 1 2 1 2 1 2 1

v v v v v v sim d d corcontent ∗ ⋅ = = } , , : ) , ( { ) , (

2 1

2 1 d d ratings

J k J j k j k j sim AVG d d cor ∈ ∈ ≠ =

slide-27
SLIDE 27

27

Distributed Prediction

The rating predictions for item j computed

by the remote systems { Rd} d∈D using the ratings contained in Vd, d∈ D are sent back to the target domain Rt

Upon receiving the set of predictions, Rt

aggregates all the predictions (including the local one) into a single value by averaging the predictions

We do not use here the Inter-Domain

Correlation.

slide-28
SLIDE 28

28

Evaluation

  • EachMovie data set - each movie belongs to 1.366 genres

(average)

  • We used only 8 genres
  • Content-based correlation used I MDb data
  • We used Cosine similarity
  • At least 6 overlapping movies to make a prediction
  • The number of nearest neighbors used to make the

prediction is 2 0

92.321 93.179 93.181 92.432 92.180 92.425 93.852 91.923 sparsity (%) 991K 681K 433K 800K 3,056K 2,209K 193K 1,166K

  • num. ratings

177 137 87 145 536 400 43 198

  • num. Movies

Thriller romance horror family drama comedy animat. action

slide-29
SLIDE 29

29

I nter-dom ain correlations: content

1.000 0.913 0.939 0.772 0.938 0.903 0.787 0.943 thriller 0.913 1.000 0.850 0.832 0.987 0.957 0.838 0.913 romance 0.939 0.850 1.000 0.739 0.873 0.868 0.765 0.902 horror 0.772 0.832 0.739 1.000 0.841 0.905 0.914 0.820 family 0.938 0.987 0.873 0.841 1.000 0.965 0.848 0.932 drama 0.903 0.957 0.868 0.905 0.965 1.000 0.913 0.935 comedy 0.787 0.838 0.765 0.914 0.848 0.913 1.0000 0.860 animat. 0.943 0.913 0.902 0.820 0.932 0.935 0.860 1.000 action thriller romance horror family drama comedy animat. action

slide-30
SLIDE 30

30

I nter-dom ain correlations: ratings

0.109 0.074 0.098 0.076 0.069 0.074 0.082 0.109 thriller 0.074 0.091 0.060 0.072 0.065 0.070 0.074 0.075 romance 0.098 0.060 0.149 0.067 0.060 0.065 0.077 0.093 horror 0.076 0.072 0.067 0.119 0.056 0.071 0.125 0.086 family 0.069 0.065 0.060 0.056 0.063 0.058 0.059 0.067 drama 0.074 0.070 0.065 0.071 0.058 0.072 0.074 0.078 comedy 0.082 0.074 0.077 0.125 0.059 0.074 0.167 0.095 animat. 0.109 0.075 0.093 0.086 0.067 0.078 0.095 0.129 action thriller romance horror family drama comedy animat. action

slide-31
SLIDE 31

31

MAE: Centralized - Local - Distributed

  • Local: more reliable similarity computation especially for users

that rated few movies (below 6% of the movies)

  • Distributed: very similar to local - there are few cases where the

average prediction is actually using information from remote domains.

0.16 0.17 0.18 0.19 0.2 0.21 0.22 0.23 below 3 3 to 6 6 to 9 9 to 12 12 to 15 15 to 18 18 to 21 21 to 24 24 to 27 27 to 30 30 to 33

  • ver 33

P e r c e nt a ge of Ra t e d M ov i e s i n t he Ta r ge t Ge nr e Cent ralized Predict ion Local Predict ion Dist ribut ed Predict ion

MAE

slide-32
SLIDE 32

32

MAE: Centralized - Distributed Peer I dentification

  • Distributed peer identification works well for medium-large user models -

the neighbors found in the remote systems tend to be the same as in the target

  • Distributed peer identification error is bounded (from below) by the local

approach

  • For small user models the peers identified are different from the local ones

and the accuracy decreases.

0.16 0.17 0.18 0.19 0.2 0.21 0.22 0.23 below 3 3 to 6 6 to 9 9 to 12 12 to 15 15 to 18 18 to 21 21 to 24 24 to 27 27 to 30 30 to 33 over 33

Percentage of Rat ed Movies in the Target Genre Centalized Prediction Distributed Peer Ident if ication

MAE

slide-33
SLIDE 33

33

I nter-Dom ain correlation Distributed Neighbor Form ation

  • Using a uniform w eighting for the similarities computed in the remote

domains does not w ork

  • Distributed neighbor formation methods - with content-based and rating-

based correlation - works in a similar way and im prove the centralized approach

  • But they do not improve the local and distributed approach.

0,16 0,17 0,18 0,19 0,2 0,21 0,22 0,23 below 3 3 to 6 6 to 9 9 to 12 12 to 15 15 to 18 18 to 21 21 to 24 24 to 27 27 to 30 30 to 33 over 33

Percentage of Rated Movies in the Target Genre Centralized Prediction CG-Uniform CG-Contents CG-Ratings

MAE

slide-34
SLIDE 34

34

Lim itations

Local and (all) distributed methods reduce

recall since they exploit smaller user profiles (lack

  • f overlapping items between target user and

neighbor)

Distributed neighbor form ation: better recall

because the user-to-user similarity is computed in all domains and then averaged – if there is enough rating overlap in some domain then ok

The genre-based partition is domain specific -

not tested general methods for item partitioning

Tested on one dataset only Experiments using "real" distributed recommender

systems were not done - privacy issues may prevent UM data exchange.

slide-35
SLIDE 35

35

Adaptive Sim ilarity Metric

All the previous approaches (except the

Centralized one) compute the user-to-user similarity using a subset of the target user and peers ratings

Is it possible to identify - for each user-item pair

the best set of ratings upon which to compute the user-to-user similarity?

Exam ple: when predicting a rating for a movie

by Schwarzenegger use only the ratings for

  • ther movies by Schwarzy and Stallone.
slide-36
SLIDE 36

36

I tem Selection

  • In adaptive user-to-user similarity select an item if:
  • there is a rating for the item in both users
  • the item has a great correlation/ im portance for the

item whose rating has to be predicted

  • I tem s Correlation W eights
  • Variance W eighting: the larger the variance of the

item, the bigger is the weight

  • I PCC: the larger is the Pearson correlation, the larger

the weight

  • Mutual I nform ation: weight is computed as the

information that predictive item provides to the knowledge of target item

  • Genre w eighting: the more genres the two items share

the larger is the weight.

slide-37
SLIDE 37

37

Perform ance of Adaptive I tem Selection

Less data can lead to

a better prediction

Selecting a small

number of

  • verlapping items

improves the accuracy

Improvement

achieved for all our used error measures (also precision and recall)

Best weighting

method depends on the error measure.

slide-38
SLIDE 38

38

Contribution

A distributed m odel for a dom ain-specialized

recommender system - components communicate in a cooperative way with a simple request-response protocol

Methods for exploiting the additional know ledge

provided by the classification of an item into a dom ain and the proof that the accuracy of CF can be improved

Methods for adapting the sim ilarity m etric to the

target prediction and improve accuracy

The validation that accuracy of CF can be im proved

by basing the rating prediction on information contained in a subset (carefully selected) of the items.

slide-39
SLIDE 39

39

More info on recommender systems: RecSys: ACM Conference on Recommender Systems 2007 and 2008

slide-40
SLIDE 40

40

slide-41
SLIDE 41

41

slide-42
SLIDE 42

42

slide-43
SLIDE 43

43

slide-44
SLIDE 44

44

Yahoo.com

slide-45
SLIDE 45

45

slide-46
SLIDE 46

46