PushTrust: An Efficient Recommendation Algorithm by Leveraging - - PowerPoint PPT Presentation

pushtrust an efficient recommendation algorithm by
SMART_READER_LITE
LIVE PREVIEW

PushTrust: An Efficient Recommendation Algorithm by Leveraging - - PowerPoint PPT Presentation

PushTrust: An Efficient Recommendation Algorithm by Leveraging Trust and Distrust Relations Rana Forsati, Iman Barjasteh, Farzan Masrour, Abdol-Hossein Esfahanian, Hayder Radha Michigan State University RecSys15 Outline Some background


slide-1
SLIDE 1

PushTrust: An Efficient Recommendation Algorithm by Leveraging Trust and Distrust Relations

Rana Forsati, Iman Barjasteh, Farzan Masrour, Abdol-Hossein Esfahanian, Hayder Radha Michigan State University RecSys’15

slide-2
SLIDE 2

Outline

  • Some background
  • Problem statement
  • Our Algorithm
  • Results

2

slide-3
SLIDE 3

Recommender Systems

Recommend ¡items to ¡users ¡to ¡maximize ¡some ¡objective(s)

Latent feature models (e.g., matrix factorization): The current most successful technique as demonstrated in KDD Cup and Netflix competition

3

slide-4
SLIDE 4

Latent Feature Models

Intuition: ratings are deeply influenced by a set of non-

  • bvious features specific to the domain

Goal: Extract latent features from existing ratings for future (unknown) predictions Items: Users: Partially observed rating matrix:

4

slide-5
SLIDE 5

Latent Features Models…

How to find latent features for users and items? Assume there are k latent features for rating

5

slide-6
SLIDE 6

Matrix Factorization

Solve ¡the ¡following ¡ ¡optimization ¡problem: Model ¡each ¡user/item ¡as ¡a ¡vector ¡of ¡features ¡(learned ¡from ¡data)

: observed ratings

Training error on

  • bserved ratings

Regularization to avoid overfitting

6

slide-7
SLIDE 7

Huge ¡number ¡of ¡ unrated ¡movies

Handling cold-start users

  • r items
  • Cold-start users: new users

who have rated only a few items

  • Cold-start items: new items without ratings

? ? ? ? ?

? ? ?

? ? ? ? ? ? ? ? ?

Sparsity of rating matrix

Example: Epinions only 0.02%

  • f matrix is observed.

Challenges

7

slide-8
SLIDE 8

Side ¡ Information ¡ about ¡Items

Rating ¡Matrix

? ? ? ? ? ? ? ? ?

Goal: exploit other sources of information to cope with these challenges

Side ¡ Information ¡about ¡Items Side ¡ Information ¡ about ¡Users

Side Information

Side ¡ Information ¡ about ¡Ratings

8

slide-9
SLIDE 9

■ Trust/Distrust Relations between users

Social Recommender Systems Spatial Recommender Systems

■ Attribute of users such as location, age,….

Side Information about Users

Social Networking Services (e.g., Facebook, Twitter)

9

slide-10
SLIDE 10

Social Information and Data Sparsity

Adding side information to resolve sparsity and cold-start problems.

Our ¡focus: trust/distrust relations ¡between ¡users

How to effectively exploit social relations of users in recommendation to boost the accuracy?

Growth of social networks Trust ¡= agreement on ratings [Guo et al., 2014] ¡

Opinion Aware Recommendation Systems

Distrust ¡= disagreement on ratings [Forsati et al., 2014] ¡

10

slide-11
SLIDE 11

Matrix Factorization

& PushTrust Algorithm

11

slide-12
SLIDE 12
  • Trust can be considered as a transitive relation and can

propogate.

  • Distrust is not transitive.
  • Distrust propagates ¡only ¡one ¡step ¡[Guha et ¡al., ¡2004]
  • Distrust can not be considered as negative of trust.

Trust versus Distrust

12

slide-13
SLIDE 13

Social Regularization with Trust and Distrust

Rating Matrix Trust Network Distrust Network

Memory based methods: distrust is used to either filter out or debug propagated trust network [Victor et al., 2011] Factorization based methods: regularization by ranking of latent features [Forsati et al., 2014]

13

slide-14
SLIDE 14

Social Regularization with Trust and Distrust

Trust+Distrust enhanced MF: the latent features of trusted users are closer and distrusted users are more distant

trust network Distrust network users arranged based on similarity of their latent features

14

slide-15
SLIDE 15

Social Regularization with Trust and Distrust

where

Regularization with Social Ranking

15

slide-16
SLIDE 16

Issues with Ranking

Computationally expensive: in large social graphs,

  • ptimization cost features can increase cubically in the

number of users

.

Only consider the trusted and distrusted friends and ignores the neutral (users with no relation) users.

The neutral friends might appear before the trust friends and be negatively influential! Not scalable to large social networks!

The top portion of ranked list might include distrusted friends due to nature of pairwise ranking model

The latent features might affected by distusted friends who appear at top of the ranked list!

16

slide-17
SLIDE 17

PushTrust

A more complete approach:

  • the trustees are PUSHED

to the top of list as much as possible.

  • The foes are PUSHED to the

bottom of list as much as possible.

  • The neutral friends are

PUSHED to the middle of list.

17

slide-18
SLIDE 18

Rewrite the social-regularized matrix factorization Here is social regularization of latent features of individual users.

PushTrust: A Convex Formulation

18

slide-19
SLIDE 19

Put neutral friends above distrusted ones Put trusted friends above neutral ones Put trusted friends above distrusted ones

where is indicator function.

non-convex indicator function

PushTrust: A Convex Formulation

: latent features of trusted friends : latent features of neutral friends : latent features of distrusted friends

19

slide-20
SLIDE 20

Indicator function:

Convex Loss Function

20

slide-21
SLIDE 21

Hinge loss:

Convex Loss Function

21

slide-22
SLIDE 22

Rewrite the social regularized matrix factorization where

PushTrust Objective

22

slide-23
SLIDE 23

Optimization Procedure

The problem is non-convex jointly in both U and V. Solution: The standard gradient descent method

23

slide-24
SLIDE 24
  • 1. Updating U while keeping V fixed:

Set , where

Optimization Procedure

24

slide-25
SLIDE 25
  • 2. Updating V while keeping V fixed:

Set , where

Optimization Procedure

25

slide-26
SLIDE 26
  • Conventional:The neutral friends of users are also

incorporated in ranking the latent features.

  • Computational: The number of constraints increases

quadratically

Key Features

26

slide-27
SLIDE 27

Experiments

27

slide-28
SLIDE 28

Experimental Evaluation

Two evaluation metrics:

Mean Absolute Error: Root Mean Squared Error: where is the set of rating that should be predicted.

28

slide-29
SLIDE 29

The Epinions Dataset

The only social rating network dataset publicly available. User trust and distrust information is included in this dataset The social network in Epinions is directed

Users Users Items Items Ratings Ratings Trust Relations Trust Relations Distrust Relations Distrust Relations Quantity Quantity 121,240 685,621 12,721,437 481,799 96,823

80% of the rating data was selected as the training set with 20% as the test data.

29

slide-30
SLIDE 30

Baseline Algorithms & Experimental Design

■ MF: Matrix Factorization ■ MF-T: Matrix Factorization with Trust information ■ MF-D: Matrix Factorization with Distrust information ■ MF-DT: Matrix Factorization with Trust and Distrust ■ PushTrust: The proposed algorithm

0. 0.25 0.5 0.75 1.

5 10

MAE Latent vector dimensions PushTrust MF MF-T MF-D MF-DT

0. 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1. 1.1 1.2 1.3

5 10

RMSE Latent vector dimensions

30

slide-31
SLIDE 31

Experiment on handling cold- start users

To evaluate different algorithms:

select 30%, 20%, and 10%

  • f the users, to be cold-start

users. For cold-start users, no rating is included in the training data consider all ratings made by cold-start users as testing data.

0.75 0.8125 0.875 0.9375 1 10% 20% 30% Cold-start users

MAE

MF MF-T MF-D MF-TD PushTrust

1 1.1875 1.375 1.5625 1.75 10% 20% 30% Cold-start users

RMSE

31

slide-32
SLIDE 32

Thank You for you attention! 32 Questions?

32