Collaborative Deep Learning and Its Variants for Recommender Systems
1
Hao Wang
Joint work with Naiyan Wang, Xingjian Shi, and Dit-Yan Yeung
Collaborative Deep Learning and Its Variants for Recommender Systems - - PowerPoint PPT Presentation
1 Collaborative Deep Learning and Its Variants for Recommender Systems Hao Wang Joint work with Naiyan Wang, Xingjian Shi, and Dit-Yan Yeung 2 Recommender Systems Rating matrix: Observed preferences: Matrix completion To predict: 3
1
Hao Wang
Joint work with Naiyan Wang, Xingjian Shi, and Dit-Yan Yeung
Observed preferences: To predict: Matrix completion Rating matrix:
2
Content information: Plots, directors, actors, etc.
3
Handcrafted features Automatically learn features Automatically learn features and
4
5
Stacked denoising autoencoders Convolutional neural networks Recurrent neural networks
6
7
8
9
10
11
Corrupted input Clean input [ Vincent et al. 2010 ]
12
Graphical model: Generative process: Objective function if using MAP:
latent vector of item j latent vector of user i rating of item j from user i
Notation:
[ Salakhutdinov et al. 2008 ]
13
Generalized SDAE Graphical model: Generative process:
corrupted input clean input weights and biases
Notation:
14
Graphical model: Collaborative deep learning SDAE
corrupted input clean input weights and biases content representation rating of item j from user i latent vector of item j latent vector of user i
Notation:
Two-way interaction
15
Perception Component Task-Specific Component Perception Variables Task Variables Hinge Variables [ Wang et al. TKDE 2016 ]
16
Graphical model: Collaborative deep learning SDAE
corrupted input clean input weights and biases content representation rating of item j from user i latent vector of item j latent vector of user i
Notation:
Two-way interaction
17
Neural network representation for degenerated CDL
18
Information flows from ratings to content
19
Information flows from content to ratings
20
Representation learning <-> recommendation
21
22
23
24
25
26
27
28
29
Content information Titles and abstracts Titles and abstracts Movie plots [ Wang et al. KDD 2011 ] [ Wang et al. IJCAI 2013 ]
30
Higher recall and mAP indicate better recommendation performance
31
When the ratings are very sparse: When the ratings are dense:
32
Exactly the same as Oord et al. 2013, we set the cutoff point at 500 for each user. A relative performance boost of about 50%
33
Moonstruck True Romance
34
Johnny English American Beauty
35
36
37
[ Li et al., CIKM 2015 ] Transformation to latent factors Transformation to latent factors Reconstruction error CDL: Marginalized CDL:
Reconstruction error
38
[ Ying et al., PAKDD 2016 ]
39
40
41
More details in http://wanghao.in/CDL.htm
Motivation:
word at a time, model documents as sequences
sequence generation under the BDL framework
“Collaborative recurrent autoencoder: recommend while learning to fill in the blanks” [ Wang et al., NIPS 2016a ]
42
Main Idea:
representation
“Collaborative recurrent autoencoder: recommend while learning to fill in the blanks” [ Wang et al., NIPS 2016a]
43
this a great idea this a great idea is encoder RNN decoder RNN wrong transition this a great idea is this a great idea <wc> encoder RNN decoder RNN Direct Denoising: Wildcard Denoising: Sentence: This is a great idea. -> This is a great idea.
44
Main Idea:
representation
sequences
“Collaborative recurrent autoencoder: recommend while learning to fill in the blanks” [ Wang et al., NIPS 2016a ]
45
46
length: 8 length: 6 length: 4 [ Wang et al., NIPS 2016a ]
47
0.08 0.18 0.22 0.16 0.21 0.10 0.04 0.01
8 length-3 vectors length-8 weight vector
vector
Use the area of the beta distribution to define the weights!
[ Wang et al., NIPS 2016a ]
48
0.13
6 length-3 vectors length-6 weight vector
vector
Use the area of the beta distribution to define the weights!
0.27 0.28 0.20 0.10 0.02 [ Wang et al., NIPS 2016a ]
49
[ Wang et al., NIPS 2016a ] Perception Component Task-Specific Component
50
[ Wang et al. AAAI 2017 ] [ Wang et al. AAAI 2015 ]
Generalized SDAE Graphical model: Generative process:
corrupted input clean input weights and biases
Notation:
51
corrupted input clean input adjacency matrix
Notation:
52
Perception Component Task-Specific Component
53
54
55
corrupted input clean input adjacency matrix
Notation:
Product of Q+1 Gaussians
Multiple networks: citation networks co-author networks
56
Network A → Relational Matrix S Relational Matrix S → Middle-Layer Representations
57
58
59
60
61
62
63
64
Does not appear in the tag lists of movies linked to ‘E.T. the Extra-Terrestrial’ Very difficult to discover this tag
65
Topic hierarchy Inter-document relation BDL-Based Topic Models
66
Perception component Task-Specific component
[ Wang et al. 2015 (AAAI) ] Unified into a probabilistic relational model for relational deep learning
Perception Component Task-Specific Component
67
68
[ Wang et al. 2017 (AAAI) ] Probabilistic SDAE Modeling relation among nodes
69
Motivation:
reinforcement learning, active learning, etc.
noise
“Natural-Parameter Networks: A Class
70
What We Want:
and testing
“Natural-Parameter Networks: A Class
71
neural networks weights/neurons as points natural-parameter networks weights/neurons as distributions
72
73
74