Relational Deep Learning: A Deep Latent Variable Model for Link - - PowerPoint PPT Presentation

relational deep learning
SMART_READER_LITE
LIVE PREVIEW

Relational Deep Learning: A Deep Latent Variable Model for Link - - PowerPoint PPT Presentation

Relational Deep Learning: A Deep Latent Variable Model for Link Prediction Hao Wang, Xingjian Shi, Dit-Yan Yeung Motivation Bayesian Deep Learning Relational Deep Learning Parameter Learning Experiments Conclusion


slide-1
SLIDE 1

Relational Deep Learning: A Deep Latent Variable Model for Link Prediction

Hao Wang, Xingjian Shi, Dit-Yan Yeung

slide-2
SLIDE 2
  • Motivation
  • Bayesian Deep Learning
  • Relational Deep Learning
  • Parameter Learning
  • Experiments
  • Conclusion
slide-3
SLIDE 3

Motivation: Link Prediction

Social Network Analysis (e.g., prediction friendship in Facebook)

slide-4
SLIDE 4

Motivation: Link Prediction

Document Networks (e.g., citation networks, co-author networks)

slide-5
SLIDE 5

Link Prediction Accuracy Links Links & Content Links & Extracted Content Feature Using DL Deep Latent Variable Model

Motivation: Deep Latent Variable Models

slide-6
SLIDE 6

Stacked denoising autoencoders Convolutional neural networks Recurrent neural networks

Typically for i.i.d. data

Motivation: Deep Latent Variable Models

slide-7
SLIDE 7
  • Motivation
  • Bayesian Deep Learning
  • Relational Deep Learning
  • Parameter Learning
  • Experiments
  • Conclusion
slide-8
SLIDE 8

Bayesian Deep Learning

Perception component Task-Specific component Bayesian deep learning (BDL)

  • Maximum a posteriori (MAP)
  • Markov chain Monte Carlo (MCMC)
  • Variational inference (VI)

Content understanding Posts by users Text in articles Target task Link prediction

[ Wang et al. 2016 ]

slide-9
SLIDE 9

[ Wang et al. 2016 ]

Bayesian Deep Learning

slide-10
SLIDE 10

Perception Component Task-Specific Component Perception Variables Task Variables Hinge Variables [ Wang et al. 2016 ]

A Principled Probabilistic Framework

slide-11
SLIDE 11
  • Motivation
  • Bayesian Deep Learning
  • Relational Deep Learning
  • Parameter Learning
  • Experiments
  • Conclusion
slide-12
SLIDE 12

Perception component: relational and deep representation learning Task-specific component: link prediction

Relational Deep Learning: Graphical Model

slide-13
SLIDE 13

Corrupted input Clean input [ Vincent et al. 2010 ]

Stacked Denoising Autoencoders (SDAE)

slide-14
SLIDE 14

Generalized SDAE Graphical model: Generative process:

corrupted input clean input weights and biases

Notation:

Probabilistic SDAE

[ Wang et al. 2015 ]

slide-15
SLIDE 15

Probabilistic SDAE Modeling relation among nodes

Relational Deep Learning

slide-16
SLIDE 16

Network of Probabilistic SDAE

Many interconnected probabilistic SDAEs with shared weights

slide-17
SLIDE 17
  • Motivation
  • Bayesian Deep Learning
  • Relational Deep Learning
  • Parameter Learning
  • Experiments
  • Conclusion
slide-18
SLIDE 18

MAP Inference

maximizing the posterior probability is equivalent to maximizing the joint log-likelihood

slide-19
SLIDE 19

MAP Inference

Prior (regularization) for link prediction parameters, weights, and biases

slide-20
SLIDE 20

MAP Inference

Generating node features from content representation with Gaussian offset

slide-21
SLIDE 21

‘Generating’ clean input from the output of probabilistic SDAE with Gaussian offset

MAP Inference

slide-22
SLIDE 22

Generating the input of Layer l from the output of Layer l-1 with Gaussian offset

MAP Inference

slide-23
SLIDE 23

MAP Inference

Generating links from Bernoulli distributions parameterized by η and φ

slide-24
SLIDE 24

Bayesian Treatment: Generalized Variational Inference

Use Laplace approximation rather than variational inference for weights/biases.

slide-25
SLIDE 25

Example: Updating φ as a Product of Gaussians

Update φ for node i as a product of two Gaussians

First Gaussian Second Gaussian

slide-26
SLIDE 26
  • Motivation
  • Bayesian Deep Learning
  • Relational Deep Learning
  • Parameter Learning
  • Experiments
  • Conclusion
slide-27
SLIDE 27

Experiments: Settings

Document Networks (e.g., citation networks)

slide-28
SLIDE 28

Experiments: Link Rank and AUC

Link rank: how high our predicted links rank in the ground truth AUC: area under curve

slide-29
SLIDE 29

Experiments: Link Rank and AUC

Link rank: how high our predicted links rank in the ground truth AUC: area under curve

slide-30
SLIDE 30

Experiments: Link Rank and AUC

Link rank: how high our predicted links rank in the ground truth AUC: area under curve

slide-31
SLIDE 31

Experiments: RDL Variants

Link rank of baselines (the first 3 columns) and RDL variants (the last 4 columns) on three datasets (L = 4) VAE: Variational Autoencoder VRAE: Variational Fair Autoencoder BLR: Bayesian Logistic Regression BSDAE1: Bayesian treatment of probabilistic SDAE (mean only) BSDAE2: Bayesian treatment of probabilistic SDAE (mean and variance) MAPRDL: RDL with MAP inference BayesRDL: RDL with full Bayesian treatment

slide-32
SLIDE 32

Experiments: Depth

Performance of RDL with different number of layers (MAP) Performance of RDL with different number of layers (Bayesian treatment)

slide-33
SLIDE 33

Case Study: RDL and RTM

t-SNE visualization of latent factors learned by RDL (left) and RTM (right). Target article: From DNA sequence to transcriptional behaviour: a quantitative approach (red): articles with links to the target article (blue): articles without links to the target article

slide-34
SLIDE 34

Case Study: RDL

t-SNE visualization of latent factors learned by RDL. Articles written in German, which are rare in the datasets Some bestselling books: The 4-Hour Work Week Mary Bell’s Complete Dehydrator Cookbook Target article: From DNA sequence to transcriptional behaviour: a quantitative approach

slide-35
SLIDE 35

Case Study: RDL ang gRTM

Key Concepts Object recognition Unsupervised learning Scale-invariant learning Top 10 link predictions made by gRTM and RDL for two articles from citeulike-a

slide-36
SLIDE 36

Case Study: RDL ang gRTM

Top 10 link predictions made by gRTM and RDL for two articles from citeulike-a Key Concepts Protein structures Protein databases

slide-37
SLIDE 37
  • Motivation
  • Bayesian Deep Learning
  • Relational Deep Learning
  • Parameter Learning
  • Experiments
  • Conclusion
slide-38
SLIDE 38

Conclusion

  • First Bayesian DL model for link prediction
  • Joint Bayesian DL is beneficial
  • Significant improvement on the state of the art
  • RDL as representation learning
slide-39
SLIDE 39

Future Work

  • Multi-relational data (co-author & citation networks)
  • Boost predictive performance
  • Discover relationship between different networks
  • GVI for other neural nets (CNN/RNN) and BayesNets
  • pSDAE + link prediction
  • pCNN + recommendation
  • pRNN + community detection
  • Replace probabilistic SDAE with other Bayesian neural nets
  • Variational autoencoders
  • Natural-parameter networks
slide-40
SLIDE 40

www.wanghao.in hwangaz@connect.ust.hk