Background p Network A ubiquitous data structure to model the - - PowerPoint PPT Presentation

background
SMART_READER_LITE
LIVE PREVIEW

Background p Network A ubiquitous data structure to model the - - PowerPoint PPT Presentation

BiNE : Bi partite N etwork E mbedding ACM SIGIR 2018, July 8, Ann Arbor Michigan, U.S.A. MingGao * , Leihui Chen * , Xiangnan He + , Aoying Zhou * * East China Normal University + National University of Singapore Background p Network A


slide-1
SLIDE 1

MingGao*, Leihui Chen*, Xiangnan He+, Aoying Zhou*

*East China Normal University +National University of Singapore

BiNE: Bipartite Network Embedding

ACM SIGIR 2018, July 8, Ann Arbor Michigan, U.S.A.

slide-2
SLIDE 2

Homogeneous Network

2 ü Item adoption ü Web visiting ü Question answering ü …

pNetwork

Ø A ubiquitous data structure to model the relationships between entities

pNetwork embedding

Ø Crucial to obtain the representations for vertices Ø Helpful to many applications, such as vertex labeling, link prediction, recommendation, and clustering, etc.

Background

Heterogeneous Network

ü Social network ü Collaboration network ü Transportation network ü …

slide-3
SLIDE 3

p Homogeneous network embedding: Ø Ignore type information of vertices (e.g., Node2vec, DeepWalk, etc.) Ø Ignore key characteristic of bipartite network -- power-law distribution of vertex degrees Heterogeneous network embedding: Ø MetaPath2vec [Dong et al, KDD’17] treats explicit and implicit relations as contributing equally

3

Drawbacks of Existing Works for Bipartite Networks

slide-4
SLIDE 4

p Background & Motivations p Proposed Method p Experiments and Results p Conclusions

Outline

slide-5
SLIDE 5

5

BiNE: Bipartite Network Embedding

!

"

!# !$ %

"

%# %$

&" " &" # &" $ &# # &$ # &$ $

… … !

"

!# !$ … … !

"

!# !$ %

"

%# %$ … … %

"

%# %$ Input Capture explicit relations Obtain implicit relations Jointly model explicit and implicit relations .2 .3 .5 1 .7 .4 .3 .5 .1 .2 .2 .6 .5 .9 .1 … … … … …

|(| |U|

.2 .1 .2 1 .7 .3 .4 .5 .5 .7 .1 .6 .5 .9 .1 … … … … …

|(| |*|

!

"

!# !$ %

"

%# %$

&" " &" # &" $ &# # &$ # &$ $

… …

BiNE

+ = (. , * , W) 2 ∶ . ∪ * → ℝ7

p Two Characteristics of BiNE

Ø Modeling the explicit and implicit relations simultaneously Ø A biased and self-adaptive random walk generator

slide-6
SLIDE 6

p Original network space

The joint probability between vertices !" and #$ is defined as:

6

Modeling Explicit Relations (Observed links)

pEmbedding space

The joint probability between vertices !" and #$ is estimated as:

pPreserving the local proximity

Minimizing the difference (KL- divergence) between the two distributions:

slide-7
SLIDE 7

p Constructing Corpus of Vertex

Sequences Ø Construct U-U and V-V networks Ø Run Self-adaptive random walker

1) # of walks starting from a vertex depends on its centrality score. 2) Length of a vertex sequence is controlled by a stop probability.

p Optimizing a point-wise classification loss to capture the high-order correlations

7

Modeling Implicit Relations (High-order relations)

!

"

!# !$ … … !

"

!# !$ %

"

%# %$ … … %

"

%# %$ !

"

!# !$ %

"

%# %$

&" " &" # &" $ &# # &$ # &$ $

… …

slide-8
SLIDE 8
  • A. Taking corpus of users !" as

example , given a sequence #, $%(=2) and a vertex &':

B. C.

p Assumption: vertices frequently co-occurred in the same context of a sequence should be assigned to similar embeddings.

8

Capturing the High-order Relations

&( &) &* &+ &, &- &. &/ #: &'

Sample High-quality and Diverse Negatives with Locality Sensitive Hashing (LSH)

slide-9
SLIDE 9

p A joint optimization framework

9

Joint Optimization

Explicit relations Implicit relations

slide-10
SLIDE 10

p Background & Motivations p Proposed Method p Experiments and Results p Conclusions

Outline

10

slide-11
SLIDE 11

p Tasks

Ø Two tasks: link prediction (classification) & recommendation (ranking)

p Datasets and Metrics pResearch Questions

Ø RQ1 Performance of BiNE compared to representative baselines Ø RQ2 Is the implicit relations helpful? Ø RQ3 Effect of random walk generator

11

Experimental Setting-up

slide-12
SLIDE 12

p Network embedding methods

Ø DeepWalk [Perozzi et al KDD 2014] Ø LINE [Tang et al WWW 2015] Ø Node2vec [Grover et al KDD 2016] Ø Metapath2vec++ [Dong et al KDD 2017]

p Link Prediction methods [Xia et al

ASONAM 2012] Ø JC (Jaccard coefficient) Ø AA (Adamic/Adar) Ø Katz (Katz index) Ø PA (Preferential attachment)

12

Baselines

p Recommendation methods

Ø BPR [Rendle et al UAI 2009] Ø RankALS [Takács et al Recsys 2012] Ø FISMauc [Kabbur et al KDD 2013]

slide-13
SLIDE 13

13

RQ1: Performance of Link Prediction

Observations:

  • 1. Data-dependent

supervised manner is more advantageous.

  • 2. Positive effect of modeling

both explicit and implicit relations into the embedding process.

  • 3. Effectiveness of modeling

the explicit and implicit relations in diffferent ways.

slide-14
SLIDE 14

14

RQ2: Performance of Recommendation

Observations:

  • 1. Positive effect of considering information of weight
  • 2. Importance of focusing on the higher-order proximities among vertices
  • 3. Jointly training is superior to separately training + post-processing
slide-15
SLIDE 15

15

Utility of Implicit Relations (RQ2)

Observation: Modeling high-order implicit relations is effective to complement with explicit relation modeling.

slide-16
SLIDE 16

16

Random Walk Generator (RQ3)

Observation: The biased and self-adaptive random walk generator contributes to learning better vertex embeddings.

slide-17
SLIDE 17

17

Random Walk Generator (RQ3)

Observation: The biased and self-adaptive random walk generator contributes to learning better vertex embeddings.

(c) Self-Adaptive generator Distribution of vertex degree DeepWalk Generator: Our Generator:

slide-18
SLIDE 18

19

Case Study

slide-19
SLIDE 19

p Conclusions

ØPropose a dedicated approach for embedding bipartite networks ØJointly model both the explicit relations and higher-order implicit relations ØExtensive experiments on several tasks of link prediction, recommendation, and visualization

p Future work

ØExtend our BiNE method to model auxiliary side info ØInvestigate how to efficiently refresh embeddings for dynamic bipartite networks ØNetwork embedding + adversarial training

20

Conclusions

slide-20
SLIDE 20

Acknowledgments

p National Natural Science Foundation of China p The Press of East China Normal University p National Research Foundation, Prime Minister’s Office, Singapore pMing Gao ()

(East China Normal University)

p Leihui Chen ()

(East China Normal University)

p Aoying Zhou ()

(East China Normal University) 25

slide-21
SLIDE 21

Code available:

Thank You for Your Attention

slide-22
SLIDE 22

p Optimizing a point-wise classification loss Ø p(!"|!#) can be approximate as: Ø Following the similar formulations, we can get the counterparts for the conditional probability p($|%#)

23

Negative Sampling

LSH-based

slide-23
SLIDE 23

p LSH-based negative sampling method Ø For a center vertex !", high-quality negatives should be the vertices that are dissimilar from !"

24

LSH-based Negative Sampling

Frequency-based or popularity-based sampling LSH-based negative sampling Strategy High frequency objects Dissimilar objects Word Embedding Useless words Network Embedding Popular items or active users

slide-24
SLIDE 24

pPerformance of BiNE with different negative sampling strategies.

Experimental Results

Observations:

  • 1. Two methods show roughly

equivalent performance in most case.

  • 2. However, there are situations

(see VisualizeUS) in which LSH- based sampling method uses dissimilar information obtained from user behavior data can generate more reasonable negative samples

29