Recent advances in document network embedding @ERIC Julien Velcin - - PowerPoint PPT Presentation

recent advances in document network embedding eric
SMART_READER_LITE
LIVE PREVIEW

Recent advances in document network embedding @ERIC Julien Velcin - - PowerPoint PPT Presentation

Dynamics On and Of Complex Networks 2020 Recent advances in document network embedding @ERIC Julien Velcin julien.velcin@univ-lyon2.fr Universit Lumire Lyon 2 - ERIC Lab Context 2 Informational landscape Projet Pulseweb (Cointet,


slide-1
SLIDE 1

Recent advances in document network embedding @ERIC

Julien Velcin

julien.velcin@univ-lyon2.fr

Université Lumière Lyon 2 - ERIC Lab

Dynamics On and Of Complex Networks 2020

slide-2
SLIDE 2

Context

2

slide-3
SLIDE 3

Informational landscape

Projet Pulseweb (Cointet, Chavalarias…) Chronolines (Nguyen et al., 2014) Metromaps (Shahaf et al., 2015) https://github.com/Erwangf/readitopics http://pulseweb.cortext.net Readitopics (Velcin et al., 2018)

3

slide-4
SLIDE 4

Document network embedding

  • Document network: “graph of vertices, where each vertex is

associated with a text document” (Tuan et al., 2014)
 e.g.: scientific articles, newspapers, social media…

  • Embedding for building a joint space for solving downstream

tasks (e.g., link prediction, node classification, community detection)

embedding

A complex system is a system composed of many components which may interact with each other. Examples of complex systems are Earth's global climate, organisms, the human brain, infrastructure such as power grid, transportation or communication systems, social and economic organizations (like cities), an ecosystem, a living cell, and ultimately the entire universe. Complex systems are systems whose behavior is intrinsically difficult to model due to the dependencies, competitions, relationships, or other types of interactions between their parts or between a given system and its environment.

classification link prediction clustering visualisation

4

slide-5
SLIDE 5

Quick survey

  • Graph/Node embedding
  • Laplacian Eigenmaps (Belkin and Niyogi, 2002)
  • DeepWalk (Perozzi et al., 2014), Node2vec (Grover and

Leskovec, 2016)

  • Graph Neural Networks (Scarselli et al., 2009)
  • Document network embedding
  • TADW (Yang et al., 2015)
  • Attention models and CANE (Tu et al., 2017)

5

slide-6
SLIDE 6

Collaborators of the DMD team

Robin Brochier Phd student (now graduated!) Antoine Gourru Phd student Adrien Guille Associate Professor Julien Jacques Professor

6

slide-7
SLIDE 7

Contributions

7

slide-8
SLIDE 8

Regularized Linear Embedding (RLE)

Given:

  • matrix of pretrained word embeddings
  • Document x Word matrix (textual information)
  • the transition matrix (graph information)
  • Goal: learn the weights

for the words composing The vector for is just a weighted sum over pretrained WE

U ∈ ℝv×k T ∈ ℝn×v A ∈ ℝ[0,1]×[0,1] pi ∈ ℝv di di = piU di

Gourru A., J. Velcin, J. Jacques and A. Guille
 Document Network Projection in Pretrained Word Embedding Space. ECIR 2020.

8

parameter to learn

slide-9
SLIDE 9

RLE (con’t)

with a tradeoff b/w textual and structural information with a squared matrix that reflects the pairwise similarity between nodes in the graph
 (here, we use )

P = (1 − λ)T + λB λ ∈ [0,1] bi = 1 ∑j Si,j ∑

j

Si,jtj S ∈ ℝn×n S = A + A2 2

9

slide-10
SLIDE 10

Evaluation

Datasets:

  • Cora (2,211 docs; 7 labels=topic; 5,001 citation links)
  • DBLP (60,744 docs; 4 labels=topic; 52,914 links)
  • New York Times (5,135 docs; 4 labels=article section;

3,050,513 links=common tag)

https://github.com/AntoineGourru/DNEmbedding

  • Task 1: node classification
  • Task 2: link prediction

10

slide-11
SLIDE 11

11

sensitivity to λ

slide-12
SLIDE 12

GVNR and GVNR-t

  • Quick reminder of DeepWalk (Perozzi et al., 2014):
  • goal: learn vector representation of nodes
  • approach: a) make multiple random walks

b) paths views as documents c) use Skip-Gram to build vectors (Mikolov et al., 2013)

12

Brochier, A., Guille and J. Velcin. Global Vectors for Node Representation. The Web Conference (WWW) 2019.

target context

slide-13
SLIDE 13

GVNR and GVNR-t

  • Following GloVe (Pennington et al., 2014, GVNR solves

regression task on the weighted cooccurrence matrix X where cells with small values are set to 0 (> threshold )

  • We’re looking for

and s.t.:

with

  • GVNR-t integrates textual information by modifying :

xmin (U, bU) (V, bV)

arg min

U,V,bU,bV n

i n

j

s(xij)(ui . vj + bU

i + bV j − log(c + xij))2

s(xij) = 1 if xij > 0 and mi ∼ B(α) else

where α is chosen s.t. m = k in average

vj

arg min

U,V,bU,bV n

i n

j

s(xij)(ui . δj . W |δj|1 + bU

i + bV j − log(c + xij))2

Brochier, A., Guille and J. Velcin. Global Vectors for Node Representation. The Web Conference (WWW) 2019.

13

cooccurrence b/w and

xi xj

U,W

1

slide-14
SLIDE 14

Results for GVNR-t

  • Classification on two citation networks (Cora with 2,708

nodes and Citeseer with 3,312 nodes)

  • Keyword recommendation on DBLP (1,397,240 documents

and 3,021,489 citation relationships)

https://github.com/brochier/gvnr

14

slide-15
SLIDE 15

Inductive Document Network Embedding (IDNE)

Topical attention

shared parameters W and T

Topical attention

xi ∈ ℝnw di ∈ ℝp σ(di . dj) xj ∈ ℝnw dj ∈ ℝp

15

Brochier R., A. Guille and J. Velcin.
 Inductive Document Network Embedding with Topic-Word Attention. ECIR 2020 (virtual).

0 = no link 1 = link

slide-16
SLIDE 16

Topical attention

16

dot products topical attention weights Z K topical vector For document

di

p p K topics K words

WT W

p

Document representation is the normalized sum over the K topics:

di ∑K

k=1 u(i|k) |xi|1

slide-17
SLIDE 17

Learning IDNE

Minimize so that:

  • S is a binary similarity matrix based on A, for instance:

L(W, T) = ∑nd

i=1 ∑nd j=1 sij log σ(ui . uj) + (1 − sij)log σ(−ui . uj)

sij = 1 if (A + A2)ij > 0 else sij = 0

17

slide-18
SLIDE 18

Results of IDNE on Cora

18

T for Transductive I for Inductive C = classification P = link prediction

slide-19
SLIDE 19

Observations

19

slide-20
SLIDE 20

Observations (con’t)

20

MCMC + theory Decision trees

slide-21
SLIDE 21

Conclusion and future works

21

slide-22
SLIDE 22

Conclusion

  • Several contributions on the embedding of documents

augmented with network information:
 RLE, GVNR-t, MATAN, IDNE, GELD

  • Use of “absolute” WE leads to good results. Can they be

improved using contextualized WE (Devlin et al., 2018)

  • Recent advances in GNN should be considered in the

future, e.g. GAT (Velikcovik et al., 2018)

22

slide-23
SLIDE 23

Future works

  • Integrating uncertainty in the modelling


(Gourru et al., 2020)

  • Moving to author embedding (Ganesh et al., 2016) and

modeling dynamics following (Balmer et al., 2017)

  • Information diffusion in information networks (work in

progress with G. Poux and S. Loudcher)

23

slide-24
SLIDE 24

References

  • Brochier R., A. Guille and J. Velcin. Inductive Document Network Embedding with Topic-

Word Attention. ECIR 2020 (virtual).

  • Brochier R., A. Guille and J. Velcin. Link Prediction with Mutual Attention for Text-

Attributed Networks. Workshop on Deep Learning for Graphs and Structured Data Embedding, colocated with WWW (Companion Volume), May 13–17, 2019, San Francisco, CA, USA.

  • Brochier R., A. Guille and J. Velcin. Global Vectors for Node Representation. The Web

Conference (WWW), May 13–17, 2019, San Francisco, CA, USA.

  • Gourru A., J. Velcin, J. Jacques and A. Guille Document Network Projection in Pretrained

Word Embedding Space. ECIR 2020 (virtual).

  • Gourru A., J. Velcin and J. Jacques. Gaussian Embedding of Linked Documents from a

Pretrained Semantic Space. IJCAI 2020.

➡ Code for GVNR and GVNR-t: https://github.com/brochier/gvnr ➡ Code for IDNE: https://github.com/brochier/idne ➡ Code for RLE and GELD: https://github.com/AntoineGourru/DNEmbedding

24