on Random Walks Farzaneh Heidari Supervisor: Manos Papagelis 1 - - PowerPoint PPT Presentation

on random walks
SMART_READER_LITE
LIVE PREVIEW

on Random Walks Farzaneh Heidari Supervisor: Manos Papagelis 1 - - PowerPoint PPT Presentation

EvoNRL: Evolving Network Representation Learning Based on Random Walks Farzaneh Heidari Supervisor: Manos Papagelis 1 networks (universal language for describing complex data) 2 Classical ML Tasks in Networks ? ? ? ? ? community


slide-1
SLIDE 1

EvoNRL: Evolving Network Representation Learning Based

  • n Random Walks

Farzaneh Heidari Supervisor: Manos Papagelis

1

slide-2
SLIDE 2

(universal language for describing complex data)

networks

2

slide-3
SLIDE 3

Classical ML Tasks in Networks

3 ? ?

node classification

? ? ?

link prediction community detection anomaly detection

?

graph similarity triangle count

slide-4
SLIDE 4

expensive computation

Limitations of Classical ML Tasks

4

(high dimension computations)

Tang, Lei, and Huan Liu. "Relational learning via latent social dimensions." Proceedings of the 15th ACM SIGKDD interna

Matrix Factorization

slide-5
SLIDE 5

extensive domain knowledge

Limitations of Classical ML Tasks

5

(task specific)

Ganapathiraju, Madhavi K., et al. "Schizophrenia interactome with 504 novel protein–protein interactions." npj Schizop

slide-6
SLIDE 6

faster computations agnostic domain knowledge

Network Representation Learning (NRL)

6

(low dimension computations) (task independent)

slide-7
SLIDE 7

Network Representation Learning (NRL)

7

several network structural properties can be learned/embedded (nodes, edges, subgraphs, graphs, …)

Low-dimension space Network

slide-8
SLIDE 8

Random Walk-based NRL

8

1 2 3 4 5 6 1 7 8 9

Feed sentences to Skip-gram NN model 4 5 3 1 6 7 8 9 2

1

3 5 8 7 6 4 5

2

1 3 5 8 7 6 5 . . . . . . . .

87

8 5 4 3 5 6 7

88

4 5 6 7 8 9

89

2 1 3 5 6 7 8

90

7 4 2 1 3 5 6

Input network Obtain a set of random walks Treat the set of random walks as sentences Learn a vector representation for each node

1 2 3 4 5 6 1 7 8 9

3 5 8 7 6 4 5

slide-9
SLIDE 9

Random Walk-based NRL

9

DeepWalk node2vec …

StaticNRL

slide-10
SLIDE 10

But…

10

real-world networks are constantly changing

slide-11
SLIDE 11

11

how can we learn representations

  • f an evolving network?
slide-12
SLIDE 12

Naive Approach

12

1 2 3 4 5 6 1 7 8 9

4 5 3 1 6 7 8 9 2

t = 0

1 2 3 4 5 6 1 7 8 9

4 5 3 1 6 7 8 9 2

1 2 3 4 5 6 1 7 8 9

4 5 3 1 6 7 8 9 2

t = 1 t = 2 StaticNRL StaticNRL StaticNRL

slide-13
SLIDE 13

Limitation #1

13

1 2 3 4 5 6 1 7 8 9

1

3 5 8 7 6 4 5

2

1 3 5 8 7 6 5 . . . . . . . .

87

8 5 4 3 5 6 7

88

4 5 6 7 8 9

89

2 1 3 5 6 7 8

90

7 4 2 1 3 5 6

1 2 3 4 5 6 1 7 8 9

time expensive

slide-14
SLIDE 14

Limitation #2

14

incomparable representations

1 2 3 4 5 6 1 7 8 9

t = 0

4 5 3 1 6 7 8 9 2

1 2 3 4 5 6 1 7 8 9

4 5 3 1 6 7 8 9 2

t = 1

1 2 3 4 5 6 1 7 8 9

Random Walks Neural Network Optimization

slide-15
SLIDE 15

EvoNRL Key Idea

15

1 2 3 4 5 6 1 7 8 9

Feed sentences to Skip-gram NN model 4 5 3 1 6 7 8 9 2

1

3 5 8 7 6 4 5

2

1 3 5 8 7 6 5 . . . . . . . .

87

8 5 4 3 5 6 7

88

4 5 6 7 8 9

89

2 1 3 5 6 7 8

90

7 4 2 1 3 5 6

Input network Obtain a set of random walks Treat the set of random walks as sentences Learn a vector representation for each node

1 2 3 4 5 6 1 7 8 9

slide-16
SLIDE 16

16

dynamically maintain a set

  • f random walks for every

change in the network

slide-17
SLIDE 17

Example

7 1 2 3 4 5 6 1 7 8 9

t = 0 t = 1

1 2 3 4 5 6 7 8 9

1

3 5 8 7 6 4 5

2

1 3 5 8 7 6 5 . . . . . . . .

87

8 5 4 3 5 6 7

88

4 5 6 7 8 9

89

2 1 3 5 6 7 8

90

7 4 2 1 3 5 6

addition of edge (1, 4)

1

3 5 8 7 6 4 5

2

1 3 5 8 7 6 5 . . . . . . . .

87

8 5 4 3 5 6 7

88

4 5 6 7 8 9

89

2 1 3 5 6 7 8

90

7 4 2 1 3 5 6

need to update the RW set 1 2 3 4 1

2

1 4 3 5 6 7 8

{

simulate the rest of the RW

17

slide-18
SLIDE 18

18

how can we efficiently maintain a set of random walks?

slide-19
SLIDE 19

EvoNRL Operations: edge addition

1 2 3 4 5 6 1 7 8 9

1

3 5 8 7 6 4 5

2

1 3 5 8 7 6 5 . . . . . . . .

87

8 5 4 3 5 6 7

88

4 5 6 7 8 9

89

2 1 3 5 6 7 8

90

7 4 2 1 3 5 6

1 2 3 4 5 6 7 8 9

1

3 5 8 7 6 4 5

2

1 3 5 8 7 6 5 . . . . . . . .

87

8 5 4 3 5 6 7

88

4 5 6 7 8 9

89

2 1 3 5 6 7 8

90

7 4 2 1 3 5 6

+ edge(n1, n2)

2

1 4 3 5 6 7 8

Operations on RW Search a node Delete a RW Insert a new RW

19

slide-20
SLIDE 20

EvoNRL Operations: edge deletion

1 2 3 4 5 6 1 7 8 9

1

3 5 8 7 6 4 5

2

1 4 5 8 7 6 5 . . . . . . . .

87

8 5 4 3 5 6 7

88

4 5 6 7 8 9

89

2 4 1 5 6 7 8

90

7 4 2 1 3 5 6

1 2 3 4 5 6 7 8 9

1

3 5 8 7 6 4 5

2

1 4 5 8 7 6 5 . . . . . . . .

87

8 5 4 3 5 6 7

88

4 5 6 7 8 9

89

2 4 1 5 6 7 8

90

7 4 2 1 3 5 6

  • edge(n1, n2)

2

1 3 4 5 6 7 8

Operations on RW Search two nodes Delete a RW Insert a new RW

20

slide-21
SLIDE 21

EvoNRL Operations: node addition

1 2 3 4 5 6 1 7 8 9

1

3 5 8 7 6 4 5

2

1 3 5 8 7 6 5 . . . . . . . .

87

8 5 4 3 5 6 7

88

4 5 6 7 8 7

89

2 1 3 5 6 7 8

90

7 4 2 1 3 5 6

1 2 3 4 5 6 7 8 9

1

3 5 8 7 6 4 5

2

1 3 5 8 7 6 5 . . . . . . . .

87

8 7 4 3 5 6 7

88

4 5 6 7 8 5

89

2 1 3 5 6 7 8

90

7 4 2 1 3 5 6

91

9 8 7 4 7 6 5

.

9 . . .

.

9 . . .

100

9 8 5 6 4 3 1

+ node(n1)

2

1 4 8 9 8 7 6

Operations on RW Search a node (node #8) Delete a RW Insert a new RW Append the RW list

{

21

slide-22
SLIDE 22

EvoNRL Operations: node deletion

1 2 3 4 5 6 1 7 8 9

1

3 5 8 7 6 4 5

2

1 3 5 8 9 6 5 . . . . . . . .

87

8 5 4 3 5 6 7

88

4 5 6 7 8 7

89

2 1 3 5 6 7 8

90

7 4 2 1 3 5 6

1 2 3 4 5 6 7 8 9

1

3 5 8 7 6 4 5

2

1 3 5 8 9 6 5 . . . . . . . .

87

8 7 4 3 5 6 7

88

4 5 6 7 8 5

89

2 1 3 5 6 7 8

90

7 4 2 1 3 5 6

91

9 8 7 4 7 6 5

.

9 . . .

.

9 . . .

100

9 8 5 6 4 3 1

  • node(n1)

2

1 4 8 5 8 7 6

Operations on RW Search two nodes Delete a RW Insert a new RW Deduct the RW list

{

22

slide-23
SLIDE 23

EvoNRL Indexing

1 2 3 4 5 6 1 7 8 9 each node is a keyword each RW is a document a set of RWs is a collection of documents

1

3 5 8 7 6 4 5

2

1 3 5 8 7 6 5 . . . . . . . .

87

8 5 4 3 5 6 7

88

4 5 6 7 8 9

89

2 1 3 5 6 7 8

90

7 4 2 1 3 5 6

Term

Frequency Postings and Positions

1

3 < 2, 1 >, < 89, 2 >, < 90, 4 >

2

2 <89, 1>, <90, 3>

3

5 <1, 1>, <2, 1>, <87, 3>, <89, 3>, <90, 5>

4

4 <1, 6>, <87, 3>, <90, 2>

5

9 <1, 2>, <1, 7>, <2, 3>, <2, 7>, <87, 5>, <88, 2>, <89, 4>, <90, 6>

6

6 <1, 5>, <2, 6>, <87, 6>, <88, 3>, <89, 3>, <90, 5>

7

5 <1, 4>, <2, 5>, <87, 7>, <88, 4>, <89, 6>, 90, 7>

8

5 <1, 3>, <2, 4>, <87, 1>, <88, 6>, <89, 7>

9

1 <88, 7>

23

slide-24
SLIDE 24

24

when is the good time to obtain a new representation of the network?

slide-25
SLIDE 25

Arriving Edge Importance

1 2 3 4 5 6 7 8 9

1

3 5 8 7 6 4 5

2

1 3 5 8 7 6 5 . . . . . . . .

87

8 5 4 3 5 6 7

88

4 5 6 7 8 7

89

2 1 3 5 6 7 8

90

7 4 2 1 3 5 6

count #RW changed edge addition

1 2 3 4 5 6 7 8 9

edge addition

1

3 5 8 7 6 4 5

2

1 3 5 8 7 6 5 . . . . . . . .

87

8 5 4 3 5 6 7

88

4 5 6 7 8 7

89

2 1 3 5 6 7 8

90

7 4 2 1 3 5 6

count #RW changed

count the #RW each edge changes

2 4

25

slide-26
SLIDE 26

Adaptive Algorithm

1 2 3 4 5 6 7 8 9

1

3 5 8 7 6 4 5

2

1 3 5 8 7 6 5 . . . . . . . .

87

8 5 4 3 5 6 7

88

4 5 6 7 8 7

89

2 1 3 5 6 7 8

90

7 4 2 1 3 5 6

edge addition count #RW changed

#RW changed # edges

repeat for upcoming edges

26

slide-27
SLIDE 27

Peak Detection

𝜐 × 𝑡𝑢𝑒[𝑢] + 𝑏𝑤𝑕[𝑢]

𝑏𝑤𝑕[𝑢]

27

slide-28
SLIDE 28

When to re-embed?

28

stream of edges

1

3 5 8 7 6 4 5

2

1 3 5 8 7 6 5 . . . . . . . .

87

8 5 4 3 5 6 7

88

4 5 6 7 8 9

89

2 1 3 5 6 7 8

90

7 4 2 1 3 5 6

(n1, n2) (n3, n4) (n5, n6) (n7, n8) (nk, nk+1) (nk+2, nk+3) (nk+4, nk+5) (nk+5, nk+6) (n1, n2) (n3, n4) (n5, n6) (n7, n8) (nk, nk+1) (nk+2, nk+3) (nk+4, nk+5) (nk+5, nk+6)

. . .

. . .

Informed decision based

  • n edge importance

(monitor #rw updated so far)

Embed Embed Embed Embed

Periodic Adaptive

slide-29
SLIDE 29

Evaluation: EvoNRL vs StaticNRL Running Time

◼EvoNRL << StaticNRL

29

slide-30
SLIDE 30

EvoNRL has the similar accuracy as StaticNRL

Accuracy: edge addition & edge deletion

30

slide-31
SLIDE 31

EvoNRL has the similar accuracy as StaticNRL

Accuracy: node addition & node deletion

31

slide-32
SLIDE 32

Time Performance

32

EvoNRL performs orders of time faster than StaticNRL

100x 𝟑𝟏𝐲

slide-33
SLIDE 33

Decision-making Performance

Periodic-50 Adaptive Periodic-100

33

slide-34
SLIDE 34

Related Work

34

1 2 3 4 5 6 1 7 9 1 2 2 4 5 1 8 9

t

1 2 1

𝐻1 𝐻𝑙 𝐻𝑜

Nguyen, Giang Hoang, et al. "Continuous-time dynamic network embeddings." Companion Proceedings of the The Web Conference

Uses outdated temporal information

slide-35
SLIDE 35

Summary

35

EvoNRL

time efficient accurate generic method

how can we learn representations of an evolving network?

slide-36
SLIDE 36

Thank you!

36

slide-37
SLIDE 37

Questions?

37

slide-38
SLIDE 38

References

[Complex Networks 2018] EvoNRL: Evolving Network Representation Learning Based on Random Walks. Farzaneh Heidari, Manos Papagelis. Proceedings of the 7th International Conference on Complex Networks and Their Applications. [KDD 2009] Relational learning via latent social dimensions. Tang Lei, Huan Liu. Proceedings of the 20th ACM SIGKDD international conference on Knowledge Discovery and Data Mining. [ACM SIGKDD 2014] Deepwalk: Online Learning of Social Representations. Perozzi, Bryan, Rami Al-Rfou, and Steven

  • Skiena. Proceedings of the 20th ACM SIGKDD international conference on Knowledge Discovery and Data Mining.

[ACM SIGKDD 2016] node2vec: Scalable Feature Learning for Networks. Aditya Grover and Jure Leskovec. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. [ACM SIGKDD 2015] PTE: Predictive Text Embedding Through Large-scale Heterogeneous Text Networks. Jian Tang, Meng Qu, and Qiaozhu Mei. Proceedings of the 21st ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. [WWW 2015] LINE: Large-scale Information Network Embedding. Jian Tang, Meng Qu, Mingzhe Wang, Ming Zhang, Jun Yan, and Qiaozhu Mei. Proceedings of the 24th International World Wide Web Conference. [ACL 2016] Diachronic Word Embeddings Reveal Statistical Laws of Semantic Change. William L. Hamilton, Jure Leskovec, and Dan Jurafsky. Proceedings of the Annual meeting of the Association for Computational Linguistics. [WWW 2018] Continuous-time dynamic network embeddings. Giang Hoang Nguyen, John Boaz Lee, Ryan A. Rossi, Nesreen K. Ahmed, Eunyee Koh, and Sungchul Kim. Proceedings of the International World Wide Web Conferences Steering Committee, 2018. 38