distributed egocentric representations of graphs for
play

Distributed, Egocentric Representations of Graphs for Detecting - PowerPoint PPT Presentation

Distributed, Egocentric Representations of Graphs for Detecting Critical Structures Ruo-Chun Tzeng 1 Shan-Hung Wu 2 1 Microsoft Inc, Taiwan 2 National Tsing Hua University, Taiwan International Conference on Machine Learning, 2019 Ruo-Chun Tzeng


  1. Distributed, Egocentric Representations of Graphs for Detecting Critical Structures Ruo-Chun Tzeng 1 Shan-Hung Wu 2 1 Microsoft Inc, Taiwan 2 National Tsing Hua University, Taiwan International Conference on Machine Learning, 2019 Ruo-Chun Tzeng and Shan-Hung Wu Ego-CNN ICML’19 1 / 9

  2. Goal To learn representations of graphs by using convolutions Ruo-Chun Tzeng and Shan-Hung Wu Ego-CNN ICML’19 1 / 9

  3. Goal To learn representations of graphs by using convolutions ...while keeping nice properties of CNNs on images: Filters detect location independent patterns Filters at a deep layer have enlarged receptive fields Ruo-Chun Tzeng and Shan-Hung Wu Ego-CNN ICML’19 1 / 9

  4. Goal To learn representations of graphs by using convolutions ...while keeping nice properties of CNNs on images: Filters detect location independent patterns Filters at a deep layer have enlarged receptive fields ...and being able to detect critical structures Ruo-Chun Tzeng and Shan-Hung Wu Ego-CNN ICML’19 1 / 9

  5. What are the critical structures? Local-scale critical structures, e.g., Alkane vs Alcohol Global-scale critical structures, e.g, Hydrocarbon Ruo-Chun Tzeng and Shan-Hung Wu Ego-CNN ICML’19 2 / 9

  6. STOA: Graph Attention Networks (GAT) 1 The (1-head) GAT learns an attention score α ij for each edge ( i , j )   h ( l ) α ij Wh ( l − 1)  � = σ + b i j  j ∈N i α ’s explicitly point out the critical structures When jointly learned with a task, α ij denotes the contribution of edge ( i , j ) to the model prediction 1 P Veličković, G Cucurull, A Casanova, A Romero, P Lio, Y Bengio, Graph attention networks, ICLR’18 Ruo-Chun Tzeng and Shan-Hung Wu Ego-CNN ICML’19 3 / 9

  7. Drawback: limited learning ability However, the (1-head) GAT suffers from limited learning ability   α ij Wh ( l − 1) h ( l )  � = σ + b i j  j ∈N i A filter W scans one node at a time does not capture the interactions between nodes Not a serious problem for node-level (e.g., classification) tasks But may severely degrade the performance of graph-level tasks Ruo-Chun Tzeng and Shan-Hung Wu Ego-CNN ICML’19 4 / 9

  8. A new way: Ego-CNN Idea: to learn critical structures just like image-based CNNs   h ( l ) α ij Wh ( l − 1) + b = σ  �  1-head GAT (ICLR’18) i j     j ∈N i W ⊛ � j ∈N i h ( l − 1) h ( l ) � � + b = σ Traditional Convolution i j h ( l ) W ⊛ � j ∈N i h ( l − 1) � � + b = σ Ego-Convolution (ours) i j For each node i , a filter W is applied to all nodes in the 1 neighborhood N i of i Use common visualization techniques (e.g., deconv) to backtrack 2 critical structures Ruo-Chun Tzeng and Shan-Hung Wu Ego-CNN ICML’19 5 / 9

  9. Challenge: variable-sized N i makes W ill-defined h ( l ) � W ⊛ � j ∈N i h ( l − 1) � = σ + b i j For images, N i can be easily defined E.g., K × K pixel block centered at i But how to define N i for graphs ? Ruo-Chun Tzeng and Shan-Hung Wu Ego-CNN ICML’19 6 / 9

  10. Challenge: variable-sized N i makes W ill-defined h ( l ) � W ⊛ � j ∈N i h ( l − 1) � = σ + b i j For images, N i can be easily defined E.g., K × K pixel block centered at i But how to define N i for graphs ? Solution: nodes that are most salient to the given task in a ego-network centered at i First layer: set N (1) as the top K unique nodes in 1 i Weisfeiler-Lehnman labeling Deep layers: N ( l ) = N ( l − 1) (just like image-based CNNs) 2 i i Ruo-Chun Tzeng and Shan-Hung Wu Ego-CNN ICML’19 6 / 9

  11. Improved learning ability on graph classification In Ego-CNN, a W at layer l can detect node interaction patterns within l -hop ego-networks 1st layer 2nd 5th Ruo-Chun Tzeng and Shan-Hung Wu Ego-CNN ICML’19 7 / 9

  12. Improved learning ability on graph classification In Ego-CNN, a W at layer l can detect node interaction patterns within l -hop ego-networks 1st layer 2nd 5th Graph classification benchmark datasets With K = 16, Ego-CNN is comparable to the state-of-the-arts Ruo-Chun Tzeng and Shan-Hung Wu Ego-CNN ICML’19 7 / 9

  13. Ego-CNN can learn critical structures WITHOUT α Backtracking W with common CNN visualization techniques (e.g., deconv) reveals critical structures Local-Scale: Alkane vs Alcohol Global-Scale: Symmetric vs Asymmetric (a) C 14 H 29 OH (c) Symmetric Isomer (b) C 82 H 165 OH (d) Asymmetric Isomer Table: Visualization of the critical structures detected by Ego-CNN Ruo-Chun Tzeng and Shan-Hung Wu Ego-CNN ICML’19 8 / 9

  14. More benefits... and let’s chat at Poster #22 Ego-CNN can detect self-similar patterns I.e., same patterns that exist at different zoom levels Commonly exist in social networks How? Ruo-Chun Tzeng and Shan-Hung Wu Ego-CNN ICML’19 9 / 9

  15. More benefits... and let’s chat at Poster #22 Ego-CNN can detect self-similar patterns I.e., same patterns that exist at different zoom levels Commonly exist in social networks How? By simply tying the weights ( W ’s) across different layers For more details, please go to Poster #22 Ruo-Chun Tzeng and Shan-Hung Wu Ego-CNN ICML’19 9 / 9

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend