predicting temporal sets with deep neural networks
play

Predicting Temporal Sets with Deep Neural Networks Le Yu, Leilei - PowerPoint PPT Presentation

Proceedings of the 26th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD 20), August 23 27, 2020, Virtual Event, CA, USA. Predicting Temporal Sets with Deep Neural Networks Le Yu, Leilei Sun*, Bowen Du, Chuanren Liu, Hui


  1. Proceedings of the 26th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD ’20), August 23– 27, 2020, Virtual Event, CA, USA. Predicting Temporal Sets with Deep Neural Networks Le Yu, Leilei Sun*, Bowen Du, Chuanren Liu, Hui Xiong, Weifeng Lv 1

  2. CONTENTS Background 01 Formalization 02 03 Motivation 04 Methodology Experiments 05 2

  3. Background 3

  4. Background Baskets in shopping Drugs in healthcare Places in traveling Courses in schools 4

  5. Related Work Existing methods usually follow a two-stage strategy (Yu et al. 2016, Hu and He, 2019): (1) set embedding (2) sequential behaviors learning Information loss The two-stage approach often leads to information loss. 5

  6. A unique perspective Existing methods: Sequential Behaviors + Set Embedding Learning Lead to information loss How would the model perform if we make prediction of temporal sets from a different perspective? This work: Learning the relationship of elements Leverage information in elements relationship as much as possible 6

  7. Formalization ⚫ L et 𝕍 = 𝑣 1 , 𝑣 2 , ⋯ , 𝑣 𝑜 , 𝕎 = {𝑤 1 , 𝑤 2 , ⋯ , 𝑤 𝑛 } denote the set of 𝑜 users and 𝑛 elements, a set 𝑇 ⊂ 𝕎 denotes the collection of elements. 𝑈 that records the 1 , 𝑇 𝑗 2 , ⋯ , 𝑇 𝑗 ⚫ Given : a sequence of sets 𝕋 𝑗 = 𝑇 𝑗 historical behaviors of user 𝑣 𝑗 ∈ 𝕍 ⚫ Goal : predict the next-period set of 𝑣 𝑗 , 𝑈+1 = 𝑔 𝑇 𝑗 መ 1 , 𝑇 𝑗 2 , ⋯ , 𝑇 𝑗 𝑈 , 𝑿 , 𝑇 𝑗 where 𝑿 is the trainable parameter. 7

  8. Our Method: DNNTSP Fuse static and dynamic information by a gated updating mechanism Learn temporal dependencies of elements in the sequence of sets Learn set-level element relationship 8

  9. Element Relationship Learning Learn element relationship in the same set. 1) Weighted Graphs Construction 𝑤 𝑗,1 , 𝑤 𝑗,2 , 𝑤 𝑗,1 , 𝑤 𝑗,3 , 𝑤 𝑗,1 , 𝑤 𝑗,2 , 2 , 𝑤 𝑗,1 , 𝑤 𝑗,2 , 0.67 , 𝑤 𝑗,2 , 𝑤 𝑗,1 , 𝑤 𝑗,2 , 𝑤 𝑗,3 , 𝑤 𝑗,1 , 𝑤 𝑗,3 , 3 , 𝑤 𝑗,1 , 𝑤 𝑗,3 , 1.0 , 𝑤 𝑗,3 , 𝑤 𝑗,1 , (𝑤 𝑗,3 , 𝑤 𝑗,2 ) 𝑤 𝑗,1 , 𝑤 𝑗,4 , 1 , 𝑤 𝑗,1 , 𝑤 𝑗,4 , 0.33 , 𝑤 𝑗,2 , 𝑤 𝑗,1 , 2 , 𝑤 𝑗,2 , 𝑤 𝑗,1 , 0.67 , … … … … … 𝑤 𝑗,1 , 𝑤 𝑗,1 , 1 , 𝑤 𝑗,1 , 𝑤 𝑗,1 , 0.33 , 𝑤 𝑗,1 , 𝑤 𝑗,3 , 𝑤 𝑗,1 , 𝑤 𝑗,4 , 𝑤 𝑗,2 , 𝑤 𝑗,2 , 1 , 𝑤 𝑗,2 , 𝑤 𝑗,2 , 0.33 , 𝑤 𝑗,3 , 𝑤 𝑗,1 , 𝑤 𝑗,3 , 𝑤 𝑗,4 , 𝑤 𝑗,3 , 𝑤 𝑗,3 , 1 , 𝑤 𝑗,3 , 𝑤 𝑗,3 , 0.33 , 𝑤 𝑗,4 , 𝑤 𝑗,1 , (𝑤 𝑗,4 , 𝑤 𝑗,3 ) 𝑤 𝑗,4 , 𝑤 𝑗,4 , 1 𝑤 𝑗,4 , 𝑤 𝑗,4 , 0.33 ( a ) ( b ) ( c ) ( d ) 2) Weighted Convolutions on Dynamic Graphs 𝑢,𝑚+1 = 𝜏 𝒄 𝑚 + 𝑢 𝑘, 𝑙 · 𝑿 𝑚 𝒅 𝑗,𝑙 𝑢,𝑚 𝒅 𝑗,𝑘 ෍ 𝐵 𝑗 𝑢 ∪{𝑘} 𝑙∈𝒪 𝑗,𝑘 trainable convolutional parameters 9

  10. Attention-based Temporal Dependency Learning Learn element temporal dependency among the sequence of sets. T 𝑫 𝑗,𝑘 𝑿 𝑟 𝑫 𝒋,𝑘 𝑿 𝑙 𝒂 𝑗,𝑘 = 𝑡𝑝𝑔𝑢𝑛𝑏𝑦 + 𝑵 𝑗 ⋅ (𝑫 𝑗,𝑘 𝑿 𝑤 ) 𝐺 ′′ 𝑢,𝑢 ′ = ቊ 0, 𝑗𝑔 𝑢 ≤ 𝑢′ 𝑁 𝑗 −∞, 𝑝𝑢ℎ𝑓𝑠𝑥𝑗𝑡𝑓 is a masked matrix. T T · 𝒂 𝑗,𝑘 𝒜 𝑗,𝑘 = 𝒂 𝑗,𝑘 · 𝒙 𝑏𝑕𝑕 a trainable parameter to learn the importance of different timestamps 10

  11. Gated Information Fusing Mine shared patterns and fuse static and dynamic representations of elements. 𝑣𝑞𝑒𝑏𝑢𝑓 = 1 − 𝛾 𝑗,𝐽 𝑘 · 𝛿 𝐽 𝑘 𝑭 𝑗,𝐽 𝑘 · 𝑭 𝑗,𝐽 𝑘 + (𝛾 𝑗,𝐽 𝑘 · 𝛿 𝐽 𝑘 ) · 𝒜 𝑗,𝑘 controls the importance of static an indicator vector and dynamic representations 11

  12. The Learning Process Prediction : 𝑣𝑞𝑒𝑏𝑢𝑓 𝒙 𝑝 + 𝒄 𝑝 𝒛 𝑗 = 𝑡𝑗𝑕𝑛𝑝𝑗𝑒 𝑭 𝑗 ෝ Loss function: 𝑂 1 𝑛 𝑀 = − 1 𝟑 𝑂 ෍ 𝑛 ෍ 𝑧 𝑗,𝑘 log ො 𝑧 𝑗,𝑘 + (1 − 𝑧 𝑗,𝑘 ) log 1 − ො 𝑧 𝑗,𝑘 + 𝝁 𝑿 𝑗 𝑘 12

  13. Datasets and Baselines Four datasets are used for evaluation, i.e. TaFeng, DC, TaoBao and TMS. Statistics of the datasets: Both classical and state-of-the-art methods: TOP, PersonalTOP, ElementTransfer, DREAM and Sets2Sets Three evaluation metrics: Recall, NDCG and PHR 13

  14. Experimental Results Our model outperforms existing methods with a significant margin. 14

  15. Experimental Results Our model is applicable to scenarios with sparse data. • Our model is better than Sets2Sets- when no empirical information is added. • Our method outperforms Sets2Sets when incorporating the component for modelling repeated behaviors. 15

  16. Applications ? ? ? ? ? ? customer patient … … 𝑈 + 1 𝑈 + 1 time time automatic treatment next-basket recommendation ? ? ? ? ? ? tourist student … 𝑈 + 1 time … 𝑈 + 1 time travel-package recommendation courses planning 16

  17. Our Team Website: https://www.brilliantasus.com/ 17

  18. Thanks! Contact Le Yu yule@buaa.edu.cn Reference Le Yu, Leilei Sun*, Bowen Du, Chuanren Liu, Hui Xiong, Weifeng Lv, Predicting Temporal Sets with Deep Neural Networks, KDD 2020 Code and data https://github.com/yule-BUAA/DNNTSP 18

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend