POLAR: Attention-based CNN for One-shot Personalized Article - - PowerPoint PPT Presentation

polar attention based cnn for one shot personalized
SMART_READER_LITE
LIVE PREVIEW

POLAR: Attention-based CNN for One-shot Personalized Article - - PowerPoint PPT Presentation

Introduction & Problem Definition Approach Experiments & Analysis POLAR: Attention-based CNN for One-shot Personalized Article Recommendation Zhengxiao Du, Jie Tang, Yuhui Ding Tsinghua University { duzx16, dingyh15 }


slide-1
SLIDE 1

Introduction & Problem Definition Approach Experiments & Analysis

POLAR: Attention-based CNN for One-shot Personalized Article Recommendation

Zhengxiao Du, Jie Tang, Yuhui Ding

Tsinghua University {duzx16, dingyh15}@mails.tsinghua.edu.cn, jietang@tsinghua.edu.cn

September 13, 2018

1 / 20

slide-2
SLIDE 2

Introduction & Problem Definition Approach Experiments & Analysis

Motivation

The publication output is growing every year (data source: DBLP)

1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011 2012 2013 2014 2015 2016 0k 50k 100k 150k 200k 250k 300k Book and Theses Conference and Workshop Papers Editorship Informal Publications Journal Articles Parts in Books or Collections Reference Works

2 / 20

slide-3
SLIDE 3

Introduction & Problem Definition Approach Experiments & Analysis

Related-Article Recommendation

Figure: An example from AMiner.org

3 / 20

slide-4
SLIDE 4

Introduction & Problem Definition Approach Experiments & Analysis

Challenge

How to provide personalized and non-personalized recommendation? How to overcome the sparsity of user feedback? How to utilize representative texts of articles effectively?

4 / 20

slide-5
SLIDE 5

Introduction & Problem Definition Approach Experiments & Analysis

Problem Definition

Definition One-shot Personalized Article Recommendation Problem

5 / 20

slide-6
SLIDE 6

Introduction & Problem Definition Approach Experiments & Analysis

Problem Definition

Definition One-shot Personalized Article Recommendation Problem Input: query article dq candidate set D = {d1, d2, · · · , dN} support set S = {( ˆ di, ˆ yi)}T

i=1 related to user u

5 / 20

slide-7
SLIDE 7

Introduction & Problem Definition Approach Experiments & Analysis

Problem Definition

Definition One-shot Personalized Article Recommendation Problem Input: query article dq candidate set D = {d1, d2, · · · , dN} support set S = {( ˆ di, ˆ yi)}T

i=1 related to user u

Output: a totally ordered set R(dq, S) ⊂ D with |R| = k

5 / 20

slide-8
SLIDE 8

Introduction & Problem Definition Approach Experiments & Analysis

One-shot Learning

Image Classification1 ˆ y =

k

  • i=1

a(ˆ x, xi)yi

1Vinyals et al., Matching Networks for One Shot Learning. 6 / 20

slide-9
SLIDE 9

Introduction & Problem Definition Approach Experiments & Analysis

One-shot Learning

Image Classification1 ˆ y =

k

  • i=1

a(ˆ x, xi)yi Article Recommendation Query article dq Support set{(di, yi)}T

i=1 1 T

T

j=1 c( ˆ

di, dj)yj the matching to the user preference(maybe missing)

1Vinyals et al., Matching Networks for One Shot Learning. 6 / 20

slide-10
SLIDE 10

Introduction & Problem Definition Approach Experiments & Analysis

One-shot Learning

Image Classification1 ˆ y =

k

  • i=1

a(ˆ x, xi)yi Article Recommendation Query article dq Support set{(di, yi)}T

i=1

ˆ si =

1 T

T

j=1 c( ˆ

di, dj)yj the matching to the user preference(maybe missing)

1Vinyals et al., Matching Networks for One Shot Learning. 6 / 20

slide-11
SLIDE 11

Introduction & Problem Definition Approach Experiments & Analysis

One-shot Learning

Image Classification1 ˆ y =

k

  • i=1

a(ˆ x, xi)yi Article Recommendation Query article dq Support set{(di, yi)}T

i=1

ˆ si = c(dq, ˆ di) + 1

T

T

j=1 c( ˆ

di, dj)yj the matching to the query article the matching to the user preference(maybe missing)

1Vinyals et al., Matching Networks for One Shot Learning. 6 / 20

slide-12
SLIDE 12

Introduction & Problem Definition Approach Experiments & Analysis

Architecture

Embedding Layer Candidate k

⃗ wdld

k

⃗ wd2

k

⃗ wd1

· · ·

Query k

⃗ wqlq

k

⃗ wq2

k

⃗ wq1

· · ·

Personalized Score

7 / 20

slide-13
SLIDE 13

Introduction & Problem Definition Approach Experiments & Analysis

Architecture

Embedding Layer Candidate k

⃗ wdld

k

⃗ wd2

k

⃗ wd1

· · ·

Query k

⃗ wqlq

k

⃗ wq2

k

⃗ wq1

· · ·

Matching Matrix Attention Matrix Conv Input Personalized Score

7 / 20

slide-14
SLIDE 14

Introduction & Problem Definition Approach Experiments & Analysis

Architecture

Embedding Layer Candidate k

⃗ wdld

k

⃗ wd2

k

⃗ wd1

· · ·

Query k

⃗ wqlq

k

⃗ wq2

k

⃗ wq1

· · ·

Matching Matrix Attention Matrix Feature Map Conv Input Convolution and Max-Pooling Hidden State Full-Connected Layer Matching Score Personalized Score

7 / 20

slide-15
SLIDE 15

Introduction & Problem Definition Approach Experiments & Analysis

Architecture

Embedding Layer Candidate k

⃗ wdld

k

⃗ wd2

k

⃗ wd1

· · ·

Query k

⃗ wqlq

k

⃗ wq2

k

⃗ wq1

· · ·

Matching Matrix Attention Matrix Feature Map Conv Input Convolution and Max-Pooling Hidden State Full-Connected Layer Support Set

ˆ d1

ˆ y1

ˆ d2

ˆ y2

ˆ dT

ˆ yT

Matching Score Personalized Score One Shot Matching

7 / 20

slide-16
SLIDE 16

Introduction & Problem Definition Approach Experiments & Analysis

Architecture

Embedding Layer Candidate k

⃗ wdld

k

⃗ wd2

k

⃗ wd1

· · ·

Query k

⃗ wqlq

k

⃗ wq2

k

⃗ wq1

· · ·

Matching Matrix Attention Matrix Feature Map Conv Input Convolution and Max-Pooling Hidden State Full-Connected Layer Support Set

ˆ d1

ˆ y1

ˆ d2

ˆ y2

ˆ dT

ˆ yT

Matching Score Personalized Score Final Score One Shot Matching

7 / 20

slide-17
SLIDE 17

Introduction & Problem Definition Approach Experiments & Analysis

Matching Matrix and Attention Matrix

Matching Matrix:(dm, dn) → Rlm×ln the similarity between the words of two articles. M(m,n)

i,j

=

  • wT

mi ·

wnj

  • wmi ·

wnj

8 / 20

slide-18
SLIDE 18

Introduction & Problem Definition Approach Experiments & Analysis

Matching Matrix and Attention Matrix

Matching Matrix:(dm, dn) → Rlm×ln the similarity between the words of two articles. M(m,n)

i,j

=

  • wT

mi ·

wnj

  • wmi ·

wnj Attention Matrix:(dm, dn) → Rlm×ln the importance of the matching signals A(m,n)

i,j

= rmi · rnj

8 / 20

slide-19
SLIDE 19

Introduction & Problem Definition Approach Experiments & Analysis

Local Weight and Global Weight

The word weight rt is the product of its local weight and global weight.

9 / 20

slide-20
SLIDE 20

Introduction & Problem Definition Approach Experiments & Analysis

Local Weight and Global Weight

The word weight rt is the product of its local weight and global weight. Global Weight: The importance of a word in the corpus(shared among different articles) υij = [IDF(tij)]β

9 / 20

slide-21
SLIDE 21

Introduction & Problem Definition Approach Experiments & Analysis

Local Weight and Global Weight

The word weight rt is the product of its local weight and global weight. Global Weight: The importance of a word in the corpus(shared among different articles) υij = [IDF(tij)]β The local weight is a little more complicated. . .

9 / 20

slide-22
SLIDE 22

Introduction & Problem Definition Approach Experiments & Analysis

Local Weight

Local Weight: The importance of a word in the article A neural network is employed to compute the local weight. The feature vector for word tij

  • xij =

wij − wi

1 1 2 2 3 3 4 4 5 5 6 6 7 7

The triangular points denote the vectors of the words in two texts The circular points denote the mean vectors of the texts. The lines with arrows denote the feature vectors 10 / 20

slide-23
SLIDE 23

Introduction & Problem Definition Approach Experiments & Analysis

Local Weight Network

The feature vector xij represents the semantic difference between the article and the term. Let u (L)

ij

be the output of the last linear layer, the output of the local weight network is µij = σ(W (L) · u (L)

ij

+ b(L)) + α α sets a lower bound for local weights.

11 / 20

slide-24
SLIDE 24

Introduction & Problem Definition Approach Experiments & Analysis

CNN & Training

The matching matrix and attention matrix are combined by element-wise multiplication and sent to a CNN.

Matching Matrix Attention Matrix Feature Map Conv Input Convolution and Max-Pooling Hidden State Full-Connected Layer Matching Score

The entire model, including the local weight network, is trained on the target task.

12 / 20

slide-25
SLIDE 25

Introduction & Problem Definition Approach Experiments & Analysis

Dataset

AMiner: papers from ArnetMiner1 Patent: patent documents from USPTO RARD (Related Article Recommendation Dataset2):from Sowiport, a digital library service provider.

1Tang et al. ArnetMiner: Extraction and Mining of Academic Social

  • Networks. In SIGKDD’2008.

2Beel et al. Rard: The related-article recommendation dataset (2017) 13 / 20

slide-26
SLIDE 26

Introduction & Problem Definition Approach Experiments & Analysis

Experimental Results

Table: Results of recommendation without personalization(%). AMiner Patent RARD Method NG@3 NG@5 NG@10 NG@3 NG@5 NG@10 NG@1 NG@3 NG@5 TF-IDF 74.3 81.8 87.5 51.8 56.4 63.4 37.6 39.8 46.3 Doc2Vec 60.0 65.8 79.1 44.6 45.6 53.5 28.4 34.0 40.0 WMD 73.0 76.3 86.2 57.4 58.5 61.9 23.4 38.2 46.8 MV-LSTM 56.2 61.2 76.2 60.2 59.0 65.0 22.2 30.7 39.3 Duet 66.6 74.4 82.6 54.5 57.5 64.6 22.3 31.1 39.8 DRMM 75.0 79.9 87.1 55.0 56.2 64.7 33.1 36.3 40.6 MatchPyramid 73.5 80.0 86.8 56.4 61.4 64.4 29.1 36.2 42.8 POLAR 80.3 85.2 90.1 67.8 69.5 73.6 42.8 46.3 51.5

1For the fairness of comparison, all models don’t involve personalization. 2NG stands for NDCG. 14 / 20

slide-27
SLIDE 27

Introduction & Problem Definition Approach Experiments & Analysis

How One-shot Personalization Can Help

Randomly divide the labeled articles into the support set and the candidate set to recommend. POLAR-OS is the proposed one-shot framework and POLAR-ALL the best model that ignores support sets in the previous part.

Table: Performance for the model with and without personalization. AMiner Patent RARD Method NDCG@1 NDCG@3 NDCG@1 NDCG@3 NDCG@1 NDCG@3 POLAR-OS 79.1 81.9 57.1 69.7 39.4 39.2 POLAR-ALL 76.1 79.2 52.3 66.2 36.5 36.5

15 / 20

slide-28
SLIDE 28

Introduction & Problem Definition Approach Experiments & Analysis

How Local and Global Weights Can Help

When computing the attention matrix, POLAR-LOC only uses local weights while POLAR-GLO only global weights.

Aminer 76 77 78 79 80 81 NDCG@3 Patent 60 62 64 66 68 RARD 43.0 43.5 44.0 44.5 45.0 45.5 46.0 46.5 47.0 POLAR-LOC POLAR-GLO POLAR-ALL

Figure: The performance of different attention matrices

16 / 20

slide-29
SLIDE 29

Introduction & Problem Definition Approach Experiments & Analysis

Case Study: How Local and Global Weights work?

2 4 6 8 2 4 6 8

matching matrix

2 4 6 8 2 4 6 8

global weights

2 4 6 8 2 4 6 8

local weights

2 4 6 8 2 4 6 8

weighted matching matrix

Figure: The visualization result of four matrices used in the matching of a pair of texts. The brighter the pixel is, the larger value it has. T1:novel robust stability criteria (for) stochastic hopfield neural networks (with) time delays. T2:new delay dependent stability criteria (for) neural networks (with) time varying delay

17 / 20

slide-30
SLIDE 30

Introduction & Problem Definition Approach Experiments & Analysis

Case Study: How Local and Global Weights work?

Table: The statistical analysis of the local and global weights

Weight Max Min Mean Std Local 2.00 1.00 1.20 0.15 Global 1.96 1.08 1.86 0.08

17 / 20

slide-31
SLIDE 31

Introduction & Problem Definition Approach Experiments & Analysis

Sensitivity Analysis of Hyperparameters

0.25 0.5 1 2 4 local weight coefficient 86.0 86.5 87.0 87.5 88.0 88.5 89.0 89.5 NDCG@10 1/8 1/4 1/2 1 global weight coefficient 86.0 86.5 87.0 87.5 88.0 88.5 89.0 89.5 NDCG@10

Figure: Performance comparison for POLAR-LOC with different αs and POLAR-GLO with different βs on the AMiner dataset

18 / 20

slide-32
SLIDE 32

Introduction & Problem Definition Approach Experiments & Analysis

Conclusion

We define the problem of one-shot personalized article recommendation. We utilize the framework of one-shot learning to deal with the sparse user feedback and propose an attention-based CNN for text similarity. We conduct experiments, whose results prove the effectiveness

  • f the proposed model.

19 / 20

slide-33
SLIDE 33

Introduction & Problem Definition Approach Experiments & Analysis

Any Questions?

20 / 20