Learning to select for a predefined ranking Aleksei Ustimenko - - PowerPoint PPT Presentation

β–Ά
learning to select for a predefined ranking
SMART_READER_LITE
LIVE PREVIEW

Learning to select for a predefined ranking Aleksei Ustimenko - - PowerPoint PPT Presentation

Learning to select for a predefined ranking Aleksei Ustimenko Alexander Vorobev Gleb Gusev Pavel Serdyukov From ranking to sorting Search engines typically order the items by some relevance score obtained from a ranker before presenting


slide-1
SLIDE 1

Learning to select for a predefined ranking

Aleksei Ustimenko Alexander Vorobev Gleb Gusev Pavel Serdyukov

slide-2
SLIDE 2

From ranking to sorting

  • Search engines typically order the items by some relevance

score obtained from a ranker before presenting the items to the user

  • Yet, online shops and social networks allow the user to

rearrange the items using some dedicated attribute (e.g. price or time)

slide-3
SLIDE 3

From ranking To sorting

slide-4
SLIDE 4

Threshold relevance?

  • It was proven that filtering with a constant threshold for relevance is

suboptimal (in terms of ranking quality metrics like DCG)

  • The optimal algorithm was suggested by (Spirin et. at at SIGIR 2015),

but it has quadratic complexity O (n2), where n – is the list size

  • Such algorithms are infeasible for search engines, we need to predict

if to filter an item by just using item features (locally), not the entire list (globally)

slide-5
SLIDE 5

LSO Problem Formulation

  • We define a selection algorithm as 𝐺 and the result of its

application to a list 𝑀 to be the selected 𝑀#

  • 𝑀# - the same ordered list as 𝑀 , but with some items filtered
  • We formulate the problem of LSO as learning from 𝐸 a selection

algorithm 𝐺 that maximizes the expected ranking quality Q of 𝑀#, where 𝑀 is sampled from some 𝑄:

Fβˆ— = arg max 𝔽/~1𝑅(𝑀#)

slide-6
SLIDE 6

Optimal Selection Predictor

  • First, we suggest to build a model 𝑁 that predicts the binary decision
  • f the infeasible optimal algorithm
  • Then we train a binary classifier 𝑁 on the training examples obtained

from that algorithm 𝑦78, π‘ƒπ‘žπ‘’78

7: />∈@,8AB..D> by minimizing logistic

loss

  • However, the logistic loss of such a classifier is still not directly related

to ranking quality Q, i.e. it is not a listwise learning-to-rank algorithm

slide-7
SLIDE 7

Direct Optimization of the Objective

  • For a document 𝑒 with features vector 𝑦F ∈ ℝH we define

probabilistic filtering rule by: 𝑄(𝐺(𝑒) = 1) = 𝜏(𝑔(𝑦F)) = 1 1 + exp(βˆ’π‘”(𝑦F))

  • Assume that decisions 𝐺(𝑒) for different 𝑒 are independent. Denote

the space of all so-defined stochastic selection algorithms by β„±.

  • We transform 𝑅 to the 𝑅QRSSTU (𝐺, 𝑀) = 𝔽V∼1X 𝑅(𝑀V)
  • And the problem to:

πΊβˆ— = arg max #βˆˆβ„± 𝔽/∼@𝑅QRSSTU(𝐺, 𝑀)

slide-8
SLIDE 8

Policy Gradient Approach

  • For i.i.d. samples of binary decisions π‘ŽB, … , π‘ŽQ ∼ 𝑄# define the estimate

(after applying the log derivative trick): πœ–π‘…QRSSTU(𝐺, 𝑀) πœ–π‘”(𝑦8) β‰ˆ 1 𝑑 m

7AB,Q

(𝑅 𝑀V> βˆ’ 𝑐) βˆ’π‘ž8

V>o

1 βˆ’ π‘ž8

BpV>o

where baseline 𝑐 ≔ 𝑅(𝑀rX

s.t) with 𝑨#,v

w.x = 1{π‘žv > 0.5}

  • And we use this functional gradient directly in the Gradient Boosted Decision

Trees learning algorithm (with implementation)

slide-9
SLIDE 9

Pre-training

After training OSP model, we use it as a starting point for

  • ur approach

Thus, we avoid getting stuck in local maxima

slide-10
SLIDE 10

Step by our poster #2 #228