SoDeep: A Sorting Deep Net to Learn Ranking Loss Surrogates June, - - PowerPoint PPT Presentation

sodeep a sorting deep net to learn ranking loss surrogates
SMART_READER_LITE
LIVE PREVIEW

SoDeep: A Sorting Deep Net to Learn Ranking Loss Surrogates June, - - PowerPoint PPT Presentation

SoDeep: A Sorting Deep Net to Learn Ranking Loss Surrogates June, 2019 Martin Engilberge, Louis Chevallier, Patrick Prez, Matthieu Cord SoDeep: A Sorting Deep Net to Learn Ranking Loss Surrogates June, 2019 Martin Engilberge, Louis


slide-1
SLIDE 1

SoDeep: A Sorting Deep Net to Learn Ranking Loss Surrogates

Martin Engilberge, Louis Chevallier, Patrick Pérez, Matthieu Cord

June, 2019

slide-2
SLIDE 2

SoDeep: A Sorting Deep Net to Learn Ranking Loss Surrogates

Martin Engilberge, Louis Chevallier, Patrick Pérez, Matthieu Cord

June, 2019

slide-3
SLIDE 3

Problem

3

Sorting Deep net to learn ranking loss surrogates

  • Metrics often define machine learning tasks
  • Goal: Use metric directly as loss function
  • Focus on ranking metrics:
  • mean Average Precision (mAP)
  • Spearman correlation
  • Recall@threshold
  • Computation of rank is non-differentiable
slide-4
SLIDE 4

Approach

4

Sorting Deep net to learn ranking loss surrogates

  • Pretrained network computes rank from scores
  • Ranking metrics expressed as a function of the rank

DNN

0.1 0.4 0.3 0.01

scores

2 1 3

ranks

0.2 0.6 0.7 0.04

GT scores

2 1 3

GT ranks Loss On ranks backpropagatethroughthe model

DNN

input rank-based loss expression

A S

slide-5
SLIDE 5

Training a differentiable sorter

5

Sorting Deep net to learn ranking loss surrogates

Using only synthetic data:

  • Uniform distribution over [-1,1]
  • Normal distribution with μ = 0 and σ = 1
  • Evenly spaced numbers in random sub-range of [-1,1]

BI-LSTM affine conv. block affine

Sorter architecture:

  • LSTM based
  • Convolution based

DNN

S

slide-6
SLIDE 6

Spearman correlation loss

6

Sorting Deep net to learn ranking loss surrogates

Spearman correlation as a loss function:

  • Spearman correlation:
slide-7
SLIDE 7

Spearman correlation loss

7

Sorting Deep net to learn ranking loss surrogates

Spearman correlation as a loss function:

  • Spearman correlation:
  • Maximizing spearman correlation:
slide-8
SLIDE 8

Spearman correlation loss

8

Sorting Deep net to learn ranking loss surrogates

Spearman correlation as a loss function:

  • Spearman correlation:
  • Maximizing spearman correlation:
  • Replacing rk

rk with the trained sorter:

slide-9
SLIDE 9

Experiments

9

Sorting Deep net to learn ranking loss surrogates

Object recognition: Evaluated on the Pascal VOC 2007 challenge

Model mAP VGG 16 89.3% SoDeep 94.0%

Memorability prediction:

Model

  • Spear. corr.

Baseline 46.0% MSE loss 48.6% SoDeep 49.4%

slide-10
SLIDE 10

Experiments

10

Sorting Deep net to learn ranking loss surrogates

Cross modal retrieval: Evaluated on MS-CoCo image/caption pairs

Caption retrieval Image retrieval Model R@1 R@5 R@10 Med. r R@1 R@5 R@10 Med. r DSVE-Loc 69.8 91.9 96.6 1 55.9 86.9 94.0 1 GXN 68.5

  • 97.9

1 56.6

  • 94.5

1 SoDeep 71.5 92.8 97.1 1 56.2 87.0 94.3 1

Query: A cat on a sofa

slide-11
SLIDE 11

Conclusion and Perspectives

11

Sorting Deep net to learn ranking loss surrogates

  • Learning an approximation of the rank function
  • Competitive results on real tasks
  • Possiblity to extend to other non-differentiable functions

Thank you for your attention ! SoDeep: A Sorting Deep Net to Learn Ranking Loss Surrogates

Poster #18