Learning Context-dependent Label Permutations for Multi-label - - PowerPoint PPT Presentation

learning context dependent label permutations for multi
SMART_READER_LITE
LIVE PREVIEW

Learning Context-dependent Label Permutations for Multi-label - - PowerPoint PPT Presentation

Learning Context-dependent Label Permutations for Multi-label Classification Jinseok Nam Amazon Alexa AI Joint work with Young-Bum Kim, Eneldo Loza Menca, Sunghyun Park, Ruhi Sarikaya and Johannes Frnkranz Mu Multi-lab label el Clas


slide-1
SLIDE 1

Learning Context-dependent Label Permutations for Multi-label Classification

Jinseok Nam

Amazon Alexa AI Joint work with Young-Bum Kim, Eneldo Loza Mencía, Sunghyun Park, Ruhi Sarikaya and Johannes Fürnkranz

slide-2
SLIDE 2

Mu Multi-lab label el Clas lassif ific icatio tion (MLC)

  • Goal: learn a function f that maps instances to a subset of labels
  • It is important to take into account label dependencies.
  • Joint probability of labels

f − − − − − − → Sea Desert Building Sky Cloud Mountain

P(y1, y2, · · · , yL|x) =

L

Y

i=1

P(yi|y<i, x)

slide-3
SLIDE 3

Ma Maximi mization

  • n of
  • f t

the j joi

  • int p

prob

  • bability
  • Traditional approaches for minimizing subset 0/1 loss:
  • (Probabilistic) classifier chain

Y = {Sea, Desert, Building, Sky, Cloud, Mountain}

  • 1. Creates a chain of L labels

Desert Sea Cloud Mountain Sky Building

Desert = 0 Desert = 0 Sea = 1 Desert = 0 Sea = 1 Cloud = 0 Desert = 0 Sea = 1 Cloud = 0 Mountain = 1 Desert = 0 Sea = 1 Cloud = 0 Mountain = 1 Sky = 1

  • 2. Train L independent classifiers

given input and partial label vector f1

Additional input features

f2 f3 f4 f5 f6 (Dembczyński et al., ICML 2010; Read et al., MLJ 2011)

slide-4
SLIDE 4

Ma Maximi mization

  • n of
  • f t

the j joi

  • int p

prob

  • bability
  • Traditional approaches for minimizing subset 0/1 loss:
  • (Probabilistic) classifier chain

Y = {Sea, Desert, Building, Sky, Cloud, Mountain}

  • 1. Creates a chain of L labels

Desert Sea Cloud Mountain Sky Building

Desert = 0 Desert = 0 Sea = 1 Desert = 0 Sea = 1 Cloud = 0 Desert = 0 Sea = 1 Cloud = 0 Mountain = 1 Desert = 0 Sea = 1 Cloud = 0 Mountain = 1 Sky = 1

  • 2. Train L independent classifiers

given input and partial label vector

  • Error-propagation at test time
  • Effect of label orders in the chain

f1

Additional input features

f2 f3 f4 f5 f6 (Dembczyński et al., ICML 2010; Read et al., MLJ 2011)

Limitations

slide-5
SLIDE 5

Ma Maximi mization

  • n of
  • f t

the j joi

  • int p

prob

  • bability
  • Traditional approaches for minimizing subset 0/1 loss:
  • (Probabilistic) classifier chain

Y = {Sea, Desert, Building, Sky, Cloud, Mountain}

  • 1. Creates a chain of L labels

Desert Sea Cloud Mountain Sky Building

Desert = 0 Desert = 0 Sea = 1 Desert = 0 Sea = 1 Cloud = 0 Desert = 0 Sea = 1 Cloud = 0 Mountain = 1 Desert = 0 Sea = 1 Cloud = 0 Mountain = 1 Sky = 1

  • 2. Train L independent classifiers

given input and partial label vector

  • Error-propagation at test time
  • Effect of label orders in the chain

f1

Additional input features

f2 f3 f4 f5 f6 (Dembczyński et al., ICML 2010; Read et al., MLJ 2011)

Limitations

Sea = 0 Sea = 0 Sea = 0 Sea = 0

slide-6
SLIDE 6

Re Recurrent Neural Networks for MLC

h2 Sea

  • 2

Building h3 Building

  • 3

Sky h4 Sky

  • 4

Mountain Mountain END h5

  • 5

h1

  • 1

Sea h0

(Nam et al., NIPS 2017)

  • Learning from a set of relevant labels in a sequential manner
  • Number of relevant labels is much smaller than the total number of labels
slide-7
SLIDE 7

Re Recurrent Neural Networks for MLC

  • Learning from a set of relevant labels in a sequential manner
  • Number of relevant labels is much smaller than the total number of labels

h2 Sea

  • 2

Building h3 Building

  • 3

Sky h4 Sky

  • 4

Mountain Mountain END h5

  • 5

h1

  • 1

Sea h0

(Nam et al., NIPS 2017)

  • Question: The effect of label permutation remain!

How to determine the target label permutation?

slide-8
SLIDE 8
  • Static label permutation for all instances
  • Arbitrary label sequence randomly chosen at the beginning
  • Label frequency distribution: freq2rare, rare2freq
  • Label structures (e.g., pairwise label dependencies)

➜ Suboptimal choice; learn from only one permutation

  • Different label permutations for individual instances
  • Choosing randomly every time
  • Learning from all possible label permutations

➜ More robust to the effect of label permutation; Computational complexity

We need MLC algorithms that learn context-dependent label permutations efficiently!

Ta Target label permutations for RNN training

slide-9
SLIDE 9
  • Static label permutation for all instances
  • Arbitrary label sequence randomly chosen at the beginning
  • Label frequency distribution: freq2rare, rare2freq
  • Label structures (e.g., pairwise label dependencies)

➜ Suboptimal choice; learn from only one permutation

  • Different label permutations for individual instances
  • Choosing randomly every time
  • Learning from all possible label permutations

➜ More robust to the effect of label permutation; Computational complexity

We need MLC algorithms that learn context-dependent label permutations efficiently!

Ta Target label permutations for RNN training

slide-10
SLIDE 10
  • Static label permutation for all instances
  • Arbitrary label sequence randomly chosen at the beginning
  • Label frequency distribution: freq2rare, rare2freq
  • Label structures (e.g., pairwise label dependencies)

➜ Suboptimal choice; learn from only one permutation

  • Different label permutations for individual instances
  • Choosing randomly every time
  • Learning from all possible label permutations

➜ More robust to the effect of label permutation; Computational complexity

We need MLC algorithms that learn context-dependent label permutations efficiently!

Ta Target label permutations for RNN training

slide-11
SLIDE 11

Mod Model ba based ed label bel per permut utation

2 1 S B x x 2 x x x x 1 x x 2 1 3 5 4 S B x 2 x 1 x 4 x 3 x 5 x

label sequence sampling computing errors & updating parameters False positive prediction True positive prediction False negative prediction

⑴ ⑵

2 1 4 3 5

true target label set:

1 2 3 4 5

sampled target label permutation:

slide-12
SLIDE 12

Po Policy gr gradi dient

2 1 S B x x 2 x x x x 1 x x 2 1 4 3 5

true target label set:

2

Generated label permutation:

1

Model prediction evaluation

rθJ(θ) = EP τ

θ

"T −1 X

i=0

rθ log Pθ(ai|si)(Ri b(si)) #

<latexit sha1_base64="IgrCFsrhHA8C8hGl8eyMxhFpsME=">ACfHicbVFdaxNBFJ1dv2r8aGwfBiFDZIwq6I9qVQEF8itK0hey63JnMJkNnP5i5Wwjr/or+M9/8Kb6Is0lQ23ph4Mw587HubzSylIY/vD8W7fv3L23c7/34OGjx7v9J3sntqyNkFNR6tKcbRSq0JOSZGWZ5WRmHMtT/n5+04/vZDGqrI4plUlkxwXhcqUQHJU2r+MC+Qa0yampSRsPwUbMIRDiHOkJefNhzZtJn8cXx3Aum0h1jKjGcS2ztNGHYZOR5Fjr96ovOVC/jbDwGmCr6BTdUQgi8Oj4AH3W4YG7VYUpL2B+E4XBfcBNEWDNi2Jmn/ezwvRZ3LgoRGa2dRWFHSoCEltGx7cW1lheIcF3LmYIG5tEmzDq+Fl46ZQ1YatwqCNftvR4O5taucO2eXh72udeT/tFlN2UHSqKqSRZic1FWa6ASuknAXBkpSK8cQGUeyuIJRoU5ObVcyFE1798E5y8HkfhOPr8ZnB0sI1jhz1lz1nAIvaOHbGPbMKmTLCf3jMv8IbeL/+F/8ofbay+t+3Z1fKf/sbepLAlg=</latexit><latexit sha1_base64="IgrCFsrhHA8C8hGl8eyMxhFpsME=">ACfHicbVFdaxNBFJ1dv2r8aGwfBiFDZIwq6I9qVQEF8itK0hey63JnMJkNnP5i5Wwjr/or+M9/8Kb6Is0lQ23ph4Mw587HubzSylIY/vD8W7fv3L23c7/34OGjx7v9J3sntqyNkFNR6tKcbRSq0JOSZGWZ5WRmHMtT/n5+04/vZDGqrI4plUlkxwXhcqUQHJU2r+MC+Qa0yampSRsPwUbMIRDiHOkJefNhzZtJn8cXx3Aum0h1jKjGcS2ztNGHYZOR5Fjr96ovOVC/jbDwGmCr6BTdUQgi8Oj4AH3W4YG7VYUpL2B+E4XBfcBNEWDNi2Jmn/ezwvRZ3LgoRGa2dRWFHSoCEltGx7cW1lheIcF3LmYIG5tEmzDq+Fl46ZQ1YatwqCNftvR4O5taucO2eXh72udeT/tFlN2UHSqKqSRZic1FWa6ASuknAXBkpSK8cQGUeyuIJRoU5ObVcyFE1798E5y8HkfhOPr8ZnB0sI1jhz1lz1nAIvaOHbGPbMKmTLCf3jMv8IbeL/+F/8ofbay+t+3Z1fKf/sbepLAlg=</latexit><latexit sha1_base64="IgrCFsrhHA8C8hGl8eyMxhFpsME=">ACfHicbVFdaxNBFJ1dv2r8aGwfBiFDZIwq6I9qVQEF8itK0hey63JnMJkNnP5i5Wwjr/or+M9/8Kb6Is0lQ23ph4Mw587HubzSylIY/vD8W7fv3L23c7/34OGjx7v9J3sntqyNkFNR6tKcbRSq0JOSZGWZ5WRmHMtT/n5+04/vZDGqrI4plUlkxwXhcqUQHJU2r+MC+Qa0yampSRsPwUbMIRDiHOkJefNhzZtJn8cXx3Aum0h1jKjGcS2ztNGHYZOR5Fjr96ovOVC/jbDwGmCr6BTdUQgi8Oj4AH3W4YG7VYUpL2B+E4XBfcBNEWDNi2Jmn/ezwvRZ3LgoRGa2dRWFHSoCEltGx7cW1lheIcF3LmYIG5tEmzDq+Fl46ZQ1YatwqCNftvR4O5taucO2eXh72udeT/tFlN2UHSqKqSRZic1FWa6ASuknAXBkpSK8cQGUeyuIJRoU5ObVcyFE1798E5y8HkfhOPr8ZnB0sI1jhz1lz1nAIvaOHbGPbMKmTLCf3jMv8IbeL/+F/8ofbay+t+3Z1fKf/sbepLAlg=</latexit><latexit sha1_base64="IgrCFsrhHA8C8hGl8eyMxhFpsME=">ACfHicbVFdaxNBFJ1dv2r8aGwfBiFDZIwq6I9qVQEF8itK0hey63JnMJkNnP5i5Wwjr/or+M9/8Kb6Is0lQ23ph4Mw587HubzSylIY/vD8W7fv3L23c7/34OGjx7v9J3sntqyNkFNR6tKcbRSq0JOSZGWZ5WRmHMtT/n5+04/vZDGqrI4plUlkxwXhcqUQHJU2r+MC+Qa0yampSRsPwUbMIRDiHOkJefNhzZtJn8cXx3Aum0h1jKjGcS2ztNGHYZOR5Fjr96ovOVC/jbDwGmCr6BTdUQgi8Oj4AH3W4YG7VYUpL2B+E4XBfcBNEWDNi2Jmn/ezwvRZ3LgoRGa2dRWFHSoCEltGx7cW1lheIcF3LmYIG5tEmzDq+Fl46ZQ1YatwqCNftvR4O5taucO2eXh72udeT/tFlN2UHSqKqSRZic1FWa6ASuknAXBkpSK8cQGUeyuIJRoU5ObVcyFE1798E5y8HkfhOPr8ZnB0sI1jhz1lz1nAIvaOHbGPbMKmTLCf3jMv8IbeL/+F/8ofbay+t+3Z1fKf/sbepLAlg=</latexit>

Label policy distribution Model parameter updates

slide-13
SLIDE 13

Expe Experiments

  • We combined two approaches! Context-dependent label permutation

learning clearly outperforms static label permutation approaches

0.45 0.60 0.75 nDCG@1 0.30 0.40 0.50 0.60 nDCG@5 100 200 0.30 0.33 0.35 ExDPple F1 100 200 0.08 0.12 0.16 0.20 0DcrR F1

freq2rDre rDre2freq fLxeG-rnG DlwDys-rnG CL3-511 α 0 CL3-511 α 0.9

Methods Example F1 Macro F1 Prec@1 Prec@3 Prec@5 Mediamill SLEEC

  • 87.82

73.45 59.17 FastXML

  • 84.22

67.33 53.04 Parabel

  • 83.91

67.12 52.99 freq2rare 66.63±0.33 39.68±0.69 90.05±0.31 74.20±0.18 58.39±0.29 rare2freq 66.95±0.26 43.33±0.62 53.67±1.31 59.57±0.78 52.49±0.37 fixed-rnd 67.21±0.25 41.85±0.90 73.95±5.20 65.58±2.31 55.55±0.83 always-rnd 66.25±0.25 34.03±0.58 89.08±0.18 73.90±0.24 59.45±0.31 CLP-RNN (α=0) 67.22±0.15 38.75±0.88 89.40±0.42 73.84±0.30 59.29±0.17 CLP-RNN (α=0.6) 67.27±0.30 36.49±0.74 91.27±0.28 75.25±0.32 59.75±0.30 Delicious SLEEC

  • 67.59

61.38 56.56 FastXML

  • 69.61

64.12 59.27 Parabel

  • 67.44

61.83 56.75 freq2rare 31.36±0.17 13.94±0.29 57.21±0.38 54.28±0.31 51.16±0.36 rare2freq 31.60±0.15 18.00±0.31 17.46±0.38 18.49±0.51 20.31±0.72 fixed-rnd 32.74±0.27 16.48±0.31 40.59±1.31 37.21±3.06 35.74±2.60 always-rnd 32.45±0.05 13.00±0.25 66.58±0.90 60.46±0.54 54.95±0.55 CLP-RNN (α=0) 34.43±0.54 17.33±0.17 69.57±0.43 61.57±0.69 55.73±0.56 CLP-RNN (α=0.9) 35.80±0.35 18.00±0.51 70.54±0.77 63.39±0.65 57.72±0.58

Poster #233