boris babenko steve branson serge belongie
play

Boris Babenko, Steve Branson, Serge Belongie University of - PowerPoint PPT Presentation

Boris Babenko, Steve Branson, Serge Belongie University of California, San Diego ICCV 2009, Kyoto, Japan Recognizing multiple categories Recognizing multiple categories Need meaningful similarity metric / feature space Recognizing


  1. Boris Babenko, Steve Branson, Serge Belongie University of California, San Diego ICCV 2009, Kyoto, Japan

  2. • Recognizing multiple categories Recognizing multiple categories – Need meaningful similarity metric / feature space

  3. • Recognizing multiple categories Recognizing multiple categories – Need meaningful similarity metric / feature space • Idea use training data to learn metric • Idea: use training data to learn metric – Goes by many names: • metric learning i l i • cue combination/weighting • kernel combination/learning • kernel combination/learning • feature selection

  4. • Learn a single global similarity metric Learn a single global similarity metric Query Image Similarity Metric Labeled Dataset Monolithic Category 1 C Category 2 Category 3 [ Jones et al. ‘03, 4 Ch Chopra et al. ‘05, t l ‘05 Category Goldberger et al. ‘05, Shakhnarovich et al. ‘05 Torralba et al. ‘08]

  5. • Learn similarity metric for each category (1 ‐ vs ‐ all) Learn similarity metric for each category (1 vs all) Query Image Similarity Metric Labeled Dataset Monolithic Category 1 C Category 2 Category 3 gory cific Categ 4 Spec Category [ Varma et al. ‘07, Frome et al. ‘07, Weinberger et al. ‘08 Nilsback et al. ’08]

  6. • Monolithic : Monolithic : – Less powerful… there is no “perfect” metric – Can generalize to new categories Can generalize to new categories • Per category : – More powerful – Do we really need thousands of metrics? – Have to train for new categories

  7. • Would like to explore space between two Would like to explore space between two extremes • Idea: • Idea: – Group categories together – Learn a few similarity metrics, one for each group L f i il i i f h

  8. • Learn a few good similarity metrics Learn a few good similarity metrics Query Image Similarity Metric Labeled Dataset Monolithic Category 1 C Category 2 MuSL Category 3 gory cific Categ 4 Spec Category

  9. • Need some framework to work with Need some framework to work with… • Boosting has many advantages: – Feature selection F t l ti – Easy implementation – Performs well

  10. • Training data: Training data: Images Category Labels • Generate pairs: – Sample negative pairs Sa p e ega e pa s ( ( , ), 1 ) 1 ( ( , ), 0 ) 0

  11. • Train similarity metric/classifier: Train similarity metric/classifier:

  12. • Choose Choose to be binary to be binary ‐‐ i e i.e. • = L1 distance over binary vectors – efficient to compute (XOR and sum) ffi i t t t (XOR d ) • For convenience: [Shakhnarovich et al. ’05, Fergus et al. ‘08]

  13. • Given some objective function Given some objective function • Boosting = gradient ascent in function space • Gradient = example weights for boosting Gradient = example weights for boosting chosen weak current strong classifier classifier other weak classifiers function space [Friedman ’01, Mason et al. ‘00]

  14. • Goal : train Goal : train and recover and recover mapping • At runtime • At runtime – To compute similarity of query image to use use y 2 Category 1 ategory 3 Category Category 4 C

  15. • Run pre ‐ processing to group categories (i e k ‐ Run pre processing to group categories (i.e. k means), then train as usual • Drawbacks: • Drawbacks: – Hacky / not elegant – Not optimal: pre ‐ processing not informed by class N i l i i f d b l confusions, etc. • How can we train & group simultaneously ?

  16. • Definitions: Definitions: Sigmoid Function Parameter

  17. • Definitions: Definitions:

  18. • Definitions: Definitions: How well works with category

  19. • Objective function: Objective function: • Each category “assigned” to classifier

  20. • Replace max with differentiable approx Replace max with differentiable approx. where where is a scalar parameter is a scalar parameter

  21. • Each training pair has Each training pair has weights weights

  22. • Intuition: Intuition: Approximation of Difficulty of pair (like regular boosting)

  23. Difficult Pair Easy Pair Assigned to Assigned to -4 -5 x 10 x 10 w 1 i i 6 6 2 2 w 2 5 i 1.5 w 3 4 i 1 3 2 0.5 1 10 20 30 10 20 30 (b (boosting iteration) ti it ti ) (boosting iteration) (b ti it ti )

  24. for for for ‐ Compute weights ‐ Train on weighted pairs end end A i Assign

  25. • Created dataset with many heterogeneous Created dataset with many heterogeneous categories 0 8 0.8 0.75 cy Accurac MuSL+retrain MuSL 0.7 k-means Rand Rand Merged categories from: Monolithic • Caltech 101 [Griffin et al.] Per Cat 0.65 • Oxford Flowers [Nilsback et al.] 5 10 15 20 • UIUC Textures [Lazebnik et al.] UIUC T t [L b ik t l ] K (n mber of classifiers) K (number of classifiers)

  26. MuSL M k ‐ mea k ns

  27. Training more metrics overfits!

  28. • Studied categorization performance vs Studied categorization performance vs number of learned metrics • Presented boosting algorithm to • Presented boosting algorithm to simultaneously group categories and train metrics metrics • Observed overfitting behavior for novel categories i

  29. • Supported by Supported by – NSF CAREER Grant #0448615 – NSF IGERT Grant DGE ‐ 0333451 NSF IGERT Grant DGE 0333451 – ONR MURI Grant #N00014 ‐ 08 ‐ 1 ‐ 0638 – UCSD FWGrid Project (NSF Infrastructure Grant UCSD FWG id P j t (NSF I f t t G t no. EIA ‐ 0303622)

  30. • Train similarity metric/classifier: Train similarity metric/classifier: • Let Let then then

  31. Matrix: 1 1 1 2 2 2 3 3 3 2 4 6 8 10 12 14 16 18 20 2 4 6 8 10 12 14 16 18 20 2 4 6 8 10 12 14 16 18 20 Caltech Caltech Caltech Flowers Textures Flowers Textures Flowers Textures ‐ High value ‐ Low value

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend