Infinite Mixture Prototypes for Few-Shot Learning Adaptively - - PowerPoint PPT Presentation

infinite mixture prototypes for few shot learning
SMART_READER_LITE
LIVE PREVIEW

Infinite Mixture Prototypes for Few-Shot Learning Adaptively - - PowerPoint PPT Presentation

Infinite Mixture Prototypes for Few-Shot Learning Adaptively inferring model capacity for simple and complex tasks Kelsey Allen, Evan Shelhamer*, Hanul Shin*, Josh Tenenbaum Few-Shot Learning by Deep Metric Learning Given few instances of a few


slide-1
SLIDE 1

Infinite Mixture Prototypes for Few-Shot Learning

Kelsey Allen, Evan Shelhamer*, Hanul Shin*, Josh Tenenbaum Adaptively inferring model capacity for simple and complex tasks

slide-2
SLIDE 2

Given few instances of a few classes, recognize a new instance:

Labeled support

Few-Shot Learning by Deep Metric Learning

Query

slide-3
SLIDE 3

Given few instances of a few classes, recognize a new instance:

Labeled support

Few-Shot Learning by Deep Metric Learning

deep net Query

slide-4
SLIDE 4

Given few instances of a few classes, recognize a new instance:

Labeled support embedding embedding

Few-Shot Learning by Deep Metric Learning

deep net Query

slide-5
SLIDE 5

Given few instances of a few classes, recognize a new instance:

Labeled support embedding embedding Unlabeled support

Few-Shot Learning by Deep Metric Learning

deep net Query

slide-6
SLIDE 6

Given few instances of a few classes, recognize a new instance:

Labeled support Unlabeled support

Few-Shot Learning by Deep Metric Learning

deep net Query embedding embedding deep net

slide-7
SLIDE 7

Given few instances of a few classes, recognize a new instance:

Labeled support Unlabeled support

Few-Shot Learning by Deep Metric Learning

deep net Query embedding embedding deep net

slide-8
SLIDE 8
  • Simple tasks might be accurately represented as uni-modal clusters
  • Complex tasks might require a more sophisticated clustering
  • A deeper/wider network may not solve both kinds of task simultaneously

Simple and Complex Tasks

Omniglot character task Omniglot super category task

slide-9
SLIDE 9

Simple and Complex Tasks

Omniglot character embeddings Omniglot super category embeddings

  • Simple tasks might be accurately represented as uni-modal clusters
  • Complex tasks might require a more sophisticated clustering
  • A deeper/wider network may not solve both kinds of task simultaneously
slide-10
SLIDE 10

Infinite Mixture Modeling

  • Represent clustering process using

Dirichlet Process mixture model

  • Unbounded number of clusters in

mixture - let data determine for itself

  • Naturally interpolates between

nearest neighbors (each data point its own cluster) and prototypes (each cluster is uni-modal Gaussian)

  • Semi-supervised and unsupervised

possible

slide-11
SLIDE 11
  • Adapt between simple and complex data distributions

by learning deep representation and inferring the number of clusters

  • Efficient inference based on DP-means

Adaptive Capacity for Simple and Complex Tasks

slide-12
SLIDE 12

Results

  • 25% absolute improvement over prototypical nets (Snell et al. 2017)

for alphabet/super-class recognition on Omniglot

  • 10% absolute improvement for super-class to sub-class transfer
  • n tiered-ImageNet
  • equal or better to fully-supervised and semi-supervised prototypical nets
  • n Omniglot and mini-ImageNet benchmarks
  • 7% absolute improvement over deep nearest neighbors
  • n mini-ImageNet
  • 20% absolute improvement in unsupervised clustering AMI

Poster 87