Learning Visual Distance Function L i Vi l Di t F ti for - - PowerPoint PPT Presentation

learning visual distance function l i vi l di t f ti for
SMART_READER_LITE
LIVE PREVIEW

Learning Visual Distance Function L i Vi l Di t F ti for - - PowerPoint PPT Presentation

Learning Visual Distance Function L i Vi l Di t F ti for Identification from one Example for Identification from one Example. Eric Nowak and Frederic Jurie E i N k d F d i J i Bertin Technologies / CNRS LEAR Group INRIA - France


slide-1
SLIDE 1

L i Vi l Di t F ti Learning Visual Distance Function for Identification from one Example for Identification from one Example.

E i N k d F d i J i Eric Nowak and Frederic Jurie Bertin Technologies / CNRS LEAR Group – INRIA - France

slide-2
SLIDE 2

This is an object you've never seen before … … can you recognize it in the following images?

slide-3
SLIDE 3

This is an object you've never seen before … … can you recognize it in the following images?

Id ifi i f O E l Identification from One Example.

“obviously” different same pose and shape but different object same pose and shape, but different object different pose and light, but same object j

slide-4
SLIDE 4

This is an object you've never seen before … … can you recognize it in the following images? Car A Car A Car A Car A Car B Car B

N P ibl !

Car A Car B Car ANot Possible! Car A Car A Car B Car B Car B Car A Car B

Cl B Class A Class B

slide-5
SLIDE 5

This is an object you've never seen before … … can you recognize it in the following images?

S( ) S( ) , S( ) S( ) , S( ) , ( ) ,

slide-6
SLIDE 6

This is an object you've never seen before … … can you recognize it in the following images? This is an object you've never seen before … can you recognize it in the following images? … can you recognize it in the following images?

S( ) S( ) < S( ) , S( ) , <

Knowledge about categories

Different Same

slide-7
SLIDE 7

O l L i f E l Our goal: Learning from one Example with Equivalence Constraints with Equivalence Constraints.

  • We want to learn a similarity measure on a generic category (e.g.

cars)

  • Given a training set of image pairs labelled «same» or «different»:

equivalence constraints

  • we can predict how similar two never seen images are
  • despite occlusions, clutter and modifications in pose, light, ...
slide-8
SLIDE 8

How to compare images ?

Distance

S( ) ,

(Euclidean)

) S( ) , S( ) ) S( ) ,

Representation Space p p (Histograms, etc.)

N t d t d t i l l Not adapted to visual classes

slide-9
SLIDE 9

How to learn the distance ? How to learn the distance ?

Negative Constaint

S=XtAX

Positive Constaint Constaint Representation Space (Hi t t ) (Histograms, etc.)

Not robust to occlusions, background

slide-10
SLIDE 10

How to be robust to occlusion How to be robust to occlusion, view point changes ? view point changes ?

Robust combination” of local distances: S=f(d1 d2 d ) S=f(d1,d2,…,dn)

slide-11
SLIDE 11

C t ti f Computation of corresponding patches corresponding patches

  • P0 in I0: sampled randomly

P0

(quadratic in size, uniform in position)

P0

  • P1 in I1: the best ZNCC match of

P0 around P0. Search region:

P1

g extension of P0 in all directions.

  • A pair of images is simplified

into the np patch pairs sampled from it. from it.

slide-12
SLIDE 12

From multiple local similarities p to one global similarity

Patch 1

P(d|same)

Patch 2

P(d|same) P(d|different)

Patch 2

P(d|same)

Patch i

P(d|same)

Patch n

d d

P(d|different) P(d|different) P(d|same) P(d|different)

d d d

P(d|different)

Likelihood->Similarity [Ferencz et al Iccv 05] [Ferencz et al. Iccv 05]

slide-13
SLIDE 13

Patch independence: a bad assumption

D1 D2 D3 D4 D2 D4 D8 D5 D6 D7

Space of patch Space of patch pairs differences =>Vector quantization

slide-14
SLIDE 14

V t ti ti f Vector quantization of pair difference pair difference

x=[ 1 0 1 1 0 1 0 1 ]

slide-15
SLIDE 15

Computation of the trees Computation of the trees

Tree creation (EXTRA-Trees [Geurts et al. ML06, Moosman et al.

NIPS06]):

– create a root node with positive and negative patch pairs. – recursively split the nodes until they contain only pos or neg pairs:

  • create ncondtrial random split conditions:

simple parametric tests on pixel intensity, gradient, geometry, etc. simple parametric tests on pixel intensity, gradient, geometry, etc. random <=> parameters drawned randomly

  • select the one with the highest information gain

lit th d i t t b d

  • split the node into two sub-nodes

Very Fast!

slide-16
SLIDE 16

Computation of the trees Computation of the trees

The positive patches of three The positive patches of three different nodes during tree construction construction

(''faces in the news'' dataset)

slide-17
SLIDE 17

F l t t Si il it From clusters to Similarity

Sim ( ) =

The similarity measure is a linear combination f th l t

(

,

)

  • f the cluster

membership

=

  • and we want:

the larger the more similar

[ 1 0 1 1 0 1 0 1 ]

  • We define the weight vector

as the normal of the linear as the normal of the linear SVM hyperplane separating the descriptors of positive and negative learn set image pairs.

slide-18
SLIDE 18

Similarity measure

  • Given 2 images ...
  • Detect corresponding patch

pairs pairs.

  • Affect them to clusters with

extremely randomized trees. extremely randomized trees.

  • The similarity measure is a

linear combination of the cluster membership. x=[ 1 0 1 1 0 0 1 1 ] [ ]

Sim ( ) = Sim (

,

) =

slide-19
SLIDE 19

Conclusions Conclusions

  • Similarity of never seen objects, given a set of similar and

different training object pairs of the same category.

  • Original method consisting in
  • Original method consisting in

– (a) finding similar patches – (b) clustering the set of patch pair differences with an ensemble of extremely randomized trees – (c) combining the cluster memberships of the pairs of local regions to make a global decision about the two images. g g

  • Can learn complex visual concepts.
  • Image polysemy‐>of pairs of “same“ and “different” defines

i l t visual concepts

  • Can automatically selects and combines most appropriate

feature types feature types

  • Future works: recognize similar object categories from a

training set of equivalence constraints.