Exploiting Multimodal Data for Image Understanding Matthieu - - PowerPoint PPT Presentation

exploiting multimodal data for image understanding
SMART_READER_LITE
LIVE PREVIEW

Exploiting Multimodal Data for Image Understanding Matthieu - - PowerPoint PPT Presentation

Exploiting Multimodal Data for Image Understanding Matthieu Guillaumin Supervised by Cordelia Schmid and Jakob Verbeek 27/09/2010 Multimodal data Webpages with images, videos, ... Videos with sound, scripts and subtitles, ... Matthieu


slide-1
SLIDE 1

Exploiting Multimodal Data for Image Understanding

Matthieu Guillaumin

Supervised by Cordelia Schmid and Jakob Verbeek

27/09/2010

slide-2
SLIDE 2

Multimodal data

Webpages with images, videos, ... Videos with sound, scripts and subtitles, ...

Matthieu Guillaumin, PhD defense 2/55

slide-3
SLIDE 3

Images with user tags

Leverage user tags available on

  • r other sources:

Tags wow San Fransisco Golden Gate Bridge SBP2005 top-f50 fog SF Chronicle 96 hours

Matthieu Guillaumin, PhD defense 3/55

slide-4
SLIDE 4

News images with captions

Exploit to identify persons, retrieve images, ...

An Iranian reads the last issue of the Farsi-language Nowruz in Tehran, Iran Wednesday, July 24, 2002. An appeals court on Wednesday confirmed the sen- tence banning Iran’s leading reformist daily Nowruz from publishing for six months and its publisher, Mohsen Mirdamadi, who is President Mohammad Khatami’s ally, from reporting for four years. Mir- damadi is head of the National Security and Foreign Policy Committee of the Iranian parliament. (AP Photo/Hasan Sarbakhshian) Chanda Rubin of the United States returns a shot during her match against Elena Dementieva of Russia at the Hong Kong Ladies Challenge January 1, 2003. Rubin beat Dementieva 6-4 6-1. (REUTERS/Bobby Yip) Matthieu Guillaumin, PhD defense 4/55

slide-5
SLIDE 5

Use of multimodal data

As additional features for classification, As labels for training (weak supervision), Or to build large collections of images automatically.

Matthieu Guillaumin, PhD defense 5/55

slide-6
SLIDE 6

Outline

1

Introduction

2

Face verification Logistic discriminant metric learning Experiments

3

News images with captions Graph-based approach for face naming Multiple-instance metric learning

4

Images with user tags Nearest neighbor image auto-annotation Experiments

5

Multimodal classification

6

Conclusion

Matthieu Guillaumin, PhD defense 6/55

slide-7
SLIDE 7

Outline

1

Introduction

2

Face verification Logistic discriminant metric learning Experiments

3

News images with captions Graph-based approach for face naming Multiple-instance metric learning

4

Images with user tags Nearest neighbor image auto-annotation Experiments

5

Multimodal classification

6

Conclusion

Matthieu Guillaumin, PhD defense 7/55

slide-8
SLIDE 8

Visual verification

Decide whether two faces images depict the same individual.

Matthieu Guillaumin, PhD defense 8/55

slide-9
SLIDE 9

Visual verification

Decide whether two faces images depict the same individual.

Matthieu Guillaumin, PhD defense 9/55

slide-10
SLIDE 10

Related work

On face recognition: Eigenfaces [Turk and Pentland, 1991] Fisherfaces [Belhummeur et al., 1997] On visual verification: Patch sampling + Forest + SVM [Nowak and Jurie, 2007] One-shot similarities [Wolf et al., 2008] Many low-level kernels + MKL [Pinto et al., 2009] “Is that you? Metric learning approaches for face identification” [Guillaumin, Verbeek and Schmid, ICCV 2009]

Matthieu Guillaumin, PhD defense 10/55

slide-11
SLIDE 11

Mahalanobis metric learning

Make positive pairs closer than negative pairs

A B C

C

  • m

p r e s s E x p a n d

Mahalanobis metrics dM(xi, xj) = (xi − xj)⊤M(xi − xj), where M is positive semidefinite (PSD). LMNN [Weinberger et al., 2005], ITML [Davis et al., 2007], MCML [Globerson and Roweis, 2005], ...

Matthieu Guillaumin, PhD defense 11/55

slide-12
SLIDE 12

Logistic discriminant metric learning (LDML)

Model the probability of (xi, xj) to have the same label as: pij = p(yi = yj|xi, xj; M, b) = σ(b − dM(xi, xj)) where σ(z) = 1/(1 + exp(−z)). 5 10 15 0.5 1 b d p = σ(b − d)

Matthieu Guillaumin, PhD defense 12/55

slide-13
SLIDE 13

Logistic discriminant metric learning (LDML)

Find M and b to maximize the likelihood on training data: L(M, b) =

  • (i,j)

p[yi=yj]

ij

(1 − pij)[yi=yj] Convex and smooth objective and convex PSD constraint:

Very effective optimization methods.

Kernelizable:

Can handle very high dimensional data.

Low-rank regularization:

Reduces the number of parameters (linear), Defines a PSD matrix, Supervised dimensionality reduction, But: objective becomes non-convex.

Desktop machine: ∼ 104 instances of 3500d in an hour.

Matthieu Guillaumin, PhD defense 13/55

slide-14
SLIDE 14

Outline

1

Introduction

2

Face verification Logistic discriminant metric learning Experiments

3

News images with captions Graph-based approach for face naming Multiple-instance metric learning

4

Images with user tags Nearest neighbor image auto-annotation Experiments

5

Multimodal classification

6

Conclusion

Matthieu Guillaumin, PhD defense 14/55

slide-15
SLIDE 15

Data set of uncontrolled face images

Labeled Faces in the Wild data set, 13233 images, 5749 individuals, standard evaluation protocol. Features: 9 locations × 3 scales × 128d SIFT → 3456d. [Everingham et al., 2006]

Matthieu Guillaumin, PhD defense 15/55

slide-16
SLIDE 16

Comparison to other metric learning

35 55 100 200 500 0.65 0.7 0.75 0.8 0.85 0.9 Projection dimensionality Accuracy L2 Eigenfaces PCA-LMNN [Weinberger] PCA-ITML [Davis] PCA-LDML [ours] LDML low rank [ours]

Matthieu Guillaumin, PhD defense 16/55

slide-17
SLIDE 17

Comparison to the state of the art

Method Setting Accuracy Eigenfaces restricted 0.600 ± 0.8 [Nowak, 2007] restricted 0.739 ± 0.5 [Wolf, 2008] restricted 0.785 ± 0.5 [Pinto, 2009] restricted 0.794 ± 0.6 LDML [ours] restricted 0.793 ± 0.6 [Kumar, 2009] restricted∗ 0.853 ± 1.2 [Wolf, 2008] unrestricted 0.793 ± 0.3 LDML [ours] unrestricted 0.838 ± 0.6 LDML+MkNN [ours] unrestricted 0.875 ± 0.4 Combined multishot [Wolf, 2009] aligned 0.895 ± 0.5

∗ relies on additional training data.

Matthieu Guillaumin, PhD defense 17/55

slide-18
SLIDE 18

Outline

1

Introduction

2

Face verification Logistic discriminant metric learning Experiments

3

News images with captions Graph-based approach for face naming Multiple-instance metric learning

4

Images with user tags Nearest neighbor image auto-annotation Experiments

5

Multimodal classification

6

Conclusion

Matthieu Guillaumin, PhD defense 18/55

slide-19
SLIDE 19

Face naming from news images

The goal is to recover the names of the faces:

German Chancellor Angela Merkel shakes hands with Chinese President Hu Jintao (. . . ) Kate Hudson and Naomi Watts, Le Divorce, Venice Film Festival - 8/31/2003.

Images as sets of faces (using face detector [Viola and Jones, 2004]), Captions as sets of labels (using NLP [Deschacht and Moens, 2006]).

Matthieu Guillaumin, PhD defense 19/55

slide-20
SLIDE 20

Face naming from news images

The goal is to recover the names of the faces:

Angela Merkel Hu Jintao German Chancellor Angela Merkel shakes hands with Chinese President Hu Jintao (. . . ) Kate Hudson Naomi Watts Kate Hudson and Naomi Watts, Le Divorce, Venice Film Festival - 8/31/2003.

Images as sets of faces (using face detector [Viola and Jones, 2004]), Captions as sets of labels (using NLP [Deschacht and Moens, 2006]).

Matthieu Guillaumin, PhD defense 20/55

slide-21
SLIDE 21

Related work

On associating names and faces (videos): Name-It system [Satoh et al., 1999] Video Google Faces and automatic naming in videos [Everingham, Sivic and Zissermann, 2006–2009] For still images: Gaussian mixture model (GMM) [Berg et al., 2004–2007] Multimodal clustering [Pham et al., 2008–2010] Identities and actions [Luo et al., 2009] Graph-based method for retrieval [Ozkan and Duygulu, 2006–2010] “Automatic face naming using caption-based supervision” [Guillaumin, Mensink, Verbeek and Schmid, CVPR 2008]

Matthieu Guillaumin, PhD defense 21/55

slide-22
SLIDE 22

Outline

1

Introduction

2

Face verification Logistic discriminant metric learning Experiments

3

News images with captions Graph-based approach for face naming Multiple-instance metric learning

4

Images with user tags Nearest neighbor image auto-annotation Experiments

5

Multimodal classification

6

Conclusion

Matthieu Guillaumin, PhD defense 22/55

slide-23
SLIDE 23

Graph-based approach

Build a similarity graph:

One vertex fi per face image, Edges are weighted with a similarity wij, One sub-graph Yn for each name n.

Find the sub-graphs Yn that maximize the sum of inner similarities: max

{Yn}

  • n
  • fi∈Yn
  • fj∈Yn

wij

Matthieu Guillaumin, PhD defense 23/55

slide-24
SLIDE 24

Optimization

As such, the global problem is intractable: Generally, the following holds:

1

Faces can only be assigned to at most one name.

2

Faces can only be assigned to a name detected in the caption.

3

Names can only be assigned to at most one face.

Approximate solution:

At document level, match detected faces with detected names, Can be solved exactly and efficiently, Iteration over documents until convergence.

Y1

f f f

Y2

f f f

Y3

f f f

Y4

f f f

Y5

f f f

f1 f2

Matthieu Guillaumin, PhD defense 24/55

slide-25
SLIDE 25

Data set and features

Labeled Yahoo! News, with around 28.000 documents. Manually annotated. Same features as previous section. Study influence of LDML on both GMM and Graph-based approach.

Matthieu Guillaumin, PhD defense 25/55

slide-26
SLIDE 26

Results

2 4 6 8 10 12 0.5 0.6 0.7 0.8 0.9 1 Number of named faces ( ×103) Precision of naming Graph LDML-Graph GMM [Berg] LDML-GMM

Matthieu Guillaumin, PhD defense 26/55

slide-27
SLIDE 27

Outline

1

Introduction

2

Face verification Logistic discriminant metric learning Experiments

3

News images with captions Graph-based approach for face naming Multiple-instance metric learning

4

Images with user tags Nearest neighbor image auto-annotation Experiments

5

Multimodal classification

6

Conclusion

Matthieu Guillaumin, PhD defense 27/55

slide-28
SLIDE 28

MildML

“Multiple-instance ML from automatically labeled bags of faces” [Guillaumin, Verbeek and Schmid, ECCV 2010] Extends LDML to handle sets of faces with sets of labels:

Define distance between pairs of images D and E: dM(D, E) = min

(i,j)∈D×E dM(xi, xj)

Define positivity and negativity of pairs of images by intersecting their label sets.

Johnny Depp, Orlando Bloom. Gore Verbinski, Jerry Bruckheimer, Johnny Depp, Keira Knightley, Orlando Bloom. Matthieu Guillaumin, PhD defense 28/55

slide-29
SLIDE 29

Results

2 4 6 8 10 12 0.6 0.8 1 Number of named faces ( ×103) Precision of naming Graph-based approach LDML MildML L2

Matthieu Guillaumin, PhD defense 29/55

slide-30
SLIDE 30

Outline

1

Introduction

2

Face verification Logistic discriminant metric learning Experiments

3

News images with captions Graph-based approach for face naming Multiple-instance metric learning

4

Images with user tags Nearest neighbor image auto-annotation Experiments

5

Multimodal classification

6

Conclusion

Matthieu Guillaumin, PhD defense 30/55

slide-31
SLIDE 31

Predicting relevance of keywords for images

... car church cloud ... road sky tree ... ? ? ? Application 1: Image annotation and retrieval

Propose a list of relevant keywords to assist human annotator. Given one or more keywords, propose a list of relevant images.

Application 2: Semantic embedding

An image is represented as a vector of word relevances.

Matthieu Guillaumin, PhD defense 31/55

slide-32
SLIDE 32

Related work

Parametric topic models: Extension of PLSA or LDA [Barnard et al., 2003] Non-parametric topic models: Multiple Bernoulli Relevance Model [Feng et al., 2004] Discriminative methods: Multiclass labeling [Carneiro et al., 2007] PAMIR [Grangier and Bengio, 2008] Local approaches: Diffusion of labels on similarity graph [Liu et al., 2009] Nearest neighbor tag transfer [Makadia et al., 2008] “TagProp: Discriminative metric learning in nearest neighbor models for image auto-annotation” [Guillaumin, Mensink, Verbeek and Schmid, ICCV 2009]

Matthieu Guillaumin, PhD defense 32/55

slide-33
SLIDE 33

TagProp: Nearest neighbor image annotation

?

Jet Plane Smoke Sky Plane Prop Clouds Jet Plane Sky Jet Plane Clouds Jet Plane Bear Polar Snow Ice Sky Jet Plane Sky Jet Plane Sky Jet Plane Bear Polar Snow Tundra

Learns the optimal visual distance to use to define neighbors, Effectively sets the number of neighbors to consider.

Matthieu Guillaumin, PhD defense 33/55

slide-34
SLIDE 34

TagProp: Nearest neighbor image annotation

Jet Plane Smoke Sky Plane Prop Clouds Jet Plane Sky Jet Plane Clouds Jet Plane Bear Polar Snow Ice Sky Jet Plane Sky Jet Plane Sky Jet Plane Bear Polar Snow Tundra

TagProp predictions: weighted sum over neighbor images,

p(yiw = +1) =

  • j

πijyjw.

Matthieu Guillaumin, PhD defense 34/55

slide-35
SLIDE 35

Rank-based weights

For every image, the k-th neighbor gets fixed weight γk, There are K parameters, where K is the neighborhood size, Effective neighborhood size set automatically.

0.05 0.1 0.15 0.2 0.25 γk 10 20 30 40 k KNN

Matthieu Guillaumin, PhD defense 35/55

slide-36
SLIDE 36

Distance-based weights

Weights πij depend smoothly on dij with exponential decrease:

πij = exp(−λdij)

  • m exp(−λdim)

λ: decay rate, effective neighborhood size. TagProp can optimize dij, e.g. combining n “base” distances:

dij = λ(1)d(1)

ij

+ λ(2)d(2)

ij

+ · · · + λ(n)d(n)

ij

One parameter λ(k) for each base distance d(k). Small number of parameters, shared by all keywords.

Matthieu Guillaumin, PhD defense 36/55

slide-37
SLIDE 37

Optimization

Maximize the log-likelihood of predictions of training data: L =

  • i
  • w

log p(yiw) Rank-based: convex objective with constraints:

∀k, γk ≥ 0

K

  • k=1

γk = 1.

Distance-based: non-convex objective with constraints:

∀k, λ(k) ≥ 0.

Optimized using projected gradient descent. With pre-computed distances and nearest neighbors, training takes only minutes with 20000 images and 300 keywords.

Matthieu Guillaumin, PhD defense 37/55

slide-38
SLIDE 38

Outline

1

Introduction

2

Face verification Logistic discriminant metric learning Experiments

3

News images with captions Graph-based approach for face naming Multiple-instance metric learning

4

Images with user tags Nearest neighbor image auto-annotation Experiments

5

Multimodal classification

6

Conclusion

Matthieu Guillaumin, PhD defense 38/55

slide-39
SLIDE 39

Features and data sets

Set of 15 standard features with base distances: Color histograms / Bag-of-SIFT, Dense sampling / Interest points, Spatial pyramids: global, landscape layout, GIST. Three benchmark data sets: Corel 5k: 5000 images, 260 words, ESP Game: 20000 images, 268 words, IAPR TC-12: 20000 images, 291 words.

Matthieu Guillaumin, PhD defense 39/55

slide-40
SLIDE 40

Features and data sets

arctic iguana den lizard fox marine grass rocks box blue brown cartoon square man white woman glacier landscape mountain lot people meadow tourist water Matthieu Guillaumin, PhD defense 40/55

slide-41
SLIDE 41

TagProp: Learned combinations of distance

G I S T D e n s e S I F T D e n s e H u e H a r r i s S I F T H a r r i s H u e D e n s e S I F T 3 D e n s e H u e 3 H a r r i s S I F T 3 H a r r i s H u e 3 R G B H S V L A B R G B 3 H S V 3 L A B 3 0.2 0.4 Contribution relative to JEC, normalized Corel 5000 ESP Game IAPR TC-12

Distance combination tends to be sparse, Weights differ between data sets.

Matthieu Guillaumin, PhD defense 41/55

slide-42
SLIDE 42

Comparison to state-of-the-art

% 10 20 30 40 50 P P P R R R COREL 5K IAPR TC-12 ESP Game +6 +10 +18 +6 +17 +2

[Feng et al., 2004] [Makadia et al., 2008] TagProp [ours]

Matthieu Guillaumin, PhD defense 42/55

slide-43
SLIDE 43

Multi-word queries

PAMIR [Grangier and Bengio, 2008], 2241 queries using one or several keywords (COREL), Easy (≥ 3 images) vs. difficult (≤ 2).

mAP All Single Multiple Easy Difficult PAMIR 26% 34% 26% 43% 22% TagProp 36% 46% 35% 55% 32%

Mean average precision: +10% globally.

Matthieu Guillaumin, PhD defense 43/55

slide-44
SLIDE 44

Outline

1

Introduction

2

Face verification Logistic discriminant metric learning Experiments

3

News images with captions Graph-based approach for face naming Multiple-instance metric learning

4

Images with user tags Nearest neighbor image auto-annotation Experiments

5

Multimodal classification

6

Conclusion

Matthieu Guillaumin, PhD defense 44/55

slide-45
SLIDE 45

Multimodal classification

Use as additionnal features for classification. Combine visual and textual kernels in SVM.

DOG (+1) not DOG (−1) DOG?

greyhound running athlete sport horse vermont cars racing dog rottweiler pets computer dual monitor

yacht canine pet locomotive black puppy cute dog Matthieu Guillaumin, PhD defense 45/55

slide-46
SLIDE 46

Results on PASCAL VOC 2007

0.2 0.4 0.6 0.8 1

a e r

  • p

l a n e b i c y c l e b i r d b

  • a

t b

  • t

t l e b u s c a r c a t c h a i r c

  • w

d i n i n g t a b l e d

  • g

h

  • r

s e m

  • t
  • r

b i k e p e r s

  • n

p

  • t

t e d p l a n t s h e e p s

  • f

a t r a i n t v m

  • n

i t

  • r

M e a n PASCAL VOC’07 Average Precision tags image image+tags

Tags (0.43) < Image (0.53) < Image+tags (0.67) Winner of PASCAL VOC’07 [Marsza lek et al.]: 0.59.

Matthieu Guillaumin, PhD defense 46/55

slide-47
SLIDE 47

Multimodal semi-supervised learning

Large pool of additional unlabeled images with tags. Tags NOT available at test time: visual categorization.

DOG (+1) Unlabeled

DOG?

greyhound running athlete sport vermont horse dog rottweiler pets canine pet

not DOG (−1)

puppy dog computer dual monitor railroads train locomotive car auto Matthieu Guillaumin, PhD defense 47/55

slide-48
SLIDE 48

Multimodal semi-supervised learning

Handful of methods that can exploit this setting explicitly: Co-training [Blum and Mitchell, 1998] “Multimodal semi-supervised learning for image classification” [Guillaumin, Verbeek and Schmid, CVPR 2010] Our proposed method, in a nutshell:

Learn a combined image+tags SVM on labeled data, Score the unlabeled multimodal data, Regress these scores with a visual scoring function on the entire training data.

Compare with baseline visual SVM on labeled data.

Matthieu Guillaumin, PhD defense 48/55

slide-49
SLIDE 49

Results of semi-supervised learning

40 40 100 100 200 200 20% 25% 30% 35% 40% 45% PASCAL VOC’07 MIR Flickr Mean AP Number of labeled training examples SVM Co-training [Blum and Mitchell, 1998] Proposed method [ours]

Proposed method improve over SVM and co-training. Especially when few training examples are used.

Matthieu Guillaumin, PhD defense 49/55

slide-50
SLIDE 50

Outline

1

Introduction

2

Face verification Logistic discriminant metric learning Experiments

3

News images with captions Graph-based approach for face naming Multiple-instance metric learning

4

Images with user tags Nearest neighbor image auto-annotation Experiments

5

Multimodal classification

6

Conclusion

Matthieu Guillaumin, PhD defense 50/55

slide-51
SLIDE 51

Contributions

New approach for face naming using graph-based method:

“Automatic face naming with caption-based supervision” (CVPR 2008)

New methods for face verification using metric learning (LDML) and nearest neighbor approaches (MkNN):

“Is that you? Metric learning approaches for face identification” (ICCV 2009)

LDML applied to the naming problem:

“Face recognition from caption-based supervision” (Technical Report, 2010)

ML extended to the multiple instance learning framework:

“Multiple instance metric learning from automatically labeled bags of faces” (ECCV 2010)

Matthieu Guillaumin, PhD defense 51/55

slide-52
SLIDE 52

Contributions

New model for image auto-annotation (TagProp):

“TagProp: Discriminative metric learning in nearest neighbor models for image auto-annotation” (ICCV 2009) “Apprentissage de distance pour l’annotation d’images par plus proches voisins” (RFIA 2010)

Obtained excellent results at the ImageCLEF 2009 and 2010 competitions, and on the MIR Flickr data set:

“INRIA-LEARs participation to ImageCLEF 2009” (CLEF workshop 2009) “Image Annotation with TagProp on the MIRFLICKR set” (ACM MIR 2010)

Proposed method to improve visual classification using multimodal training data:

“Multimodal semi-supervised learning for image classification” (CVPR 2010)

Matthieu Guillaumin, PhD defense 52/55

slide-53
SLIDE 53

Conclusion

Learning adapted similarities significantly improves recognition, clustering and classification, To some extent, these similarities can be learned from weak text-based supervision, More generally, multimodal data indeed improves visual recognition and image understanding.

Matthieu Guillaumin, PhD defense 53/55

slide-54
SLIDE 54

Future challenges

Improve textual analysis to “denoise” labels, Explore other multimodal data (e.g., videos), Design methods for web-scale data sets: at constant annotation cost, can weakly supervised learning outperform supervised learning?

Matthieu Guillaumin, PhD defense 54/55

slide-55
SLIDE 55

Exploiting Multimodal Data for Image Understanding

Matthieu Guillaumin

Supervised by Cordelia Schmid and Jakob Verbeek

27/09/2010