Local Features and Kernels for Classifcation of Texture and Object - - PowerPoint PPT Presentation

local features and kernels for classifcation of texture
SMART_READER_LITE
LIVE PREVIEW

Local Features and Kernels for Classifcation of Texture and Object - - PowerPoint PPT Presentation

Introduction Our Image Classification Framework Influence of background Summary Local Features and Kernels for Classifcation of Texture and Object Categories: A Comprehensive Study J. Zhang 1 M. Marszaek 1 S. Lazebnik 2 C. Schmid 1 1 INRIA


slide-1
SLIDE 1

Introduction Our Image Classification Framework Influence of background Summary

Local Features and Kernels for Classifcation of Texture and Object Categories: A Comprehensive Study

  • J. Zhang1
  • M. Marszałek1
  • S. Lazebnik2
  • C. Schmid1

1INRIA Rhône-Alpes, LEAR - GRAVIR

Montbonnot, France

2Beckman Institute, University of Illinois

Urbana, USA

Beyond Patches Workshop, 2006

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

slide-2
SLIDE 2

Introduction Our Image Classification Framework Influence of background Summary

Overview

We have built an extensible image classification framework

sparse local features bag-of-features image representation non-linear Support Vector Machines (SVMs) for classification

We have evaluated various elements of the framework on

4 texture datasets (UIUCTex, KTH-TIPS, Brodatz, CUReT) 5 object category datasets (Xerox7, Caltech6, Caltech101, Graz, PASCAL 2005)

The conclusions hold over the datasets We have performed a detailed evaluation of the background influence

to check whether we can exploit context information to evaluate the robustness against background clutter

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

slide-3
SLIDE 3

Introduction Our Image Classification Framework Influence of background Summary

Outline

1

Our Image Classification Framework Framework components Comparison with state-of-the-art

2

Influence of background Context information Non-represenative training set

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

slide-4
SLIDE 4

Introduction Our Image Classification Framework Influence of background Summary Framework components Comparison with state-of-the-art

Outline

1

Our Image Classification Framework Framework components Comparison with state-of-the-art

2

Influence of background Context information Non-represenative training set

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

slide-5
SLIDE 5

Introduction Our Image Classification Framework Influence of background Summary Framework components Comparison with state-of-the-art

Overview

Salient image regions (interest “points”) are detected Regions are locally described with feature vectors Features are quantized or clustered Histograms or signatures are classified with SVMs

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

Image → Interest points → Local descriptors → Bag-of-features → Classification

slide-6
SLIDE 6

Introduction Our Image Classification Framework Influence of background Summary Framework components Comparison with state-of-the-art

Detectors

We have evaluated two widely used detectors

Harris-Laplace — detects corners Laplacian — detects blobs

Laplacian demonstrates slightly higher performance...

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

Image → Interest points → Local descriptors → Bag-of-features → Classification

slide-7
SLIDE 7

Introduction Our Image Classification Framework Influence of background Summary Framework components Comparison with state-of-the-art

Detectors

We have evaluated two widely used detectors

Harris-Laplace — detects corners Laplacian — detects blobs

Laplacian demonstrates slightly higher performance... ...the combination, however, performs even better The two detectors capture complementary information

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

Image → Interest points → Local descriptors → Bag-of-features → Classification

slide-8
SLIDE 8

Introduction Our Image Classification Framework Influence of background Summary Framework components Comparison with state-of-the-art

Descriptors

We have evaluated three descriptors

SIFT — gradient orientation histogram SPIN — rotation invariant histogram of intensities RIFT — rotation invariant version of SIFT

SIFT performs the best, SPIN slightly worse, RIFT seems to loose important information Again, combining SIFT with SPIN improves performance as those descriptors are complementary Adding RIFT does not help

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

Image → Interest points → Local descriptors → Bag-of-features → Classification

slide-9
SLIDE 9

Introduction Our Image Classification Framework Influence of background Summary Framework components Comparison with state-of-the-art

Description invariance

We need invariance to recognize objects observed under varying conditions We have seen that invariance leads to information loss How much invariance do we need?

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

Image → Interest points → Local descriptors → Bag-of-features → Classification

slide-10
SLIDE 10

Introduction Our Image Classification Framework Influence of background Summary Framework components Comparison with state-of-the-art

Description invariance

We need invariance to recognize objects observed under varying conditions We have seen that invariance leads to information loss How much invariance do we need? No more than necessary We needed scale invariance in our experiments Rotation invariance helped only for UIUCTex We have not observed any improvement due to affine adaptation of interest “points” — object recognition is different from matching

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

Image → Interest points → Local descriptors → Bag-of-features → Classification

slide-11
SLIDE 11

Introduction Our Image Classification Framework Influence of background Summary Framework components Comparison with state-of-the-art

Visual vocabulary

Bag-of-words representation has proven its usefullness in text classification Visual words are created by clustering the observed features

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

Image → Interest points → Local descriptors → Bag-of-features → Classification

slide-12
SLIDE 12

Introduction Our Image Classification Framework Influence of background Summary Framework components Comparison with state-of-the-art

Bag-of-Features

Given a vocabulary, we can quantize the feature vector space by assigning each

  • bserved feature to the

closest visual word Given an image, we can create a histogram of words’ occurence Alternatively, we can cluster the set of features Note that there are no common underlying words in this case, the words are adapted to an image Note that both approaches ignore spatial relationships between features

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

Image → Interest points → Local descriptors → Bag-of-features → Classification

slide-13
SLIDE 13

Introduction Our Image Classification Framework Influence of background Summary Framework components Comparison with state-of-the-art

Support Vector Machines

We use non-linear Support Vector Machines to classify histograms and signatures The decision function has the following form g(x) =

  • i αiyiK(xi, x) − b

We use extended Gaussian kernels K(xj, xk) = exp

  • − 1

A D(xj, xk)

  • D(xj, xk) is a similarity measure
  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

Image → Interest points → Local descriptors → Bag-of-features → Classification

slide-14
SLIDE 14

Introduction Our Image Classification Framework Influence of background Summary Framework components Comparison with state-of-the-art

χ2 kernel

To compare histograms, we use χ2 distance D(U, W) = 1 2

m

  • i=1

(ui − wi)2 ui + wi Efficient to compute It is bin-to-bin measure, so common underlying words are necessary

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

Image → Interest points → Local descriptors → Bag-of-features → Classification

slide-15
SLIDE 15

Introduction Our Image Classification Framework Influence of background Summary Framework components Comparison with state-of-the-art

EMD kernel

To compare signatures, we use Earth Mover’s Distance D(U, W) = m

i=1

n

j=1 fij d(ui, wj)

m

i=1

n

j=1 fij

Requires solving a linear programming problem to determine the fij flow We have to define the ground distance d(ui, wj) between features No vocabulary construction is necessary

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

Image → Interest points → Local descriptors → Bag-of-features → Classification

slide-16
SLIDE 16

Introduction Our Image Classification Framework Influence of background Summary Framework components Comparison with state-of-the-art

Which kernel to choose?

Both perform comparably EMD kernel does not require an expensive vocabulary construction — short training times χ2 kernel is faster to compute — short testing times

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

Image → Interest points → Local descriptors → Bag-of-features → Classification

slide-17
SLIDE 17

Introduction Our Image Classification Framework Influence of background Summary Framework components Comparison with state-of-the-art

Outline

1

Our Image Classification Framework Framework components Comparison with state-of-the-art

2

Influence of background Context information Non-represenative training set

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

slide-18
SLIDE 18

Introduction Our Image Classification Framework Influence of background Summary Framework components Comparison with state-of-the-art

Texture datasets

T01 (bark) T02 (bark) T03 (bark) T04 (wood) T05 (wood) T06 (wood) T07 (water) T08 (granite) T09 (marble) T10 (stone) S1 S2 S3 S1 S2 S3 I1 I2 Sponge Cotton Felt Plaster Styrofoam

Methods UIUCTex KTH-TIPS Brodatz CUReT

  • ur

98.3 ± 0.5 95.5 ± 1.3 95.4 ± 0.3 95.3 ± 0.4 Hayman 92.0 ± 1.3 94.8 ± 1.2 95.0 ± 0.8 98.6 ± 0.2 Lazebnik 96.4 ± 0.9 91.3 ± 1.4 89.8 ± 1.0 72.5 ± 0.7 VZ-joint 78.4 ± 2.0 92.4 ± 2.1 92.9 ± 0.8 96.0 ± 0.4

  • G. Gabor

65.2 ± 2.0 90.0 ± 2.0 87.9 ± 1.0 92.4 ± 0.5

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

slide-19
SLIDE 19

Introduction Our Image Classification Framework Influence of background Summary Framework components Comparison with state-of-the-art

Object category datasets

bikes books buildings cars faces phones trees airplanes cars (rear) cars (side) faces motorbikes wildcats bikes people background bikes people background accordion carside pagoda scorpion ibis anchor (93%) (92%) (89%) (15%) (8%) (7%)

Methods Xerox7 CalTech6 Graz Pascal CalTech101 test set1 test set2

  • ur

94.3 97.9 90.0 92.8 74.3 53.9

  • thers

82.0 96.6 83.7 94.6 70.5 43

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

slide-20
SLIDE 20

Introduction Our Image Classification Framework Influence of background Summary Framework components Comparison with state-of-the-art

PASCAL VOC challenge 2005 dataset

bike cars motorbikes people

training test set 1 test set 2

Note the object annotations

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

slide-21
SLIDE 21

Introduction Our Image Classification Framework Influence of background Summary Context information Non-represenative training set

Outline

1

Our Image Classification Framework Framework components Comparison with state-of-the-art

2

Influence of background Context information Non-represenative training set

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

slide-22
SLIDE 22

Introduction Our Image Classification Framework Influence of background Summary Context information Non-represenative training set

Overview

PASCAL VOC challenge 2005 dataset comes with detailed object annotation Using the provided bounding boxes we can approximately separate foreground and background features We perform experiments on five meta-datasets

FF — foreground features BF — background features AF — all features (FF + BF) AF-CONST — foreground features with static scene background features AF-RAND — foreground features with background features of a random

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

slide-23
SLIDE 23

Introduction Our Image Classification Framework Influence of background Summary Context information Non-represenative training set

How important is the context?

0.2 0.4 0.6 0.8 1 0.2 0.4 0.6 0.8 1 True Positive Rate False Positive Rate PASCAL testset 1 BF (bicycles) BF (cars) BF (motorbikes) BF (people) 0.2 0.4 0.6 0.8 1 0.2 0.4 0.6 0.8 1 True Positive Rate False Positive Rate PASCAL testset 2 BF (bicycles) BF (cars) BF (motorbikes) BF (people)

We train and test on BF — still better than random method Background features carry a significant ammount of context information

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

slide-24
SLIDE 24

Introduction Our Image Classification Framework Influence of background Summary Context information Non-represenative training set

Can we use this information?

0.6 0.7 0.8 0.9 1 0.1 0.2 0.3 0.4 True Positive Rate False Positive Rate PASCAL testset 1, people FF AF-CONST AF AF-RAND 0.6 0.7 0.8 0.9 1 0.1 0.2 0.3 0.4 True Positive Rate False Positive Rate PASCAL testset 2, people FF AF-CONST AF AF-RAND

Training and testing on FF gives better results than AF Due to background clutter we cannot use the context information to improve the classification results

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

slide-25
SLIDE 25

Introduction Our Image Classification Framework Influence of background Summary Context information Non-represenative training set

Can we deal with background clutter at all?

0.6 0.7 0.8 0.9 1 0.1 0.2 0.3 0.4 True Positive Rate False Positive Rate PASCAL testset 1, people FF AF-CONST AF AF-RAND 0.6 0.7 0.8 0.9 1 0.1 0.2 0.3 0.4 True Positive Rate False Positive Rate PASCAL testset 2, people FF AF-CONST AF AF-RAND

Training and testing on AF-CONST is still better than AF and often close to FF We can easily deal with static background

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

slide-26
SLIDE 26

Introduction Our Image Classification Framework Influence of background Summary Context information Non-represenative training set

Outline

1

Our Image Classification Framework Framework components Comparison with state-of-the-art

2

Influence of background Context information Non-represenative training set

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

slide-27
SLIDE 27

Introduction Our Image Classification Framework Influence of background Summary Context information Non-represenative training set

It’s easy to get biased

What happens if the test images are not represented well by the training set?

0.6 0.7 0.8 0.9 1 0.1 0.2 0.3 0.4 True Positive Rate False Positive Rate PASCAL testset 1, people FF / AF AF-CONST / AF AF / AF AF-RAND / AF 0.6 0.7 0.8 0.9 1 0.1 0.2 0.3 0.4 True Positive Rate False Positive Rate PASCAL testset 2, people FF / AF AF-CONST / AF AF / AF AF-RAND / AF

Training on AF and AF-RAND is significantly better than on FF or AF-CONST One should not train on too easy examples, it is better to choose too hard ones

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

slide-28
SLIDE 28

Introduction Our Image Classification Framework Influence of background Summary

Summary

We have created an effective image classification framework that outperforms the state-of-the-art We have evaluated the parameters of the framework on a wide range of datesets and were able to deliver general design conclusions We have evaluated the influence that the background has

  • n bag-of-features methods
  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

slide-29
SLIDE 29

Introduction Our Image Classification Framework Influence of background Summary

Future work

Spatial Weigting for

  • bject recognition

Spatial Pyramid Matching for scene classification

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification

slide-30
SLIDE 30

Introduction Our Image Classification Framework Influence of background Summary

Thank you for your attention

I will be glad to answer your questions

  • J. Zhang, M. Marszałek, S. Lazebnik, C. Schmid

Local Features and Kernels for Image Classification