Fast Discriminative Visual Codebooks using Randomized Clusering - - PowerPoint PPT Presentation

fast discriminative visual codebooks using randomized
SMART_READER_LITE
LIVE PREVIEW

Fast Discriminative Visual Codebooks using Randomized Clusering - - PowerPoint PPT Presentation

Fast Discriminative Visual Codebooks using Randomized Clusering Forests Frank Moosmann, Bill Triggs, and Frederic Jurie Presented by: Andrew F. Dreher CS 395T - Spring 2007 Contributions 1) Creating visual words using classification


slide-1
SLIDE 1

Fast Discriminative Visual Codebooks using Randomized Clusering Forests

Frank Moosmann, Bill Triggs, and Frederic Jurie Presented by: Andrew F. Dreher CS 395T - Spring 2007

slide-2
SLIDE 2

Contributions

1) Creating visual “words” using classification trees 2) Small ensembles of randomized trees can outperform k-means clustering Using stochasticity to improve accuracy

slide-3
SLIDE 3

Trees as “Words”

slide-4
SLIDE 4

Visual “Words”

1) High dimensional vectors; typically extracted features or clusters of features summarized at a point 2) Clusters forming is usually performed using k-means clustering 3) Used with “bag of words” methods derived from text processing

slide-5
SLIDE 5

Trees as “words”

1) Trees are trained as classifiers 2) Leaves are used as “words” Represent a classified cluster of visual features Provides spacial information and intuition lacking in k-means 3) Classification is a separate stage (using SVM) over the leaves

slide-6
SLIDE 6

Information Gain with Entropy

1) Useful with limited number of values 2) Often prefers “pure” nodes Randomization of thresholds helps create different splits and trees Paper parameters Smin and Tmax

[0, 1] Completely random trees [1, D] Discriminative trees (classic ID3)

slide-7
SLIDE 7

Basic Example of Entropy

slide-8
SLIDE 8

Basic Example of Entropy

1

slide-9
SLIDE 9

Basic Example of Entropy

1 3 2

slide-10
SLIDE 10

Basic Example of Entropy

slide-11
SLIDE 11

Experiments

slide-12
SLIDE 12

General Overview

1) Descriptors - Dataset dependent HSV color (768-D vector) Wavelet (768-D vector) Created from HSV using Haar transform SIFT (128-D vector) 2) Performance Metrics 1) Receiver Operating Characteristic (ROC) 2) Equal Error Rate (EER)

slide-13
SLIDE 13

Haar Wavelet

1) First known wavelet 2) Not continuous or differentiable 3) Described as:

Source: Wikipedia (http://en.wikipedia.org/wiki/Haar_wavelet)

{

1 0 ≤ x ≤ ½

  • 1 ½ ≤ x ≤ 1

1 0 otherwise

f (x) =

slide-14
SLIDE 14

Specific Parameters

1) Descriptors: Color Wavelet 2) Tree parameters: Smin = 0.5; Tmax ≈ 50 3) Dataset: GRAZ-02 Three categories 300 Images from each category ½ for training; ½ for testing

slide-15
SLIDE 15

Spacial Results

Posterior probabilities at a given position to be labeled “bike”

slide-16
SLIDE 16

Spacial Results

Posterior probabilities at a given position to be labeled “bike”

slide-17
SLIDE 17

Category vs. Negative

10 20 30 40 50 60 70 80 90 100 Bikes vs. None Cars vs. None

70.9 76.5 79.8 84.1 79.9 84.4

Unsegmented Segmented Opelt et. al

GRAZ-02 Average EER by Category

slide-18
SLIDE 18

Parameters for ERC-Forest vs. K-Means

1) 20,000 total features (only 67 per image) 2) 1000 spacial bins per tree; 5 trees 3) 8000 sampled patches to create global histogram 4) 20,000 windows per image for k-means

slide-19
SLIDE 19

ERC-Forest vs. K-Means

0.9 0.85 0.8 0.75 0.7 0.65 0.6 100 1,000 10,000 100,000 Number of Features per Image to Create Histogram

ERC-Forest K-Means Unsupervied Forest + MI Binarisation Unsupervied Forest

Bike versus Negative Classification

slide-20
SLIDE 20

Other Results

slide-21
SLIDE 21

Pascal Challenge Dataset

10 20 30 40 50 60 70 80 90 100

96.0 94.0 90.1 95.8

Motorbikes Bicycles People Cars

EER by Category using SIFT Descriptor

slide-22
SLIDE 22

Pascal Horses Dataset

1) Highly variable images 2) SIFT Descriptors 3) 100 Patches per image for training 4) 10,000 patches per image for testing 5) Average EER: 85.3%

slide-23
SLIDE 23

Conclusion

1) Method uses forest of randomized classification trees to create a vocabulary Good classification, reasonable training 2) Uses two (2) stage processing Use forest to obtain descriptive “word” Classify “word” using another method 3) Stochasticity improves accuracy

slide-24
SLIDE 24

Thank You