BIL-722 ADVANCED TOPICS IN COMPUTER VISION a da Ba , N10266943 - - PowerPoint PPT Presentation

bil 722 advanced
SMART_READER_LITE
LIVE PREVIEW

BIL-722 ADVANCED TOPICS IN COMPUTER VISION a da Ba , N10266943 - - PowerPoint PPT Presentation

BIL-722 ADVANCED TOPICS IN COMPUTER VISION a da Ba , N10266943 Paper: Searching for objects driven by context Authors: Bogdan Alexe, Nicolas Heess, Yee Whye Teh, Vittorio Ferrari PURPOSE: OBJECT DETECTION Among many problems, all


slide-1
SLIDE 1

BIL-722 ADVANCED TOPICS IN COMPUTER VISION

Çağdaş Baş, N10266943 Paper: Searching for objects driven by context Authors: Bogdan Alexe, Nicolas Heess, Yee Whye Teh, Vittorio Ferrari

slide-2
SLIDE 2

PURPOSE: OBJECT DETECTION

 Among many problems, all the methods exhaustively search the object

with help of the sliding windows approach.

 All the methods evaluates all the possible windows.  This process is very slow and also unnatural.

 Cognitive search shows that humans don’t do that. Instead search

intelligently.

slide-3
SLIDE 3

PROPOSITION: INTELLIGENT SEARCH

 Learn an object’s relative

position to its surroundings.

 An ideal search strategy would

be like this:

1.

W1 is sky, cars occur below sky so look below.

2.

W2 is road, cars occur on the road, look just below the road

3.

There is a car part inside W3, look surrounding patches.

4.

W4 is a car.

Figure Credit: Alexe Bogdan

slide-4
SLIDE 4

OVERVIEW OF THE METHOD

Figure Credit: Alexe Bogdan

slide-5
SLIDE 5

ALGORITHM IN A NUTSHELL

1.

Method randomly picks one window at the beginning.

2.

Search Policy 𝜌𝑇:

1.

Similar position/appearance duo searched in the training set.

2.

Each of these similar patches votes for a new position.

3.

Method accumulates these votes as probability maps and decides where to look next.

3.

Output Policy 𝜌𝑃:

1.

If current window similar enough to a car, search is over.

slide-6
SLIDE 6

ALGORITHM IN DETAIL: FEATURE VECTOR

 A window is represented by these vector:

𝑥𝑚 = 𝑦𝑚, 𝑧𝑚, 𝑡𝑚 , 𝑧𝑢

 Window features 𝑧𝑢 consists of:

 Normalized location and scale of the window  HOG Histogram of the window  Classifier score  Displacement vector:

 Intersection over union with the ground truth box  Normalized Hamming distance to the ground truth box  Absolute difference in the window classifier with the ground truth box

Position Feature vectors Scale

slide-7
SLIDE 7

ALGORITHM IN DETAIL: SEARCH POLICY

 Extract uniformly distributed windows from all the training images,

store features.

 For a test image: 1.

Select a window, find it’s K-NN from training windows.

2.

Map new window and acquire the new probability map.

3.

Choose next window with the highest probability:

slide-8
SLIDE 8

ALGORITHM IN DETAIL: SEARCH POLICY (2)

 Calculate probability map with the new window in test image 𝑥𝑢

 𝑥𝑢: Current window in test image.  𝑥𝑚: Window from training set.

Feature similarity kernel Spatial Smoothing Kernel

slide-9
SLIDE 9

ALGORITHM IN DETAIL: SEARCH POLICY (3)

 Normalize each probability map and integrate all the past maps.  Integrate all maps to form the overall probability map using

exponentially decaying mixture.

Feature similarity kernel Spatial Smoothing Kernel

slide-10
SLIDE 10

ALGORITHM IN DETAIL: OUTPUT POLICY

 After 𝑈 iteration, output a single window which has highest classification

score amongst all: 𝑥𝑝𝑣𝑢 = 𝑏𝑠𝑕max

𝑢

𝑑(𝑥𝑢)

 This is a downside. Method assumes that there is only one instance in

the image.

slide-11
SLIDE 11

ALGORITHM IN DETAIL: LEARNING WEIGHTS

 There is a weight for each class in similarity kernel stage.  This weights defines each patch’s importance for each object class.

slide-12
SLIDE 12

OBJECT CLASSIFIER

 An object classifier is trained for each class.

 For each class, one root HOG filter and several part HOG filters are trained.  Root and part filters summed with weights according to Felzenswab’s work.

 For each class, training split is used for classifier learning.

slide-13
SLIDE 13

EXPERIMENTS

 Experiments conducted on PASCAL

VOC 2010 dataset.

 A highly challenging dataset which contains 20 object classes witch bounding box

annotations.

 Validation set is used for testing.  Mean Average Precision over all classes and detection rate and number

  • f windows evaluated by the detector used as performance measures.
slide-14
SLIDE 14

EXPERIMENTS: QUANTITATIVE

slide-15
SLIDE 15

EXPERIMENTS: QUALITATIVE

slide-16
SLIDE 16

EXPERIMENTS: QUALITATIVE

 Comparison of ? With Felzenszwalb et al. PAMI 2010

slide-17
SLIDE 17

EXPERIMENTS: PERFORMANCE

 Experiments run on a Intel i7 processor powered PC.  It can be seen that compared window count is significantly lower than

the usual deformable part model approach.

 It is said that deformable part model approach takes 92s while

proposed method takes only 2s.

slide-18
SLIDE 18

PROS - CONS

 Pros:

 Fast and logical search  Can be applied with any classifier/feature

 Cons:

 Assumes only one instance exists.  Dataset dependent?

slide-19
SLIDE 19

THANKS