9.54 Class 16 Features for recognition supervised, unsupervised - - PowerPoint PPT Presentation

9 54 class 16
SMART_READER_LITE
LIVE PREVIEW

9.54 Class 16 Features for recognition supervised, unsupervised - - PowerPoint PPT Presentation

9.54 Class 16 Features for recognition supervised, unsupervised and innate Shimon Ullman + Tomaso Poggio Danny Harari + Daniel Zysman + Darren Seibert Visual recognition The initial input is just image intensities Object Categories -- We


slide-1
SLIDE 1

9.54 Class 16

Features for recognition

supervised, unsupervised and innate

Shimon Ullman + Tomaso Poggio Danny Harari + Daniel Zysman + Darren Seibert

slide-2
SLIDE 2

Visual recognition

slide-3
SLIDE 3

The initial input is just image intensities

slide-4
SLIDE 4
  • - We perceive the world in term of objects and classes
  • - Large variability within a each class

Object Categories

slide-5
SLIDE 5

Individual Recognition

slide-6
SLIDE 6

Object parts

Headlight Window Door knob Back wheel Mirror Front wheel Headlight Window Bumper

slide-7
SLIDE 7

Categorization: dealing with class variability

slide-8
SLIDE 8

Class Non-class

Natural for the brain, difficult computationally

slide-9
SLIDE 9

Unsupervised Classification

slide-10
SLIDE 10

Features and Classifiers

slide-11
SLIDE 11

Features and Classifiers

slide-12
SLIDE 12

Generic Features

Simple (wavelets) Complex (Geons)

Image features Classifier

slide-13
SLIDE 13

Visual Class: Similar Configurations of Shared Image Components

slide-14
SLIDE 14

What will be optimal image building- blocks for the class?

slide-15
SLIDE 15

Optimal Class Components?

  • Large features are too rare
  • Small features are found

everywhere Find features that carry the highest amount of information

slide-16
SLIDE 16

Mutual Information I(C,F) I(F,C) = H(C) – H(C|F)

  

  

   

C c F f F f

f c P Log f c P f p f F C H f p F C H )) ( ( ) ( ) ( ) ( ) ( ) (

  • Definition of MI as the difference between

the class entropy and conditional entropy of the class given a feature:

  • Definition of entropy:

 

C c

c P Log c P C H )) ( ( ) ( ) (

  • Definition of conditional entropy:
slide-17
SLIDE 17

Mutual Information I(C,F)

Class: 1 1 1 1 Feature 1 1 1 1

I(F,C) = H(C) – H(C|F)

slide-18
SLIDE 18

Computing MI from Examples

  • Mutual information can be measured from

examples:

  • 100 Faces

100 Non-faces Feature: 44 times 6 times Mutual information: 0.1525

H(C) = 1, H(C|F) = 0.8475

Simple neural-network approximations

slide-19
SLIDE 19

Optimal classification features

  • Theoretically: maximizing delivered information minimizes classification

error

Error = H – I(C;F)

  • In practice: informative object components can be identified in training images
slide-20
SLIDE 20

Mutual Info vs. Threshold

0.00 20.00 40.00 Detection threshold Mutual Info

forehead hairline mouth eye nose nosebridge long_hairline chin twoeyes

Selecting Fragments

‘Imprinting’ many receptive fields and selecting a subset

slide-21
SLIDE 21

Adding a New Fragment

(Avoiding redundancy by max-min selection) ?

ΔMI Maxi Mink ΔMI (Fi, Fk) Compare new fragments Fi to all the previous ones. Select F which maximizes the additional information

Competition between units with similar responses

slide-22
SLIDE 22

Highly Informative Face Fragments

Optimal receptive fields for Faces

Ullman et al Nature Neuroscience 2002

slide-23
SLIDE 23

Horse-class features Car-class features

Informative class features

slide-24
SLIDE 24

Informative fragments with positions

∑wk Fk > θ

On all detected fragments within their regions

slide-25
SLIDE 25

Star model

Detected fragments ‘vote’ for the center location Find location with maximal vote In variations, a popular state-of-the art scheme

slide-26
SLIDE 26

Image parts informative for classification

Fergus, Perona, Zisserman 2003 Agarwal, Roth 2002

Ullman, Sali 1999

slide-27
SLIDE 27

Variability of Airplanes Detected

slide-28
SLIDE 28

Image representation for recognition HoG Descriptor

Dallal, N & Triggs, B. Histograms of Oriented Gradients for Human Detection

slide-29
SLIDE 29

Object model using HoG

slide-30
SLIDE 30

fMRI

Functional Magnetic Resonance Imaging

slide-31
SLIDE 31

Looking for Class Features in the Brain: fMRI

Lerner, Epshtein Ullman Malach JCON 2008

slide-32
SLIDE 32

Class-fragments and Activation

Malach et al 2008

slide-33
SLIDE 33

EEG

slide-34
SLIDE 34

Informative Fragments: ERP Study

Harel, Ullman, Epshtein, Bentin

slide-35
SLIDE 35

ERP

FACE FEATURES milliseconds 200 400 600 200 400 600 milliseconds Left Hemisphere Right Hemisphere Posterior-Temporal sites FACE FEATURES milliseconds 200 400 600 200 400 600 milliseconds Left Hemisphere Right Hemisphere Posterior-Temporal sites

MI 1 — MI 2 — MI 3 — MI 4 — MI 5 —

Harel, Ullman,Epshtein, Bentin Vis Res 2007

slide-36
SLIDE 36

Features for object segregation:

Innate mechanisms for unsupervised learning

slide-37
SLIDE 37

Object Segregation

Object 1 Object 2 Background

slide-38
SLIDE 38

Object segregation is learned

[Kellman & Spelke 1983; Spelke 1990; Kestenbaum et al., 1987]

5 months

Even basic Gestalt cues are initially missing

[Schmidt et al. 1986]

slide-39
SLIDE 39

Object segregation is learned

Adults

slide-40
SLIDE 40

It all begins with motion

slide-41
SLIDE 41

It all begins with motion

Grouping by common motion precedes figural goodness

[Spelke 1990 - review]

Motion discontinuities provide an early cue for occlusion boundaries

[Granrud et al. 1984]

slide-42
SLIDE 42

Our model

Static segregation Local occlusion boundaries Object form Motion discontinuities Common motion

Boundary General Accurate Noisy Incomplete Global Object-specific Complete Inaccurate

Motion-based segregation

Dorfman, Harari & Ullman, CogSci 2013

slide-43
SLIDE 43

Intensity edges?

Boundary

slide-44
SLIDE 44

Occlusion cues

Extremal edges Convexity T-junctions

[Ghose & Palmer 2010]

Boundary

slide-45
SLIDE 45

Familiar object

Global

slide-46
SLIDE 46

How does it actually work?

slide-47
SLIDE 47

Moving object

Motion

slide-48
SLIDE 48

Motion Boundary Global

Figure Ground Unknown

slide-49
SLIDE 49

Need many examples for good results (1000+) Boundary

Informative boundary features

slide-50
SLIDE 50

Prediction

Figure

  • r

Ground? Figure

  • r

Ground? Novel object, novel background

78% success Using 100,000 training examples

Boundary

slide-51
SLIDE 51

Entire image

Boundary

Figure Background

slide-52
SLIDE 52

Learning an object

Standard object recognition algorithm

Learns local features and their relative locations

Global

slide-53
SLIDE 53

Detection

Global

slide-54
SLIDE 54

Combining information sources

Combined Boundary

Accurate Noisy & Incomplete

Global

Complete Inaccurate

Figure Background

slide-55
SLIDE 55

More complex algorithms

Default GrabCut With segregation cue

[Rother et al. 2004]

slide-56
SLIDE 56

More complex algorithms

Default GrabCut With segregation cue

[Rother et al. 2004]

slide-57
SLIDE 57

Object segregation - summary

  • Static segregation is learned from motion
  • Two simple mechanisms:

Boundary

Motion discontinuities  Occlusion boundaries

(Need a rich library, including extremal edges)

Global

Common motion  Object form

  • These mechanisms work in synergy
  • This is enough to get started,

adult segregation is much more complex

slide-58
SLIDE 58

Summary

  • Features are important for many visual tasks

such as object recognition and segregation.

  • Features can be learned in a supervised

manner given labeled examples.

  • Features can be also learned in an

unsupervised manner using statistical regularities or domain-specific cues.