6.891 Computer Vision and Applications Prof. Trevor. Darrell - - PowerPoint PPT Presentation

6 891
SMART_READER_LITE
LIVE PREVIEW

6.891 Computer Vision and Applications Prof. Trevor. Darrell - - PowerPoint PPT Presentation

6.891 Computer Vision and Applications Prof. Trevor. Darrell Lecture 14: Unsupervised Category Learning Gestalt Principles Segmentation by Clustering K-Means Graph cuts Segmentation by Fitting Hough transform


slide-1
SLIDE 1

1

6.891

Computer Vision and Applications

  • Prof. Trevor. Darrell

Lecture 14:

– Unsupervised Category Learning – Gestalt Principles – Segmentation by Clustering

  • K-Means
  • Graph cuts

– Segmentation by Fitting

  • Hough transform
  • Fitting

Readings: F&P Ch. 14, 15.1-15.2

slide-2
SLIDE 2

2

(Un)Supervised Learning

  • Methods in last two lectures presume:

– Segmentation – Labeling – Alignment

  • What can we do with unsupervised (weakly

supervised) data?

  • Clustering / Generative Model Approach…
slide-3
SLIDE 3

3

Representation

Use a scale invariant, scale sensing feature keypoint detector (like the first steps of Lowe’s SIFT).

From: Rob Fergus http://www.robots.ox

[Slide from Bradsky & Thrun, Stanford]

.ac.uk/%7Efergus/

slide-4
SLIDE 4

4

Features for Category Learning

A direct appearance model is taken around each located key. This is then normalized by it’s detected scale to an 11x11 window. PCA further reduces these features.

From: Rob Fergus http://www.robots.ox

[Slide from Bradsky & Thrun, Stanford]

.ac.uk/%7Efergus/

slide-5
SLIDE 5

5 From: Rob Fergus http://www.robots.ox.ac.uk/%7Efergus/

[Slide from Bradsky & Thrun, Stanford]

slide-6
SLIDE 6

6

Learning

  • Fit with E-M (this example is a 3 part model)
  • We start with the dual problem of what to fit and where to fit it.

From: Rob Fergus http://www.robots.ox.ac.uk/%7Efergus/

Assume that an object instance is the only consistent thing somewhere in a scene. We don’t know where to start, so we use the initial random parameters.

  • 1. (M) We find the best (consistent across

images) assignment given the params.

  • 2. (E) We refit the feature detector
  • params. and repeat until converged.
  • Note that there isn’t much

consistency

  • 3. This repeats until it converges at the

most consistent assignment with maximized parameters across images. [Slide from Bradsky & Thrun, Stanford]

slide-7
SLIDE 7

7

Data

Slide from Li Fei-Fei http://www.vision.caltech.edu/feifeili/Resume.htm

[Slide from Bradsky & Thrun, Stanford]

slide-8
SLIDE 8

8

Learned Model

From: Rob Fergus http://www.robots.ox.ac.uk/%7Efergus/

The shape model. The mean location is indicated by the cross, with the ellipse showing the uncertainty in location. The number by each part is the probability of that part being present.

slide-9
SLIDE 9

9

Recognition

From: Rob Fergus http://www.robots.ox.ac.uk/%7Efergus/

slide-10
SLIDE 10

10

Result: Unsupervised Learning

Slide from Li Fei-Fei http://www.vision.caltech.edu/feifeili/Resume.htm

[Slide from Bradsky & Thrun, Stanford]

slide-11
SLIDE 11

11 From: Rob Fergus http://www.robots.ox.ac.uk/%7Efergus/

slide-12
SLIDE 12

12

Segmentation and Line Fitting

  • Gestalt grouping
  • Background subtraction
  • K-Means
  • Graph cuts
  • Hough transform
  • Iterative fitting

(Next time: Probabilistic segmentation)

slide-13
SLIDE 13

13

Segmentation and Grouping

  • Motivation: vision is often

simple inference, but for segmentation

  • Obtain a compact

representation from an image/motion sequence/set of tokens

  • Should support application
  • Broad theory is absent at

present

  • Grouping (or clustering)

– collect together tokens that “belong together”

  • Fitting

– associate a model with tokens – issues

  • which model?
  • which token goes to which

element?

  • how many elements in the

model?

slide-14
SLIDE 14

14

General ideas

  • Tokens

– whatever we need to group (pixels, points, surface elements, etc., etc.)

  • Top down

segmentation

– tokens belong together because they lie on the same object

  • Bottom up

segmentation

– tokens belong together because they are locally coherent

  • These two are not

mutually exclusive

slide-15
SLIDE 15

15

Why do these tokens belong together?

slide-16
SLIDE 16

16

What is the figure?

slide-17
SLIDE 17

17

Basic ideas of grouping in humans

  • Figure-ground

discrimination

– grouping can be seen in terms of allocating some elements to a figure, some to ground – impoverished theory

  • Gestalt properties

– A series of factors affect whether elements should be grouped together

slide-18
SLIDE 18

18

slide-19
SLIDE 19

19

slide-20
SLIDE 20

20

slide-21
SLIDE 21

21

slide-22
SLIDE 22

22

slide-23
SLIDE 23

23

Occlusion is an important cue in grouping.

slide-24
SLIDE 24

24

Consequence: Groupings by Invisible Completions

* Images from Steve Lehar’s Gestalt papers: http://cns-alumni.bu.edu/pub/slehar/Lehar.html

slide-25
SLIDE 25

25

And the famous…

slide-26
SLIDE 26

26

And the famous invisible dog eating under a tree:

slide-27
SLIDE 27

27

Technique: Background Subtraction

  • If we know what the

background looks like, it is easy to identify “interesting bits”

  • Applications

– Person in an office – Tracking cars on a road – surveillance

  • Approach:

– use a moving average to estimate background image – subtract from current frame – large absolute values are interesting pixels

  • trick: use morphological
  • perations to clean up

pixels

slide-28
SLIDE 28

28

slide-29
SLIDE 29

29

low thresh high thresh EM (later) 80x60

slide-30
SLIDE 30

30

low thresh high thresh EM (later) 160x120

slide-31
SLIDE 31

31

Static Background Modeling Examples

[MIT Media Lab Pfinder / ALIVE System]

slide-32
SLIDE 32

32

Static Background Modeling Examples

[MIT Media Lab Pfinder / ALIVE System]

slide-33
SLIDE 33

33

Static Background Modeling Examples

[MIT Media Lab Pfinder / ALIVE System]

slide-34
SLIDE 34

34

Dynamic Background

BG Pixel distribution is non-stationary:

[MIT AI Lab VSAM]

slide-35
SLIDE 35

35

Mixture of Gaussian BG model

Staufer and Grimson tracker: Fit per-pixel mixture model to observed distrubution.

[MIT AI Lab VSAM]

slide-36
SLIDE 36

36

Background Subtraction Principles

Wallflower: Principles and Practice of Background Maintenance, by Kentaro Toyama, John Krumm, Barry Brumitt, Brian Meyers. P1: P2: P3: P4: P5:

slide-37
SLIDE 37

37

Background Techniques Compared

From the Wallflower Paper

slide-38
SLIDE 38

38

Segmentation as clustering

  • Cluster together (pixels, tokens, etc.) that belong

together…

  • Agglomerative clustering

– attach closest to cluster it is closest to – repeat

  • Divisive clustering

– split cluster along best boundary – repeat

  • Dendrograms

– yield a picture of output as clustering process continues

slide-39
SLIDE 39

39

Clustering Algorithms

slide-40
SLIDE 40

40

slide-41
SLIDE 41

41

K-Means

  • Choose a fixed number of

clusters

  • Choose cluster centers and

point-cluster allocations to minimize error

  • can’t do this by search,

because there are too many possible allocations.

  • Algorithm

– fix cluster centers; allocate points to closest cluster – fix allocation; compute best cluster centers

  • x could be any set of

features for which we can compute a distance (careful about scaling) x j − µi

2 j∈elements of i'th cluster

     

i∈clusters

slide-42
SLIDE 42

42

K-Means

slide-43
SLIDE 43

43

Image Clusters on intensity (K=5) Clusters on color (K=5)

K-means clustering using intensity alone and color alone

slide-44
SLIDE 44

44

Image Clusters on color

K-means using color alone, 11 segments

slide-45
SLIDE 45

45

K-means using color alone, 11 segments.

Color alone

  • ften will not

yeild salient segments!

slide-46
SLIDE 46

46

K-means using colour and position, 20 segments

Still misses goal of perceptually pleasing segmentation! Hard to pick K…

slide-47
SLIDE 47

47

Mean Shift Segmentation

http://www.caip.rutgers.edu/~comanici/MSPAMI/msPamiResults.html

slide-48
SLIDE 48

48

Mean Shift Algorithm

Mean Shift Algorithm

  • 1. Choose a search window size.
  • 2. Choose the initial location of the search window.
  • 3. Compute the mean location (centroid of the data) in the search window.
  • 4. Center the search window at the mean location computed in Step 3.
  • 5. Repeat Steps 3 and 4 until convergence.

The mean shift algorithm seeks the “mode” or point of highest density of a data distribution:

slide-49
SLIDE 49

49

Mean Shift Setmentation Algorithm

  • 1. Convert the image into tokens (via color, gradients, texture measures etc).
  • 2. Choose initial search window locations uniformly in the data.
  • 3. Compute the mean shift window location for each initial position.
  • 4. Merge windows that end up on the same “peak” or mode.
  • 5. The data these merged windows traversed are clustered together.

Mean Shift Segmentation

*Image From: Dorin Comaniciu and Peter Meer, Distribution Free Decomposition of Multivariate Data, Pattern Analysis & Applications (1999)2:22–30

slide-50
SLIDE 50

50

Mean Shift Segmentation Results:

http://www.caip.rutgers.edu/~comanici/MSPAMI/msPamiResults.html

slide-51
SLIDE 51

51

Graph-Theoretic Image Segmentation

Build a weighted graph G=(V,E) from image V:image pixels E: connections between pairs of nearby pixels region same the to belong j & i y that probabilit :

ij

W

slide-52
SLIDE 52

52

Graphs Representations

                1 1 1 1 1 1 1 1 a d b c e Adjacency Matrix

* From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

slide-53
SLIDE 53

53

Weighted Graphs and Their Representations

                ∞ ∞ ∞ ∞ ∞ ∞ 1 7 2 1 6 7 6 4 3 2 4 1 3 1 a e d c b 6 Weight Matrix

* From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

slide-54
SLIDE 54

54

Boundaries of image regions defined by a number of attributes

– Brightness/color – Texture – Motion – Stereoscopic depth – Familiar configuration [Malik]

slide-55
SLIDE 55

55

Measuring Affinity

Intensity aff x, y

( )= exp −

1 2σ i

2

    I x

( )− I y ( )

2

( )

      Distance aff x, y

( )= exp −

1 2σ d

2

    x − y

2

( )

      Color aff x, y

( )= exp −

1 2σ t

2

    c x

( )− c y ( )

2

( )

     

slide-56
SLIDE 56

56

Eigenvectors and affinity clusters

  • Simplest idea: we want a

vector a giving the association between each element and a cluster

  • We want elements within

this cluster to, on the whole, have strong affinity with one another

  • We could maximize
  • But need the constraint
  • This is an eigenvalue

problem - choose the eigenvector of A with largest eigenvalue aTAa aTa = 1

slide-57
SLIDE 57

57

Example eigenvector

points eigenvector matrix

slide-58
SLIDE 58

58

Example eigenvector

points eigenvector matrix

slide-59
SLIDE 59

59

Scale affects affinity

σ=.2 σ=.1 σ=.2 σ=1

slide-60
SLIDE 60

60

Scale affects affinity

σ=.1 σ=.2 σ=1

slide-61
SLIDE 61

61

Some Terminology for Graph Partitioning

  • How do we bipartition a graph:

∅ = ∩ ∈ ∈∑

=

B A with B A,

), , W( B) A, (

v u

v u cut

disjoint y necessaril not A' and A A' A,

) , ( W ) A' A, (

∈ ∈

=

v u

v u assoc

[Malik]

slide-62
SLIDE 62

62

Minimum Cut

A cut of a graph G is the set of edges S such that removal of S from G disconnects G. Minimum cut is the cut of minimum weight, where weight of cut <A,B> is given as

( )

( )

∈ ∈

=

B y A x

y x w B A w

,

, ,

* From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

slide-63
SLIDE 63

63

Minimum Cut and Clustering

* From Khurram Hassan-Shafique CAP5415 Computer Vision 2003

slide-64
SLIDE 64

64

Drawbacks of Minimum Cut

  • Weight of cut is directly proportional to the

number of edges in the cut.

Cuts with lesser weight than the ideal cut Ideal Cut

* Slide from Khurram Hassan-Shafique CAP5415 Computer Vision 2003

slide-65
SLIDE 65

65

Normalized cuts

  • First eigenvector of affinity

matrix captures within cluster similarity, but not across cluster difference

  • Min-cut can find degenerate

clusters

  • Instead, we’d like to maximize

the within cluster similarity compared to the across cluster difference

  • Write graph as V, one cluster as

A and the other as B

  • Maximize

where cut(A,B) is sum of weights that straddle A,B; assoc(A,V) is sum of all edges with one end in A. I.e. construct A, B such that their within cluster similarity is high compared to their association with the rest of the graph

cut(A,B) assoc(A,V) cut(A,B) assoc(B,V) +

slide-66
SLIDE 66

66

Solving the Normalized Cut problem

  • Exact discrete solution to Ncut is NP-complete

even on regular grid,

– [Papadimitriou’97]

  • Drawing on spectral graph theory, good

approximation can be obtained by solving a generalized eigenvalue problem.

[Malik]

slide-67
SLIDE 67

67

Normalized Cut As Generalized Eigenvalue problem

... ) , ( ) , ( ; 1 1 ) 1 ( ) 1 )( ( ) 1 ( 1 1 ) 1 )( ( ) 1 ( ) V B, ( ) B A, ( ) V A, ( B) A, ( B) A, ( = = − − − − + + − + = + =

∑ ∑ >

i x T T T T

i i D i i D k D k x W D x D k x W D x assoc cut assoc cut Ncut

i

  • after simplification, we get

. 1 }, , 1 { with , ) ( ) , ( = − ∈ − = D y b y Dy y y W D y B A Ncut

T i T T

[Malik]

slide-68
SLIDE 68

68

Normalized cuts

  • Instead, solve the generalized eigenvalue problem
  • which gives
  • Now look for a quantization threshold that maximizes the criterion ---

i.e all components of y above that threshold go to one, all below go to - b

maxy yT D − W

( )y

( ) subject to yTDy = 1 ( )

D − W

( )y = λDy

slide-69
SLIDE 69

69

Brightness Image Segmentation

slide-70
SLIDE 70

70

Brightness Image Segmentation

slide-71
SLIDE 71

71

slide-72
SLIDE 72

72

Results on color segmentation

slide-73
SLIDE 73

73

Motion Segmentation with Normalized Cuts

  • Networks of spatial-temporal connections:
  • Motion “proto-volume” in space-time
slide-74
SLIDE 74

74

slide-75
SLIDE 75

75

Comparison of Methods

Normalizes A. Finds k eigenvectors, forms X. Normalizes X, clusters rows Affinity A, User inputs k Ng, Jordan, Weiss Finds k eigenvectors of A, forms

  • V. Normalizes rows of V. Forms

Q = VV’. Segments by Q. Q(i,j)=1 -> same cluster Affinity A, User inputs k Scott/ Longuet-Higgins 2nd smallest generalized eigenvector Also recursive D-A with D a degree matrix Shi/Malik 1st x: Recursive procedure Affinity A Perona/ Freeman Procedure/Eigenvectors used Matrix used Authors

Ax x λ =

( , ) ( , )

j

D i i A i j = ∑

( ) D A x Dx λ − =

Nugent, Stanberry UW STAT 593E

slide-76
SLIDE 76

76

Advantages/Disadvantages

  • Perona/Freeman

– For block diagonal affinity matrices, the first eigenvector finds points in the “dominant”cluster; not very consistent

  • Shi/Malik

– 2nd generalized eigenvector minimizes affinity between groups by affinity within each group; no guarantee, constraints

Nugent, Stanberry UW STAT 593E

slide-77
SLIDE 77

77

Advantages/Disadvantages

  • Scott/Longuet-Higgins

– Depends largely on choice of k – Good results

  • Ng, Jordan, Weiss

– Again depends on choice of k – Claim: effectively handles clusters whose

  • verlap or connectedness varies across clusters

Nugent, Stanberry UW STAT 593E

slide-78
SLIDE 78

Affinity Matrix Perona/Freeman Shi/Malik Scott/Lon.Higg 1st eigenv. 2nd gen. eigenv. Q matrix Affinity Matrix Perona/Freeman Shi/Malik Scott/Lon.Higg 1st eigenv. 2nd gen. eigenv. Q matrix Affinity Matrix Perona/Freeman Shi/Malik Scott/Lon.Higg 1st eigenv. 2nd gen. eigenv. Q matrix Nugent, Stanberry UW STAT 593E

slide-79
SLIDE 79

79

Segmentation and Line Fitting

  • Gestalt grouping
  • Background subtraction
  • K-Means
  • Graph cuts
  • Hough transform
  • Iterative fitting
slide-80
SLIDE 80

80

Fitting

  • Choose a parametric
  • bject/some objects to

represent a set of tokens

  • Most interesting case is

when criterion is not local

– can’t tell whether a set of points lies on a line by looking only at each point and the next.

  • Three main questions:

– what object represents this set of tokens best? – which of several objects gets which token? – how many objects are there? (you could read line for object here, or circle, or ellipse

  • r...)
slide-81
SLIDE 81

81

Fitting and the Hough Transform

  • Purports to answer all three

questions

– in practice, answer isn’t usually all that much help

  • We do for lines only
  • A line is the set of points (x, y)

such that

  • Different choices of θ, d>0 give

different lines

  • For any (x, y) there is a one

parameter family of lines through this point, given by

  • Each point gets to vote for each

line in the family; if there is a line that has lots of votes, that should be the line passing through the points

sinθ

( )x + cosθ ( )y + d = 0

sinθ

( )x + cosθ ( )y + d = 0

slide-82
SLIDE 82

82

tokens votes

slide-83
SLIDE 83

83

Mechanics of the Hough transform

  • Construct an array

representing θ, d

  • For each point, render the

curve (θ, d) into this array, adding one at each cell

  • Difficulties

– how big should the cells be? (too big, and we cannot distinguish between quite different lines; too small, and noise causes lines to be missed)

  • How many lines?

– count the peaks in the Hough array

  • Who belongs to which

line?

– tag the votes

  • Hardly ever satisfactory in

practice, because problems with noise and cell size defeat it

slide-84
SLIDE 84

84

tokens votes

slide-85
SLIDE 85

85

slide-86
SLIDE 86

86

slide-87
SLIDE 87

87

slide-88
SLIDE 88

88

Line fitting

What criteria to optimize when fitting a line to a set of points?

slide-89
SLIDE 89

89

“Least Squares” “Total Least Squares” Line fitting can be max. likelihood - but choice of model is important

slide-90
SLIDE 90

90

Who came from which line?

  • Assume we know how many lines there are
  • but which lines are they?

– easy, if we know who came from which line

  • Three strategies

– Incremental line fitting – K-means – Probabilistic (later!)

slide-91
SLIDE 91

91

slide-92
SLIDE 92

92

Incremental line fitting

slide-93
SLIDE 93

93

Incremental line fitting

slide-94
SLIDE 94

94

Incremental line fitting

slide-95
SLIDE 95

95

Incremental line fitting

slide-96
SLIDE 96

96

Incremental line fitting

slide-97
SLIDE 97

97

slide-98
SLIDE 98

98

K-means line fitting

slide-99
SLIDE 99

99

K-means line fitting

slide-100
SLIDE 100

100

K-means line fitting

slide-101
SLIDE 101

101

K-means line fitting

slide-102
SLIDE 102

102

K-means line fitting

slide-103
SLIDE 103

103

K-means line fitting

slide-104
SLIDE 104

104

K-means line fitting

slide-105
SLIDE 105

105

Robustness

  • As we have seen, squared error can be a source of

bias in the presence of noise points

– One fix is EM - we’ll do this shortly – Another is an M-estimator

  • Square nearby, threshold far away

– A third is RANSAC

  • Search for good points

(Next lecture….)

slide-106
SLIDE 106

106

Segmentation and Line Fitting

Lecture 14:

– Unsupervised Category Learning – Gestalt Principles – Segmentation by Clustering

  • K-Means
  • Graph cuts

– Segmentation by Fitting

  • Hough transform
  • Fitting

Readings: F&P Ch. 14, 15.1-15.2 (Next time: Finish fitting, Probabilistic segmentation; FP 15.4-5, 16)

slide-107
SLIDE 107

107

Visual learning is inefficient

Slide from Li Fei-Fei http://www.vision.caltech.edu/feifeili/Resume.htm

[Slide from Bradsky & Thrun, Stanford]

slide-108
SLIDE 108

108

Slide from Li Fei-Fei http://www.vision.caltech.edu/feifeili/Resume.htm

[Slide from Bradsky & Thrun, Stanford]

slide-109
SLIDE 109

109

This guy is wearing a haircut This guy is wearing a haircut called a “Mullet” called a “Mullet” [Slide from Bradsky & Thrun, Stanford]

slide-110
SLIDE 110

110

Find the Mullets…

One-Shot Learning

[Slide from Bradsky & Thrun, Stanford]

slide-111
SLIDE 111

111

One-Shot Learning

“The appearance of the categories we know and … the variability in their appearance, gives us important information on what to expect in a new category”

1.

  • L. Fei-Fei, R. Fergus and P. Perona, “A Bayesian Approach to

Unsupervised One-Shot Learning of Object Categories” ICCV 03. 2.

  • R. Fergus, P. Perona and A.Zisserman, “Object Class Recognition by

Unsupervised Scale-Invariant Learning”, CVPR 03.

  • http://www.vision.caltech.edu/html-files/publications.html

[Slide from Bradsky & Thrun, Stanford]

slide-112
SLIDE 112

112

Learn meta-parameters

set to 1.0 Training set Shape Training set Appearance Shape Appearance Model Params

[Fei Fei et al. 2003]

Learn [Slide from Bradsky & Thrun, Stanford]