Shape Co-analysis and constrained clustering Daniel Cohen-Or - - PowerPoint PPT Presentation

shape co analysis and constrained clustering
SMART_READER_LITE
LIVE PREVIEW

Shape Co-analysis and constrained clustering Daniel Cohen-Or - - PowerPoint PPT Presentation

Shape Co-analysis and constrained clustering Daniel Cohen-Or Tel-Aviv University 1 High-level Shape analysis [Mehra et al. 08] [Fu et al. 08] [Kalograkis et al. 10] Shape abstraction Upright orientation Learning segmentation [Kim et al.


slide-1
SLIDE 1

Shape Co-analysis and constrained clustering

Daniel Cohen-Or

Tel-Aviv University

1

slide-2
SLIDE 2

High-level Shape analysis

2

Shape abstraction

[Mitra et al. 10] [Mehra et al. 08] [Fu et al. 08]

Upright orientation Illustrating assemblies

[Wang et al. 11]

Symmetry hierarchy

Learning segmentation

[Kalograkis et al. 10]

Exploration of shape collections

[Kim et al. 12]

slide-3
SLIDE 3

Segmentation and Correspondence

3

Segmentation Correspondence

slide-4
SLIDE 4

Individual vs. Co-segmentation

4

slide-5
SLIDE 5

Individual vs. Co-segmentation

5

slide-6
SLIDE 6

Challenge

6

Similar geometries can be associated with different semantics

slide-7
SLIDE 7

Challenge

7

Similar semantics can be represented by different geometries

slide-8
SLIDE 8

Large set are more challenging

8

Methods do not give perfect results

slide-9
SLIDE 9

11

[Sidi et al.11]

Descriptor-based unsupervised co-segmentation

slide-10
SLIDE 10

co-segmentation

12

Clustering in feature space

slide-11
SLIDE 11

Clustering (basic stuff)

Takes a set of points,

slide-12
SLIDE 12

Clustering (basic stuff)

Takes a set of points, and groups them into several separate clusters

slide-13
SLIDE 13

Clustering is not easy…

– Clean separation to groups not always possible – Must make “hard splitting” decisions – Number of groups not always known, or can be very difficult to determine from data

slide-14
SLIDE 14

Clustering is hard!

slide-15
SLIDE 15

Clustering is hard!

Hard to determine number of clusters

slide-16
SLIDE 16

Clustering is hard!

Hard to determine number of clusters

slide-17
SLIDE 17

Clustering is hard!

Hard to decide where to split clusters

slide-18
SLIDE 18

Clustering

Hard to decide where to split clusters

slide-19
SLIDE 19

Clustering

  • Two general types of input for Clustering:

– Spatial Coordinates (points, feature space), or – Inter‐object Distance matrix

slide-20
SLIDE 20

Clustering

Spatial Coordinates (points, feature space), or Inter‐object Distance matrix

?

K‐Means, EM, Mean‐Shift, Linkage, DBSCAN, Spectral Clustering

slide-21
SLIDE 21

Clustering 101

slide-22
SLIDE 22

Initial co-segmentation

  • Over-segmentation mapped to a descriptor

space (geodesic distance, shape diameter function, normal histogram)

24

High‐dimensional feature space

slide-23
SLIDE 23

Co‐segmentation

Points represent some kind of object parts, and we want to cluster them as means to co‐ segment the set of objects

slide-24
SLIDE 24

Clustering

  • Underlying assumptions behind all clustering

algorithms:

– Neighboring points imply similar parts.

slide-25
SLIDE 25

Clustering

  • Underlying assumptions behind all clustering

algorithms:

– Distant points imply dissimilar parts.

slide-26
SLIDE 26

Clustering

  • When assumptions fail, result is not useful:

– Similar parts are distant in feature space

slide-27
SLIDE 27

Clustering

  • When assumptions fail, result is not useful:

– Dissimilar parts are close in feature space

slide-28
SLIDE 28

Clustering

  • Assumptions might fail because:

– Data is difficult to analyze – Similarity/Dissimilarity of data not well defined – Feature space is insufficient or distorted

slide-29
SLIDE 29

Add training set of labeled data (pre‐clustered)

Supervised Clustering

slide-30
SLIDE 30

Then clustering becomes easy…

Supervised Clustering

slide-31
SLIDE 31

Supervised segmentation

34

Labeled shape

Head Nec k Torso Leg Tail Ear

Input shape Training shapes

[Kalogerakis et al.10, van Kaick et al. 11]

slide-32
SLIDE 32

Semi‐Supervised Clustering

  • Supervision as pair‐wise constraints:

– Must Link and Cannot‐Link

Can’t Link Must Link

slide-33
SLIDE 33

Semi‐Supervised Clustering

  • Cluster data while respecting constraints
slide-34
SLIDE 34

Learning from labeled and unlabeled data

39

slide-35
SLIDE 35

Supervised learning

40

slide-36
SLIDE 36

Unsupervised learning

41

slide-37
SLIDE 37

Semi-supervised learning

42

slide-38
SLIDE 38

Constrained clustering

43

Cannot Link Must Link

slide-39
SLIDE 39

Active Co‐analysis of a Set of Shapes Wang et al. SIGGRAPH ASIA 2012

slide-40
SLIDE 40

Active Co-Analysis

  • A semi-supervised method for co-segmentation

with minimal user input

45

slide-41
SLIDE 41

Automatically suggest the user which constraints can be effective

Initial Co-segmentation Constrained Clustering Final result Active Learning 46

slide-42
SLIDE 42

Cannot Link Must Link

Constrained Clustering

53

slide-43
SLIDE 43

Spring System

  • A spring system is used to re‐embed all the

points in the feature space, so the result of clustering will satisfy constraints.

slide-44
SLIDE 44

Spring System

  • Result of clustering after re‐embedding

(mistakes marked with circle):

Result of Spring Re‐embedding and Clustering Final Result

slide-45
SLIDE 45

Spring System

Neighbor Spring Must‐Link Can’t Link Random Springs

slide-46
SLIDE 46
slide-47
SLIDE 47

64

slide-48
SLIDE 48

65

Constrained clustering & Co-segmentation

slide-49
SLIDE 49

Uncertain points

  • “Uncertain” points are located using the

Silhouette Index:

Darker points have lower confidence

68

slide-50
SLIDE 50

Silhouette Index

  • Silhouette Index of node x:

69

slide-51
SLIDE 51

Constraint Suggestion

  • Pick super-faces with lowest confidence

?

  • Pick the highest confidence super-faces
  • Ask the user to add constraints between such pairs

70

slide-52
SLIDE 52

71

slide-53
SLIDE 53

Candelabra: 28 shapes, 164 super-faces,24 constraints

73

slide-54
SLIDE 54

Fourleg: 20 shapes, 264 super-faces,69 constraints

74

slide-55
SLIDE 55

Tele-alien: 200 shapes, 1869 super-faces,106 constraints

75

slide-56
SLIDE 56

Vase: 300 shapes, 1527 super-faces,44 constraints

76

slide-57
SLIDE 57

Cannot‐Link Springs

slide-58
SLIDE 58

Constraints as Features

CVPR 2013

  • Goal: Modify data so distances fit constraints
  • Basic idea:

– Convert constraints into extra‐features that are added to the data (augmentation) – Recalculate the distances – Unconstrained clustering of the modified data – Clustering result more likely to satisfy constraints

  • Apply this idea to Cannot‐Link constraints
  • Must‐Link constraints handled differently
slide-59
SLIDE 59

Cannot‐link Constraints

  • Points should be distant.
  • What value should be given: D(c1, c2) = X ?

– Should relate to max(D(x, y)), but how?

  • If modified, how to restore triangle‐inequality?
slide-60
SLIDE 60

Constraints as Features

  • Solution:

– Add extra‐dimension, where Cannot‐Link pair is far away (±1):

slide-61
SLIDE 61

Constraints as Features

  • Solution:

– Add extra‐dimension, where Cannot‐Link pair is far away (±1): – What values should other points be given?

slide-62
SLIDE 62

Constraints as Features

  • Values of other points:

– Points closer to c1 should have values closer to +1, – Points closer to c2 should have values closer to ‐1

  • Formulation:
  • Simple distance does not convey real

closeness.

slide-63
SLIDE 63

Constraints as Features

  • Point A should be “closer” to c1, despite

smaller Euclidean distance.

A c1 c2 A

slide-64
SLIDE 64

Constraints as Features

  • Use a Diffusion Map, where this holds true.

A c1 c2 A

slide-65
SLIDE 65

Constraints as Features

  • Diffusion Maps related to random walk

process on a graph

  • Affinity Matrix:
  • Eigen‐Analysis of normalized A forms a

Diffusion Map:

slide-66
SLIDE 66

Constraints as Features

  • Use Diffusion Map distances:
  • Calculate value of each point in new

dimension:

slide-67
SLIDE 67

Constraints as Features

  • Create new distance matrix, of distances in the

new extra dimension:

  • Add distance matrix per Cannot‐Link:
  • Cluster data by modified distance matrix
slide-68
SLIDE 68

Constraints as Features

Original Springs Features

slide-69
SLIDE 69

Constraints as Features!!!

Unconstrained clustering of the modified data

slide-70
SLIDE 70

Results – UCI (CVPR 2013)

slide-71
SLIDE 71

Summary

  • A new semi‐supervised clustering method.
  • Constraints are embedded into the data,

reducing the problem to an unconstrained setting.

slide-72
SLIDE 72

Thank you!

105

slide-73
SLIDE 73
slide-74
SLIDE 74
slide-75
SLIDE 75
slide-76
SLIDE 76
slide-77
SLIDE 77
slide-78
SLIDE 78
slide-79
SLIDE 79
slide-80
SLIDE 80
slide-81
SLIDE 81

Thank you!

114