INFO 4300 / CS4300 Information Retrieval slides adapted from - - PowerPoint PPT Presentation

info 4300 cs4300 information retrieval slides adapted
SMART_READER_LITE
LIVE PREVIEW

INFO 4300 / CS4300 Information Retrieval slides adapted from - - PowerPoint PPT Presentation

INFO 4300 / CS4300 Information Retrieval slides adapted from Hinrich Sch utzes, linked from http://informationretrieval.org/ IR 20/26: Linear Classifiers and Flat clustering Paul Ginsparg Cornell University, Ithaca, NY 10 Nov 2009 1 /


slide-1
SLIDE 1

INFO 4300 / CS4300 Information Retrieval slides adapted from Hinrich Sch¨ utze’s, linked from http://informationretrieval.org/

IR 20/26: Linear Classifiers and Flat clustering

Paul Ginsparg

Cornell University, Ithaca, NY

10 Nov 2009

1 / 92

slide-2
SLIDE 2

Discussion 6, 12 Nov

For this class, read and be prepared to discuss the following: Jeffrey Dean and Sanjay Ghemawat, MapReduce: Simplified Data Processing on Large Clusters. Usenix SDI ’04, 2004. http://www.usenix.org/events/osdi04/tech/full papers/dean/dean.pdf See also (Jan 2009):

http://michaelnielsen.org/blog/write-your-first-mapreduce-program-in-20-minutes/

part of lectures on “google technology stack”:

http://michaelnielsen.org/blog/lecture-course-the-google-technology-stack/

(including PageRank, etc.)

2 / 92

slide-3
SLIDE 3

Overview

1

Recap

2

Linear classifiers

3

> two classes

4

Clustering: Introduction

5

Clustering in IR

6

K-means

3 / 92

slide-4
SLIDE 4

Outline

1

Recap

2

Linear classifiers

3

> two classes

4

Clustering: Introduction

5

Clustering in IR

6

K-means

4 / 92

slide-5
SLIDE 5

Poisson Distribution

Bernoulli process with N trials, each probability p of success: p(m) = N m

  • pm(1 − p)N−m .

Probability p(m) of m successes, in limit N very large and p small, parametrized by just µ = Np (µ = mean number of successes). For N ≫ m, we have

N! (N−m)! = N(N − 1) · · · (N − m + 1) ≈ Nm,

so N

m

N! m!(N−m)! ≈ Nm m! , and

p(m) ≈ 1 m!Nm µ N m 1− µ N N−m ≈ µm m! lim

N→∞

  • 1− µ

N N = e−µµm m! (ignore (1 − µ/N)−m since by assumption N ≫ µm). N dependence drops out for N → ∞, with average µ fixed (p → 0). The form p(m) = e−µ µm

m! is known as a Poisson distribution

(properly normalized: ∞

m=0 p(m) = e−µ ∞ m=0 µm m! = e−µ ·eµ = 1).

5 / 92

slide-6
SLIDE 6

Poisson Distribution for µ = 10

p(m) = e−10 10m

m!

0.02 0.04 0.06 0.08 0.1 0.12 0.14 5 10 15 20 25 30

Compare to power law p(m) ∝ 1/m2.1

6 / 92

slide-7
SLIDE 7

Classes in the vector space

x x x x

⋄ ⋄ ⋄ ⋄ ⋄ ⋄

China Kenya UK

Should the document ⋆ be assigned to China, UK or Kenya? Find separators between the classes Based on these separators: ⋆ should be assigned to China How do we find separators that do a good job at classifying new documents like ⋆?

7 / 92

slide-8
SLIDE 8

Rocchio illustrated: a1 = a2, b1 = b2, c1 = c2

x x x x

⋄ ⋄ ⋄ ⋄ ⋄ ⋄

China Kenya UK

a1 a2 b1 b2 c1 c2

8 / 92

slide-9
SLIDE 9

kNN classification

kNN classification is another vector space classification method. It also is very simple and easy to implement. kNN is more accurate (in most cases) than Naive Bayes and Rocchio. If you need to get a pretty accurate classifier up and running in a short time . . . . . . and you don’t care about efficiency that much . . . . . . use kNN.

9 / 92

slide-10
SLIDE 10

kNN is based on Voronoi tessellation

x x x x x x x x x x x

⋄ ⋄ ⋄ ⋄ ⋄ ⋄ ⋄ ⋄ ⋄ ⋄ ⋄

1NN, 3NN classifica- tion decision for star?

10 / 92

slide-11
SLIDE 11

Exercise

⋆ x x x x x x x x x x

  • How is star classified by:

(i) 1-NN (ii) 3-NN (iii) 9-NN (iv) 15-NN (v) Rocchio

11 / 92

slide-12
SLIDE 12

kNN: Discussion

No training necessary

But linear preprocessing of documents is as expensive as training Naive Bayes. You will always preprocess the training set, so in reality training time of kNN is linear.

kNN is very accurate if training set is large. Optimality result: asymptotically zero error if Bayes rate is zero. But kNN can be very inaccurate if training set is small.

12 / 92

slide-13
SLIDE 13

Outline

1

Recap

2

Linear classifiers

3

> two classes

4

Clustering: Introduction

5

Clustering in IR

6

K-means

13 / 92

slide-14
SLIDE 14

Linear classifiers

Linear classifiers compute a linear combination or weighted sum

i wixi of the feature values.

Classification decision:

i wixi > θ?

. . . where θ (the threshold) is a parameter. (First, we only consider binary classifiers.) Geometrically, this corresponds to a line (2D), a plane (3D) or a hyperplane (higher dimensionalities) Assumption: The classes are linearly separable. Can find hyperplane (=separator) based on training set Methods for finding separator: Perceptron, Rocchio, Naive Bayes – as we will explain on the next slides

14 / 92

slide-15
SLIDE 15

A linear classifier in 1D

x1 A linear classifier in 1D is a point described by the equation w1x1 = θ The point at θ/w1 Points (x1) with w1x1 ≥ θ are in the class c. Points (x1) with w1x1 < θ are in the complement class c.

15 / 92

slide-16
SLIDE 16

A linear classifier in 2D

A linear classifier in 2D is a line described by the equation w1x1 + w2x2 = θ Example for a 2D linear classifier Points (x1 x2) with w1x1 + w2x2 ≥ θ are in the class c. Points (x1 x2) with w1x1 + w2x2 < θ are in the complement class c.

16 / 92

slide-17
SLIDE 17

A linear classifier in 3D

A linear classifier in 3D is a plane described by the equation w1x1 + w2x2 + w3x3 = θ Example for a 3D linear classifier Points (x1 x2 x3) with w1x1 + w2x2 + w3x3 ≥ θ are in the class c. Points (x1 x2 x3) with w1x1 + w2x2 + w3x3 < θ are in the complement class c.

17 / 92

slide-18
SLIDE 18

Rocchio as a linear classifier

Rocchio is a linear classifier defined by:

M

  • i=1

wixi = w · x = θ where the normal vector w = µ(c1) − µ(c2) and θ = 0.5 ∗ (| µ(c1)|2 − | µ(c2)|2). (follows from decision boundary | µ(c1) − x| = | µ(c2) − x|)

18 / 92

slide-19
SLIDE 19

Naive Bayes classifier

(Just like BIM, see lecture 13)

  • x represents document, what is p(c|

x) that document is in class c? p(c| x) = p( x|c)p(c) p( x) p(¯ c| x) = p( x|¯ c)p(¯ c) p( x)

  • dds :

p(c| x) p(¯ c| x) = p( x|c)p(c) p( x|¯ c)p(¯ c) ≈ p(c) p(¯ c)

  • 1≤k≤nd p(tk|c)
  • 1≤k≤nd p(tk|¯

c) log odds : log p(c| x) p(¯ c| x) = log p(c) p(¯ c) +

  • 1≤k≤nd

log p(tk|c) p(tk|¯ c)

19 / 92

slide-20
SLIDE 20

Naive Bayes as a linear classifier

Naive Bayes is a linear classifier defined by:

M

  • i=1

wixi = θ where wi = log

  • p(ti|c)/p(ti|¯

c)

  • ,

xi = number of occurrences of ti in d, and θ = − log

  • p(c)/p(¯

c)

  • .

(the index i, 1 ≤ i ≤ M, refers to terms of the vocabulary) Linear in log space

20 / 92

slide-21
SLIDE 21

kNN is not a linear classifier

x x x x x x x x x x x

⋄ ⋄ ⋄ ⋄ ⋄ ⋄ ⋄ ⋄ ⋄ ⋄ ⋄

Classification decision based on majority of k nearest neighbors. The decision boundaries between classes are piecewise linear . . . . . . but they are not linear classifiers that can be described as M

i=1 wixi = θ.

21 / 92

slide-22
SLIDE 22

Example of a linear two-class classifier

ti wi x1i x2i ti wi x1i x2i prime 0.70 1 dlrs

  • 0.71

1 1 rate 0.67 1 world

  • 0.35

1 interest 0.63 sees

  • 0.33

rates 0.60 year

  • 0.25

discount 0.46 1 group

  • 0.24

bundesbank 0.43 dlr

  • 0.24

This is for the class interest in Reuters-21578. For simplicity: assume a simple 0/1 vector representation x1: “rate discount dlrs world” x2: “prime dlrs” Exercise: Which class is x1 assigned to? Which class is x2 assigned to? We assign document d1 “rate discount dlrs world” to interest since

  • wT ·

d1 = 0.67 · 1 + 0.46 · 1 + (−0.71) · 1 + (−0.35) · 1 = 0.07 > 0 = b. We assign d2 “prime dlrs” to the complement class (not in interest) since

  • wT ·

d2 = −0.01 ≤ b. (dlr and world have negative weights because they are indicators for the competing class currency)

22 / 92

slide-23
SLIDE 23

Which hyperplane?

23 / 92

slide-24
SLIDE 24

Which hyperplane?

For linearly separable training sets: there are infinitely many separating hyperplanes. They all separate the training set perfectly . . . . . . but they behave differently on test data. Error rates on new data are low for some, high for others. How do we find a low-error separator? Perceptron: generally bad; Naive Bayes, Rocchio: ok; linear SVM: good

24 / 92

slide-25
SLIDE 25

Linear classifiers: Discussion

Many common text classifiers are linear classifiers: Naive Bayes, Rocchio, logistic regression, linear support vector machines etc. Each method has a different way of selecting the separating hyperplane

Huge differences in performance on test documents

Can we get better performance with more powerful nonlinear classifiers? Not in general: A given amount of training data may suffice for estimating a linear boundary, but not for estimating a more complex nonlinear boundary.

25 / 92

slide-26
SLIDE 26

A nonlinear problem

0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0

Linear classifier like Rocchio does badly on this task. kNN will do well (assuming enough training data)

26 / 92

slide-27
SLIDE 27

A linear problem with noise

Figure 14.10: hypothetical web page classification scenario: Chinese-only web pages (solid circles) and mixed Chinese-English web (squares). linear class boundary, except for three noise docs

27 / 92

slide-28
SLIDE 28

Which classifier do I use for a given TC problem?

Is there a learning method that is optimal for all text classification problems? No, because there is a tradeoff between bias and variance. Factors to take into account:

How much training data is available? How simple/complex is the problem? (linear vs. nonlinear decision boundary) How noisy is the problem? How stable is the problem over time?

For an unstable problem, it’s better to use a simple and robust classifier.

28 / 92

slide-29
SLIDE 29

Outline

1

Recap

2

Linear classifiers

3

> two classes

4

Clustering: Introduction

5

Clustering in IR

6

K-means

29 / 92

slide-30
SLIDE 30

How to combine hyperplanes for > 2 classes?

?

(e.g.: rank and select top-ranked classes)

30 / 92

slide-31
SLIDE 31

One-of problems

One-of or multiclass classification

Classes are mutually exclusive. Each document belongs to exactly one class. Example: language of a document (assumption: no document contains multiple languages)

31 / 92

slide-32
SLIDE 32

One-of classification with linear classifiers

Combine two-class linear classifiers as follows for one-of classification:

Run each classifier separately Rank classifiers (e.g., according to score) Pick the class with the highest score

32 / 92

slide-33
SLIDE 33

Any-of problems

Any-of or multilabel classification

A document can be a member of 0, 1, or many classes. A decision on one class leaves decisions open on all other classes. A type of “independence” (but not statistical independence) Example: topic classification Usually: make decisions on the region, on the subject area, on the industry and so on “independently”

33 / 92

slide-34
SLIDE 34

Any-of classification with linear classifiers

Combine two-class linear classifiers as follows for any-of classification:

Simply run each two-class classifier separately on the test document and assign document accordingly

34 / 92

slide-35
SLIDE 35

Outline

1

Recap

2

Linear classifiers

3

> two classes

4

Clustering: Introduction

5

Clustering in IR

6

K-means

35 / 92

slide-36
SLIDE 36

What is clustering?

(Document) clustering is the process of grouping a set of documents into clusters of similar documents. Documents within a cluster should be similar. Documents from different clusters should be dissimilar. Clustering is the most common form of unsupervised learning. Unsupervised = there are no labeled or annotated data.

36 / 92

slide-37
SLIDE 37

Data set with clear cluster structure

0.0 0.5 1.0 1.5 2.0 0.0 0.5 1.0 1.5 2.0 2.5

37 / 92

slide-38
SLIDE 38

Classification vs. Clustering

Classification: supervised learning Clustering: unsupervised learning Classification: Classes are human-defined and part of the input to the learning algorithm. Clustering: Clusters are inferred from the data without human input.

However, there are many ways of influencing the outcome of clustering: number of clusters, similarity measure, representation of documents, . . .

38 / 92

slide-39
SLIDE 39

Outline

1

Recap

2

Linear classifiers

3

> two classes

4

Clustering: Introduction

5

Clustering in IR

6

K-means

39 / 92

slide-40
SLIDE 40

The cluster hypothesis

Cluster hypothesis. Documents in the same cluster behave similarly with respect to relevance to information needs. All applications in IR are based (directly or indirectly) on the cluster hypothesis.

40 / 92

slide-41
SLIDE 41

Applications of clustering in IR

Application What is Benefit Example clustered? Search result clustering search results more effective infor- mation presentation to user Scatter-Gather (subsets

  • f)

col- lection alternative user inter- face: “search without typing” Collection clustering collection effective information presentation for ex- ploratory browsing McKeown et al. 2002, news.google.com Cluster-based retrieval collection higher efficiency: faster search Salton 1971

41 / 92

slide-42
SLIDE 42

Search result clustering for better navigation

42 / 92

slide-43
SLIDE 43

Scatter-Gather

43 / 92

slide-44
SLIDE 44

Global navigation: Yahoo

44 / 92

slide-45
SLIDE 45

Global navigation: MESH (upper level)

45 / 92

slide-46
SLIDE 46

Global navigation: MESH (lower level)

46 / 92

slide-47
SLIDE 47

Note: Yahoo/MESH are not examples of clustering. But they are well known examples for using a global hierarchy for navigation. Some examples for global navigation/exploration based on clustering:

Cartia Themescapes Google News

47 / 92

slide-48
SLIDE 48

Global navigation combined with visualization (1)

48 / 92

slide-49
SLIDE 49

Global navigation combined with visualization (2)

49 / 92

slide-50
SLIDE 50

Global clustering for navigation: Google News

http://news.google.com

50 / 92

slide-51
SLIDE 51

Clustering for improving recall

To improve search recall:

Cluster docs in collection a priori When a query matches a doc d, also return other docs in the cluster containing d

Hope: if we do this: the query “car” will also return docs containing “automobile”

Because clustering groups together docs containing “car” with those containing “automobile”. Both types of documents contain words like “parts”, “dealer”, “mercedes”, “road trip”.

51 / 92

slide-52
SLIDE 52

Data set with clear cluster structure

0.0 0.5 1.0 1.5 2.0 0.0 0.5 1.0 1.5 2.0 2.5

Exercise: Come up with an algorithm for finding the three clusters in this case

52 / 92

slide-53
SLIDE 53

Document representations in clustering

Vector space model As in vector space classification, we measure relatedness between vectors by Euclidean distance . . . . . . which is almost equivalent to cosine similarity. Almost: centroids are not length-normalized. For centroids, distance and cosine give different results.

53 / 92

slide-54
SLIDE 54

Issues in clustering

General goal: put related docs in the same cluster, put unrelated docs in different clusters.

But how do we formalize this?

How many clusters?

Initially, we will assume the number of clusters K is given.

Often: secondary goals in clustering

Example: avoid very small and very large clusters

Flat vs. hierarchical clustering Hard vs. soft clustering

54 / 92

slide-55
SLIDE 55

Flat vs. Hierarchical clustering

Flat algorithms

Usually start with a random (partial) partitioning of docs into groups Refine iteratively Main algorithm: K-means

Hierarchical algorithms

Create a hierarchy Bottom-up, agglomerative Top-down, divisive

55 / 92

slide-56
SLIDE 56

Hard vs. Soft clustering

Hard clustering: Each document belongs to exactly one cluster.

More common and easier to do

Soft clustering: A document can belong to more than one cluster.

Makes more sense for applications like creating browsable hierarchies You may want to put a pair of sneakers in two clusters:

sports apparel shoes

You can only do that with a soft clustering approach.

For soft clustering, see course text: 16.5,18

Today: Flat, hard clustering Next time: Hierarchical, hard clustering

56 / 92

slide-57
SLIDE 57

Flat algorithms

Flat algorithms compute a partition of N documents into a set of K clusters. Given: a set of documents and the number K Find: a partition in K clusters that optimizes the chosen partitioning criterion Global optimization: exhaustively enumerate partitions, pick

  • ptimal one

Not tractable

Effective heuristic method: K-means algorithm

57 / 92

slide-58
SLIDE 58

Outline

1

Recap

2

Linear classifiers

3

> two classes

4

Clustering: Introduction

5

Clustering in IR

6

K-means

58 / 92

slide-59
SLIDE 59

K-means

Perhaps the best known clustering algorithm Simple, works well in many cases Use as default / baseline for clustering documents

59 / 92

slide-60
SLIDE 60

K-means

Each cluster in K-means is defined by a centroid. Objective/partitioning criterion: minimize the average squared difference from the centroid Recall definition of centroid:

  • µ(ω) = 1

|ω|

  • x∈ω
  • x

where we use ω to denote a cluster. We try to find the minimum average squared difference by iterating two steps:

reassignment: assign each vector to its closest centroid recomputation: recompute each centroid as the average of the vectors that were assigned to it in reassignment

60 / 92

slide-61
SLIDE 61

K-means algorithm

K-means({ x1, . . . , xN}, K) 1 ( s1, s2, . . . , sK) ← SelectRandomSeeds({ x1, . . . , xN}, K) 2 for k ← 1 to K 3 do µk ← sk 4 while stopping criterion has not been met 5 do for k ← 1 to K 6 do ωk ← {} 7 for n ← 1 to N 8 do j ← arg minj′ | µj′ − xn| 9 ωj ← ωj ∪ { xn} (reassignment of vectors) 10 for k ← 1 to K 11 do µk ←

1 |ωk|

  • x∈ωk

x (recomputation of centroids) 12 return { µ1, . . . , µK}

61 / 92

slide-62
SLIDE 62

Set of points to be clustered

b b b b b b b b b b b b b b b b b b b b

62 / 92

slide-63
SLIDE 63

Random selection of initial cluster centers

× ×

b b b b b b b b b b b b b b b b b b b b

Centroids after convergence?

63 / 92

slide-64
SLIDE 64

Assign points to closest center

b b b b b b b b b b b b b b b b b b b b

× ×

64 / 92

slide-65
SLIDE 65

Assignment

2 1 1 2 1 1 1 1 1 1 1 1 1 1 1 2 1 1 2 2

× ×

65 / 92

slide-66
SLIDE 66

Recompute cluster centroids

2 1 1 2 1 1 1 1 1 1 1 1 1 1 1 2 1 1 2 2

× ×

× ×

66 / 92

slide-67
SLIDE 67

Assign points to closest centroid

b b b b b b b b b b b b b b b b b b b b

× ×

67 / 92

slide-68
SLIDE 68

Assignment

2 2 1 2 1 1 1 1 1 1 1 2 1 1 1 2 1 1 2 2

× ×

68 / 92

slide-69
SLIDE 69

Recompute cluster centroids

2 2 1 2 1 1 1 1 1 1 1 2 1 1 1 2 1 1 2 2

× ×

× ×

69 / 92

slide-70
SLIDE 70

Assign points to closest centroid

b b b b b b b b b b b b b b b b b b b b

× ×

70 / 92

slide-71
SLIDE 71

Assignment

2 2 2 2 1 1 1 1 1 1 1 2 1 1 1 2 1 1 2 2

× ×

71 / 92

slide-72
SLIDE 72

Recompute cluster centroids

2 2 2 2 1 1 1 1 1 1 1 2 1 1 1 2 1 1 2 2

× ×

× ×

72 / 92

slide-73
SLIDE 73

Assign points to closest centroid

b b b b b b b b b b b b b b b b b b b b

× ×

73 / 92

slide-74
SLIDE 74

Assignment

2 2 2 2 1 1 1 1 2 1 1 2 1 1 1 2 1 1 2 2

× ×

74 / 92

slide-75
SLIDE 75

Recompute cluster centroids

2 2 2 2 1 1 1 1 2 1 1 2 1 1 1 2 1 1 2 2

× ×

× ×

75 / 92

slide-76
SLIDE 76

Assign points to closest centroid

b b b b b b b b b b b b b b b b b b b b

× ×

76 / 92

slide-77
SLIDE 77

Assignment

2 2 2 2 1 1 1 1 2 2 1 2 1 1 1 1 1 1 2 1

× ×

77 / 92

slide-78
SLIDE 78

Recompute cluster centroids

2 2 2 2 1 1 1 1 2 2 1 2 1 1 1 1 1 1 2 1

× ×

× ×

78 / 92

slide-79
SLIDE 79

Assign points to closest centroid

b b b b b b b b b b b b b b b b b b b b

× ×

79 / 92

slide-80
SLIDE 80

Assignment

2 2 2 2 1 1 1 1 2 2 1 2 1 1 1 1 1 1 1 1

× ×

80 / 92

slide-81
SLIDE 81

Recompute cluster centroids

2 2 2 2 1 1 1 1 2 2 1 2 1 1 1 1 1 1 1 1

× ×

× ×

81 / 92

slide-82
SLIDE 82

Assign points to closest centroid

b b b b b b b b b b b b b b b b b b b b

× ×

82 / 92

slide-83
SLIDE 83

Assignment

2 2 2 2 1 1 1 1 2 2 1 1 1 1 1 1 1 1 1 1

× ×

83 / 92

slide-84
SLIDE 84

Recompute cluster centroids

2 2 2 2 1 1 1 1 2 2 1 1 1 1 1 1 1 1 1 1

× ×

× ×

84 / 92

slide-85
SLIDE 85

Centroids and assignments after convergence

2 2 2 2 1 1 1 1 2 2 1 1 1 1 1 1 1 1 1 1

× ×

85 / 92

slide-86
SLIDE 86

K-means is guaranteed to converge

Proof: The sum of squared distances (RSS) decreases during reassignment.

RSS = sum of all squared distances between document vector and closest centroid

(because each vector is moved to a closer centroid) RSS decreases during recomputation. (We will show this on the next slide.) There is only a finite number of clusterings. Thus: We must reach a fixed point. (assume that ties are broken consistently)

86 / 92

slide-87
SLIDE 87

Recomputation decreases average distance

RSS = K

k=1 RSSk – the residual sum of squares (the “goodness”

measure) RSSk( v) =

  • x∈ωk
  • v −

x2 =

  • x∈ωk

M

  • m=1

(vm − xm)2 ∂RSSk( v) ∂vm =

  • x∈ωk

2(vm − xm) = 0 vm = 1 |ωk|

  • x∈ωk

xm The last line is the componentwise definition of the centroid! We minimize RSSk when the old centroid is replaced with the new

  • centroid. RSS, the sum of the RSSk, must then also decrease

during recomputation.

87 / 92

slide-88
SLIDE 88

K-means is guaranteed to converge

But we don’t know how long convergence will take! If we don’t care about a few docs switching back and forth, then convergence is usually fast (< 10-20 iterations). However, complete convergence can take many more iterations.

88 / 92

slide-89
SLIDE 89

Optimality of K-means

Convergence does not mean that we converge to the optimal clustering! This is the great weakness of K-means. If we start with a bad set of seeds, the resulting clustering can be horrible.

89 / 92

slide-90
SLIDE 90

Exercise: Suboptimal clustering

1 2 3 4 1 2 3

× × × × × ×

d1 d2 d3 d4 d5 d6 What is the optimal clustering for K = 2? Do we converge on this clustering for arbitrary seeds di1, di2?

90 / 92

slide-91
SLIDE 91

Initialization of K-means

Random seed selection is just one of many ways K-means can be initialized. Random seed selection is not very robust: It’s easy to get a suboptimal clustering. Better heuristics:

Select seeds not randomly, but using some heuristic (e.g., filter

  • ut outliers or find a set of seeds that has “good coverage” of

the document space) Use hierarchical clustering to find good seeds (next class) Select i (e.g., i = 10) different sets of seeds, do a K-means clustering for each, select the clustering with lowest RSS

91 / 92

slide-92
SLIDE 92

Time complexity of K-means

Computing one distance of two vectors is O(M). Reassignment step: O(KNM) (we need to compute KN document-centroid distances) Recomputation step: O(NM) (we need to add each of the document’s < M values to one of the centroids) Assume number of iterations bounded by I Overall complexity: O(IKNM) – linear in all important dimensions However: This is not a real worst-case analysis. In pathological cases, the number of iterations can be much higher than linear in the number of documents.

92 / 92