Clustering: Models and Algorithms
Shikui Tu 2019-02-28
1
Clustering: Models and Algorithms Shikui Tu 2019-02-28 1 Outline - - PowerPoint PPT Presentation
Clustering: Models and Algorithms Shikui Tu 2019-02-28 1 Outline Clustering K-mean clustering, hierarchical clustering Adaptive learning (online learning) CL, FSCL, RPCL Gaussian Mixture Models (GMM)
1
2
3
! xt ||! - xt ||2
||! - x1 ||2 + ||! - x2 ||2 + ||! - x3 ||2
!
5
6
http://www2.imm.dtu.dk/pubdb/views/edoc_download.php/3274/pdf/imm3274.pdf
We have the following data: We want to cluster the data into two clusters (red and blue)
7
rnk = 1 if and only if data point xn is assigned to cluster k;
minimize
k = 1, 2; K = 2 clusters n = 1, …, N; N: the total number of points. rn1 = 1 rn2 = 0
µ2 µ1
We need to calculate { rnk }and { µk }.
8
µ2 µ1
Since the points have been assigned to cluster 1 or cluster 2, we calculate
µ1 = mean of the points in cluster 1
Or formally
µ2 = mean of the points in cluster 2
We call it the M Step.
9
µ2 µ1
|| xn – µ1 ||2 < || xn – µ2 ||2 We should assign point xn to cluster 1, because Then,
rn1 = 1 rn2 = 0
Or formally We call it the E Step
10
µ2 µ1
11
E Step
Assign the points to the nearest cluster:
Steps
Equal distance line
12
M Step
Calculate the means of the points in each cluster:
Steps 13
E Step
Assign the points to the nearest cluster:
Steps 14
M Step
Calculate the means of the points in each cluster:
Steps 15
Initialization E-Step M-Step E-Step M-Step M-Step E-Step E-Step Convergence If J does not change, or { µ1, µ2 } do not change, then the algorithm converges.
16
– No, the nearest local optimum, depending on initialization
19
20
Adapted from Blei, D. Hierarchial Cluster [PwerPoint slides]. www.cs.princeton.edu/courses/archive/spr08/cos424/slides/clustering-2.pdf
21
Adapted from Blei, D. Hierarchial Cluster [PwerPoint slides]. www.cs.princeton.edu/courses/archive/spr08/cos424/slides/clustering-2.pdf
22
a∈A,b∈B d(a, b)
a∈A,b∈B d(a, b)
a∈A,b∈B d(a, b)
23
Jain, A. K., Murty, M. N., Flynn, P. J. (1999) "Data Clustering: A Review". ACM Computing Surveys (CSUR), 31(3), p.264-323, 1999.
24
Distance
25
26
x1 x2 xN
27
x1 x2 xN
28
29
30
The idea is to penalize the frequent winners:
31
When k is pre-assigned to 5. the frequency sensitive mechanism also brings the extra one into data to disturb the correct locations of others
32
(Xu, Krzyzak, & Oja, 1992 , 1993) The RPCL differs from FSCL by implementing pj,t as follows:
where γ approximately takes a number between 0.05 and 0.1 for controlling the penalizing strength.
33
Rival penalized mechanism makes extra agents driven far away.
53