10601
play

10601 Machine Learning Hierarchical clustering Reading: Bishop: - PowerPoint PPT Presentation

10601 Machine Learning Hierarchical clustering Reading: Bishop: 9-9.2 Second half: Overview Clustering - Hierarchical, semi-supervised learning Graphical models - Bayesian networks, HMMs, Reasoning under uncertainty Putting it


  1. 10601 Machine Learning Hierarchical clustering Reading: Bishop: 9-9.2

  2. Second half: Overview • Clustering - Hierarchical, semi-supervised learning • Graphical models - Bayesian networks, HMMs, Reasoning under uncertainty • Putting it together - Model / feature selection, Boosting, dimensionality reduction • Advanced classification - SVM

  3. What is Clustering? • Organizing data into clusters such that there is • high intra-cluster similarity • low inter-cluster similarity • Informally, finding natural groupings among objects. • Why do we want to do that? • Any REAL application?

  4. Example: clusty

  5. Example: clustering genes • Microarrays measures the activities of all genes in different conditions • Clustering genes can help determine new functions for unknown genes • An early “killer application” in this area – The most cited (12,309) paper in PNAS!

  6. Unsupervised learning • Clustering methods are unsupervised learning techniques - We do not have a teacher that provides examples with their labels • We will also discuss dimensionality reduction, another unsupervised learning method later in the course

  7. Outline • Distance functions • Hierarchical clustering • Number of clusters

  8. What is Similarity? The quality or state of being similar; likeness; resemblance; as, a similarity of features. Webster's Dictionary Similarity is hard to define, but… “ We know it when we see it ” The real meaning of similarity is a philosophical question. We will take a more pragmatic approach.

  9. Defining Distance Measures Definition : Let O 1 and O 2 be two objects from the universe of possible objects. The distance (dissimilarity) between O 1 and O 2 is a real number denoted by D ( O 1 , O 2 ) gene1 gene2 0.23 3 342.7

  10. gene2 gene1 Inside these black boxes: d('' '', ' , '') ') = = 0 0 d(s d(s, '') , '') = = d(' d('', ', some function on two variables s) = | |s| -- -- i.e. length of s d(s1+ch1, , s2+ch2) = m min( d(s1, (might be simple or very s2) + if ch1=ch2 then 0 else 1 f fi, d(s1+ch1, , s2) 2) + 1, 1, d(s d(s1, 1, s2+ch2 h2) complex) + 1 ) ) 3  A few examples: d ( x , y )  ( x i  y i ) 2 • Euclidian distance • Similarity rather than distance i • Can determine similar trends • Correlation coefficient  ( x i   x )( y i   y ) ฀  s ( x , y )  i  x  y ฀ 

  11. Outline • Distance measure • Hierarchical clustering • Number of clusters

  12. Desirable Properties of a Clustering Algorithm • Scalability (in terms of both time and space) • Ability to deal with different data types • Minimal requirements for domain knowledge to determine input parameters • Interpretability and usability Optional - Incorporation of user-specified constraints

  13. Two Types of Clustering • Partitional algorithms: Construct various partitions and then evaluate them by some criterion • Hierarchical algorithms: Create a hierarchical decomposition of the set of objects using some criterion (focus of this class) Bottom up or top down Top down Partitional Hierarchical

  14. (How-to) Hierarchical Clustering The number of dendrograms with n Bottom-Up (agglomerative): Starting leafs = (2 n -3)!/[(2 ( n -2) ) ( n -2)!] with each item in its own cluster, find the best pair to merge into a new cluster. Number Number of Possible Repeat until all clusters are fused of Leafs Dendrograms 2 1 together. 3 3 4 15 5 105 ... … 10 34,459,425

  15. We begin with a distance matrix which contains the distances between every pair of objects in our database. 0 8 8 7 7 0 2 4 4 0 3 3 D( , ) = 8 0 1 D( , ) = 1 0

  16. Bottom-Up (agglomerative): Starting with each item in its own cluster, find the best pair to merge into a new cluster. Repeat until all clusters are fused together. Consider all Choose … possible the best merges…

  17. Bottom-Up (agglomerative): Starting with each item in its own cluster, find the best pair to merge into a new cluster. Repeat until all clusters are fused together. Consider all Choose possible … the best merges… Consider all Choose … possible the best merges…

  18. Bottom-Up (agglomerative): Starting with each item in its own cluster, find the best pair to merge into a new cluster. Repeat until all clusters are fused together. Consider all Choose possible … the best merges… Consider all Choose possible … the best merges… Consider all Choose … possible the best merges…

  19. Bottom-Up (agglomerative): Starting with each item in its own cluster, find the best pair to merge into a new cluster. Repeat until all clusters are fused together. Consider all Choose possible … the best merges… But how do we compute distances between clusters rather than objects? Consider all Choose possible … the best merges… Consider all Choose … possible the best merges…

  20. Computing distance between clusters: Single Link • cluster distance = distance of two closest members in each class - Potentially long and skinny clusters

  21. Computing distance between clusters: : Complete Link • cluster distance = distance of two farthest members + tight clusters

  22. Computing distance between clusters: Average Link • cluster distance = average distance of all pairs the most widely used measure Robust against noise

  23. Example: single link 1 2 3 4 5   1 0   2 2 0     3 6 3 0   4 10 9 7 0       5 9 8 5 4 0 5 4 3 2 1

  24. Example: single link 1 2 3 4 5 ( 1 , 2 ) 3 4 5   1 0   ( 1 , 2 ) 0     2 2 0   3 3 0     3 6 3 0     4 9 7 0   4 10 9 7 0     5 8 5 4 0     5 9 8 5 4 0 5    d min{ d , d } min{ 6 , 3 } 3 ( 1 , 2 ), 3 1 , 3 2 , 3    4 d min{ d , d } min{ 10 , 9 } 9 ( 1 , 2 ), 4 1 , 4 2 , 4    3 d min{ d , d } min{ 9 , 8 } 8 ( 1 , 2 ), 5 1 , 5 2 , 5 2 1

  25. Example: single link 1 2 3 4 5 ( 1 , 2 ) 3 4 5 ( 1 , 2 , 3 ) 4 5   1 0   ( 1 , 2 ) 0     ( 1 , 2 , 3 ) 0   2 2 0     3 3 0     4 7 0   3 6 3 0     4 9 7 0   5  5 4 0    4 10 9 7 0     5 8 5 4 0     5 9 8 5 4 0 5    d min{ d , d } min{ 9 , 7 } 7 ( 1 , 2 , 3 ), 4 ( 1 , 2 ), 4 3 , 4 4    d min{ d , d } min{ 8 , 5 } 5 ( 1 , 2 , 3 ), 5 ( 1 , 2 ), 5 3 , 5 3 2 1

  26. Example: single link 1 2 3 4 5 ( 1 , 2 ) 3 4 5 ( 1 , 2 , 3 ) 4 5   1 0   ( 1 , 2 ) 0     ( 1 , 2 , 3 ) 0   2 2 0     3 3 0     4 7 0   3 6 3 0     4 9 7 0   5  5 4 0    4 10 9 7 0     5 8 5 4 0     5 9 8 5 4 0 5   d min{ d , d } 5 4 ( 1 , 2 , 3 ), ( 4 , 5 ) ( 1 , 2 , 3 ), 4 ( 1 , 2 , 3 ), 5 3 2 1

  27. Single linkage 7 6 5 4 3 Height represents 2 distance between objects 1 29 2 6 11 9 17 10 13 24 25 26 20 22 30 27 1 3 8 4 12 5 14 23 15 16 18 19 21 28 7 / clusters Average linkage

  28. Summary of Hierarchal Clustering Methods • No need to specify the number of clusters in advance. • Hierarchical structure maps nicely onto human intuition for some domains • They do not scale well: time complexity of at least O( n 2 ), where n is the number of total objects. • Like any heuristic search algorithms, local optima are a problem. • Interpretation of results is (very) subjective.

  29. But what are the clusters? In some cases we can determine the “correct” number of clusters. However, things are rarely this clear cut, unfortunately.

  30. One potential use of a dendrogram is to detect outliers The single isolated branch is suggestive of a data point that is very different to all others Outlier

  31. Example: clustering genes • Microarrays measures the activities of all genes in different conditions • Clustering genes can help determine new functions for unknown genes

  32. Partitional Clustering • Nonhierarchical, each instance is placed in exactly one of K non-overlapping clusters. • Since the output is only one set of clusters the user has to specify the desired number of clusters K.

  33. K-means Clustering: Finished! Re- assign and move centers, until … no objects changed membership. 5 expression in condition 2 4 k 1 3 2 k 2 k 3 1 0 0 1 2 3 4 5

  34. Gaussian mixture clustering

  35. Clustering methods: Comparison Hierarchical K-means GMM naively, O( N 3 ) Running fastest (each fast (each time iteration is iteration is linear) linear) Assumptions requires a strong strongest similarity / assumptions assumptions distance measure Input none K (number of K (number of parameters clusters) clusters) Clusters subjective (only a exactly K exactly K tree is returned) clusters clusters

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend