CS 730/830: Intro AI Unsuperv. Learning asst 11 posted Wheeler - - PowerPoint PPT Presentation

cs 730 830 intro ai
SMART_READER_LITE
LIVE PREVIEW

CS 730/830: Intro AI Unsuperv. Learning asst 11 posted Wheeler - - PowerPoint PPT Presentation

CS 730/830: Intro AI Unsuperv. Learning asst 11 posted Wheeler Ruml (UNH) Lecture 23, CS 730 1 / 13 Unsuperv. Learning Overview Bottom-Up RANSAC Break k -Means An Algorithm EM Basic Clustering Unsupervised


slide-1
SLIDE 1

CS 730/830: Intro AI

  • Unsuperv. Learning

Wheeler Ruml (UNH) Lecture 23, CS 730 – 1 / 13

asst 11 posted

slide-2
SLIDE 2

Unsupervised Learning

  • Unsuperv. Learning

■ Overview ■ Bottom-Up ■ RANSAC ■ Break ■ k-Means ■ An Algorithm ■ EM ■ Basic Clustering ■ Summary ■ EOLQs

Wheeler Ruml (UNH) Lecture 23, CS 730 – 2 / 13

slide-3
SLIDE 3

Overview

  • Unsuperv. Learning

■ Overview ■ Bottom-Up ■ RANSAC ■ Break ■ k-Means ■ An Algorithm ■ EM ■ Basic Clustering ■ Summary ■ EOLQs

Wheeler Ruml (UNH) Lecture 23, CS 730 – 3 / 13

modeling = predicting = understanding clustering finding ‘structure’ in data

slide-4
SLIDE 4

Bottom-Up Unsupervised Learning

  • Unsuperv. Learning

■ Overview ■ Bottom-Up ■ RANSAC ■ Break ■ k-Means ■ An Algorithm ■ EM ■ Basic Clustering ■ Summary ■ EOLQs

Wheeler Ruml (UNH) Lecture 23, CS 730 – 4 / 13

explain the data all-at-once vs piece-by-piece? repeat make a model to explain a minimal amount of data check how much of the total data the model explains repeat until model fits a decent amount of the data when found, remove explained data from the set until hard to find a decent model or not enough data left

slide-5
SLIDE 5

Random Sample Consensus (RANSAC)

  • Unsuperv. Learning

■ Overview ■ Bottom-Up ■ RANSAC ■ Break ■ k-Means ■ An Algorithm ■ EM ■ Basic Clustering ■ Summary ■ EOLQs

Wheeler Ruml (UNH) Lecture 23, CS 730 – 5 / 13

given data, find a set of explanatory models: repeat repeat many times randomly pick minimum data to fit model find inliers repeat until no change fit model to inliers find new inliers if best model has enough inliers record model remove inliers from data until best model not good enough or not enough data left

slide-6
SLIDE 6

Break

  • Unsuperv. Learning

■ Overview ■ Bottom-Up ■ RANSAC ■ Break ■ k-Means ■ An Algorithm ■ EM ■ Basic Clustering ■ Summary ■ EOLQs

Wheeler Ruml (UNH) Lecture 23, CS 730 – 6 / 13

asst 10

asst 11

projects: four weeks from yesterday!

Tue May 5: 9-noon: project presentations

Mon May 11 2pm: final paper PDF and project tarball via email

slide-7
SLIDE 7

k-Means Clustering

  • Unsuperv. Learning

■ Overview ■ Bottom-Up ■ RANSAC ■ Break ■ k-Means ■ An Algorithm ■ EM ■ Basic Clustering ■ Summary ■ EOLQs

Wheeler Ruml (UNH) Lecture 23, CS 730 – 7 / 13

Naive Bayes model: choose class, generate attributes independently mixture model: choose class, generate data P(x|θ) =

  • k

P(C = k|θk)P(x|C = k, θk) eg, for mixture of Gaussians, P(x|C = k, µk, σ2

k) =

1

  • 2σ2

exp

  • −(x − µk)2

2σ2

k

slide-8
SLIDE 8

An Algorithm

  • Unsuperv. Learning

■ Overview ■ Bottom-Up ■ RANSAC ■ Break ■ k-Means ■ An Algorithm ■ EM ■ Basic Clustering ■ Summary ■ EOLQs

Wheeler Ruml (UNH) Lecture 23, CS 730 – 8 / 13

Means represent the center of a cluster/class Values for the means are the model Model changes based on the classes assigned to the data init the k means somehow repeat until cluster assignments do not change: Assign each data point to the mean nearest to it Calculate new means for the data assigned to each cluster

slide-9
SLIDE 9

An Algorithm

  • Unsuperv. Learning

■ Overview ■ Bottom-Up ■ RANSAC ■ Break ■ k-Means ■ An Algorithm ■ EM ■ Basic Clustering ■ Summary ■ EOLQs

Wheeler Ruml (UNH) Lecture 23, CS 730 – 8 / 13

Means represent the center of a cluster/class Values for the means are the model Model changes based on the classes assigned to the data init the k means somehow repeat until cluster assignments do not change: Assign each data point to the mean nearest to it Calculate new means for the data assigned to each cluster Example

slide-10
SLIDE 10

An Algorithm

  • Unsuperv. Learning

■ Overview ■ Bottom-Up ■ RANSAC ■ Break ■ k-Means ■ An Algorithm ■ EM ■ Basic Clustering ■ Summary ■ EOLQs

Wheeler Ruml (UNH) Lecture 23, CS 730 – 8 / 13

Means represent the center of a cluster/class Values for the means are the model Model changes based on the classes assigned to the data init the k means somehow repeat until cluster assignments do not change: Assign each data point to the mean nearest to it Calculate new means for the data assigned to each cluster Example Is the classification optimal? What is it optimizing?

slide-11
SLIDE 11

Expectation-Maximization

  • Unsuperv. Learning

■ Overview ■ Bottom-Up ■ RANSAC ■ Break ■ k-Means ■ An Algorithm ■ EM ■ Basic Clustering ■ Summary ■ EOLQs

Wheeler Ruml (UNH) Lecture 23, CS 730 – 9 / 13

model parameters θ (eg, µ, σ2, P(C = k))

  • bserved variables xj

hidden variables Cj init the θk somehow repeat until done: E: compute expected values of hidden vars: P(Cj = k|xj, θk) eg by αP(C = k)P(xj|C = k, θk) M: maximize data likelihood using current estimates: θk, with each xj weighted by P(Cj = k|xj), eg by θ ← argmax

θ

  • z

P(Z = z|x, θ)P(x, Z = z|θ)

slide-12
SLIDE 12

Expectation-Maximization

  • Unsuperv. Learning

■ Overview ■ Bottom-Up ■ RANSAC ■ Break ■ k-Means ■ An Algorithm ■ EM ■ Basic Clustering ■ Summary ■ EOLQs

Wheeler Ruml (UNH) Lecture 23, CS 730 – 9 / 13

model parameters θ (eg, µ, σ2, P(C = k))

  • bserved variables xj

hidden variables Cj init the θk somehow repeat until done: E: compute expected values of hidden vars: P(Cj = k|xj, θk) eg by αP(C = k)P(xj|C = k, θk) M: maximize data likelihood using current estimates: θk, with each xj weighted by P(Cj = k|xj), eg by θ ← argmax

θ

  • z

P(Z = z|x, θ)P(x, Z = z|θ) greedy increase of data likelihood

slide-13
SLIDE 13

Expectation-Maximization

  • Unsuperv. Learning

■ Overview ■ Bottom-Up ■ RANSAC ■ Break ■ k-Means ■ An Algorithm ■ EM ■ Basic Clustering ■ Summary ■ EOLQs

Wheeler Ruml (UNH) Lecture 23, CS 730 – 10 / 13

Features

Probabilistic clustering

Explicit model

Locally optimal Issues

Number of classes (means, Gaussians, etc.)

Local maxima

slide-14
SLIDE 14

Agglomerative Clustering

  • Unsuperv. Learning

■ Overview ■ Bottom-Up ■ RANSAC ■ Break ■ k-Means ■ An Algorithm ■ EM ■ Basic Clustering ■ Summary ■ EOLQs

Wheeler Ruml (UNH) Lecture 23, CS 730 – 11 / 13

dendrogram O(n2) vs O(kn) AutoClass

slide-15
SLIDE 15

Summary

  • Unsuperv. Learning

■ Overview ■ Bottom-Up ■ RANSAC ■ Break ■ k-Means ■ An Algorithm ■ EM ■ Basic Clustering ■ Summary ■ EOLQs

Wheeler Ruml (UNH) Lecture 23, CS 730 – 12 / 13

supervised learning: learning a function or a density unsupervised learning: explaining data reinforcement learning: learning how to act

slide-16
SLIDE 16

EOLQs

  • Unsuperv. Learning

■ Overview ■ Bottom-Up ■ RANSAC ■ Break ■ k-Means ■ An Algorithm ■ EM ■ Basic Clustering ■ Summary ■ EOLQs

Wheeler Ruml (UNH) Lecture 23, CS 730 – 13 / 13

What question didn’t you get to ask today?

What’s still confusing?

What would you like to hear more about? Please write down your most pressing question about AI and put it in the box on your way out. Thanks!