CS 559: Machine Learning Fundamentals and Applications 6th Set of Notes
Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215
1
CS 559: Machine Learning Fundamentals and Applications 6 th Set of - - PowerPoint PPT Presentation
1 CS 559: Machine Learning Fundamentals and Applications 6 th Set of Notes Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215 Project Proposal Typical experiments
1
HMMs, etc.
etc.
2
3
4
Pattern Classification, Chapter 3 5
Pattern Classification, Chapter 3 6
Pattern Classification, Chapter 3 7
i into a one dimensional
Pattern Classification, Chapter 3 8
Pattern Classification, Chapter 3 9
1
2
2 1
Pattern Classification, Chapter 3 10
2 1 2 1
2 1
Pattern Classification, Chapter 3 11
2 1
Pattern Classification, Chapter 3 12
2 1
Pattern Classification, Chapter 3 13
2 1
Pattern Classification, Chapter 3 14
Pattern Classification, Chapter 3 15
Pattern Classification, Chapter 3 16
Pattern Classification, Chapter 3 17
Pattern Classification, Chapter 3 18
Pattern Classification, Chapter 3 19
Pattern Classification, Chapter 3 20
W has full rank (the inverse exists), we can convert
Pattern Classification, Chapter 3 21
Pattern Classification, Chapter 3 22
– it has full rank, don’t have to solve for eigenvalues
Pattern Classification, Chapter 3 23
Pattern Classification, Chapter 3 24
Pattern Classification, Chapter 3 25
Pattern Classification, Chapter 3 26
mean of all data mean of class i
c-1 corresponding eigenvectors
Pattern Classification, Chapter 3 27
Pattern Classification, Chapter 3 28
Pattern Classification, Chapter 3 29
30
31
32
1. Assume some functional form for P(Y|X) 2. Estimate parameters of P(Y|X) directly from training data
1. Assume some functional form for P(X|Y), P(X) 2. Estimate parameters of P(X|Y), P(X) directly from training data 3. Use Bayes rule to calculate P(Y|X= xi)
33 Slides by T. Mitchell (CMU)
Slides by S. Srihari (U. Buffalo) 34
– Model class-conditional pdfs and prior probabilities – “Generative” since sampling can generate synthetic data points – Popular models
– Directly estimate posterior probabilities – No attempt to model underlying probability distributions – Focus computational resources on given task– better performance – Popular models
Slides by S. Srihari (U. Buffalo) 35
36 Slides by T. Mitchell (CMU)
– Prior information about the structure of the data is often most naturally specified through a generative model P(X|Y)
more square jaw, etc.
– The generative approach does not directly target the classification model P(Y|X) since the goal of generative training is P(X|Y) – If the data x are complex, finding a suitable generative data model P(X|Y) is a difficult task – Since each generative model is separately trained for each class, there is no competition amongst the models to explain the data – The decision boundary between the classes may have a simple form, even if the data distribution of each class is complex
Barber, Ch. 13 37
Barber, Ch. 13 38
39
40
41
0 = 0
(1)+…+ a
(d) = 0
42
43
tx +
i0
44
45
ij defined by:
ij
ij is given by:
46
Pattern Classification, Chapter 5 47
48