CS 559: Machine Learning Fundamentals and Applications 9th Set of Notes
Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215
1
CS 559: Machine Learning Fundamentals and Applications 9 th Set of - - PowerPoint PPT Presentation
1 CS 559: Machine Learning Fundamentals and Applications 9 th Set of Notes Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215 Overview Logistic Regression Notes
1
2
– Introduction – Linear Discriminant
– Kernel Trick
– Multi-class SVMs
3
4
2)
5
6
7
8
9
10
11
12
13
14
15
16
17
18
2)
19
20
Pattern Classification, Chapter 5 21
22
23
24
25
26
27
28
29
30
(k)
31
32
33
34
35
36
(k)
37
(k)
38
39
40
41
42
43
(1) = [0.25, 0.25, 0.25, 0.25, 0.25]
44
45
46
47
(1)=[0,0.5, 0.5, 0, 0 ],
48
49
50
(1) = [1 1 1]
51
52
53
54
ta(4)=[-1 -5 -6]*[0 1 -4]=19>0
ta(4)=[1 2 1]*[0 1 -4]=-2<0
55
(k) =const
56
57
58
– Choose positive constants b1, b , b2,…, b ,…, bn – Try to find weight vector a a such that atyi = b = bi for all samples yi – If we can find such a vector, then a a is a solution because the bi’s are positive – Consider all the samples (not just the misclassified ones)
59
60
61
62
63
64
Pattern Classification, Chapter 5 65
66
– That is if all elements of vector Ya Ya are positive – where ε may be negative
– If the approximation is not good, εi may be large and negative, for some i, thus bi + + εi will be negative and a is not a separating hyperplane
67
th element of Ya
ta
t (βa) <
68
69
70
944 . 045 . 1 66 . 2 a
71
3 < 0
72
73
74
75
(k)=η(1) (1)/k
(k) converges to the MSE
76
77
78
– http://research.microsoft.com/en- us/um/people/cburges/papers/SVMTutorial.pdf
– http://www.support-vector.net/icml-tutorial.pdf
– http://www.csie.ntu.edu.tw/~cjlin/libsvm/
– http://www.csie.ntu.edu.tw/~cjlin/liblinear/
– http://svmlight.joachims.org/
– https://sites.google.com/site/wujx2001/home/power-mean-svm
79
80
81
82
83
84
85
86
87
88
89
90
91
txi
92
93
94
95
96
97
98
2 )
99
100
101
the margin, independent of the dimensionality
102
txj
103
104
105
i and xj
106
107
txj + c)
108
n k jk ik j i
1
109
n k k k k k
y x y x K
1
2
2
n k k k H
1
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125