Machine Learning
Linear Classifiers: Expressiveness
1
Linear Classifiers: Expressiveness Machine Learning 1 Lecture - - PowerPoint PPT Presentation
Linear Classifiers: Expressiveness Machine Learning 1 Lecture outline Linear models: Introduction What functions do linear classifiers express? 2 Where are we? Linear models: Introduction What functions do linear classifiers
1
2
3
4
5
x1 x2 x3 y 1 1 1 1 1 1 1 1 1 1 1 1 1
6
x1 x2 x3 y x1 + x2 + x3 – 3 sign
1
1
1 1
1
1 1
1 1
1 1 1 1 1
7
x1 x2 x3 y x1 + x2 + x3 – 3 sign
1
1
1 1
1
1 1
1 1
1 1 1 1 1 Negations are okay too. In general, use 1 − 𝑦 in the linear threshold unit if 𝑦 is negated 𝑧 = 𝑦! ∧ 𝑦" ∧ ¬𝑦# corresponds to 𝑦1 + 𝑦2 + 1 − 𝑦3 ≥ 3
8
x1 x2 x3 y x1 + x2 + x3 – 3 sign
1
1
1 1
1
1 1
1 1
1 1 1 1 1 Exercise: What would the linear threshold function be if the conjunctions here were replaced with disjunctions? Negations are okay too. In general, use 1 − 𝑦 in the linear threshold unit if 𝑦 is negated 𝑧 = 𝑦! ∧ 𝑦" ∧ ¬𝑦# corresponds to 𝑦1 + 𝑦2 + 1 − 𝑦3 ≥ 3
9
x1 x2 x3 y x1 + x2 + x3 – 3 sign
1
1
1 1
1
1 1
1 1
1 1 1 1 1 Exercise: What would the linear threshold function be if the conjunctions here were replaced with disjunctions? Negations are okay too. In general, use 1 − 𝑦 in the linear threshold unit if 𝑦 is negated 𝑧 = 𝑦! ∧ 𝑦" ∧ ¬𝑦# corresponds to 𝑦1 + 𝑦2 + 1 − 𝑦3 ≥ 3 Questions?
10
11
12
Questions?
13
Can’t draw a line to separate the two classes Questions?
(The XOR function)
14
15
x
16
x
17
x x2
18
(-2, 4) x x2
19
x x2
20
21
Training data is almost separable, except for some noise How much noise do we allow for?
22
b +w1 x1 + w2x2=0 Training data is almost separable, except for some noise How much noise do we allow for?
23
24
25
restricting the learner only to hyperplanes that go through the origin May not be expressive enough
26
If b is zero, then we are restricting the learner only to hyperplanes that go through the origin May not be expressive enough
27