1
600.325/425 Declarative Methods - J. Eisner 1
A few methods for learning binary classifiers
600.325/425 Declarative Methods - J. Eisner 2
Fundamental Problem of Machine Learning: It is ill-posed
slide thanks to Tom Dietterich (modified)
600.325/425 Declarative Methods - J. Eisner 3
Learning Appears Impossible
There are 216 = 65536 possible
boolean functions over four input features.
Why? Such a function is defined
by 24 = 16 rows. So its output column has 16 slots for answers. There are 216 ways it could fill those in.
slide thanks to Tom Dietterich (modified) x x x x y
- ?
- ?
- ?
- ?
- ?
- ?
- ?
- ?
- ?
- ?
? ? ? ? ? ? spam detection
600.325/425 Declarative Methods - J. Eisner 4
Learning Appears Impossible
There are 216 = 65536 possible
boolean functions over four input features.
Why? Such a function is defined
by 24 = 16 rows. So its output column has 16 slots for answers. There are 216 ways it could fill those in.
We can’t figure out which one is
correct until we’ve seen every possible input-output pair.
After 7 examples, we still have 9
slots to fill in, or 29 possibilities.
slide thanks to Tom Dietterich (modified) x x x x y
- ?
- ?
- ?
- ?
- ?
? ? ? ? spam detection
600.325/425 Declarative Methods - J. Eisner 5
Solution: Work with a restricted hypothesis space
We need to generalize from our few training examples! Either by applying prior knowledge or by guessing, we
choose a space of hypotheses H that is smaller than the space of all possible Boolean functions:
simple conjunctive rules m-of-n rules linear functions multivariate Gaussian joint probability distributions etc.
slide thanks to Tom Dietterich (modified)
600.325/425 Declarative Methods - J. Eisner 6
Illustration: Simple Conjunctive Rules
There are only 16
simple conjunctions (no negation)
Try them all! But no simple rule
explains our 7 training examples.
The same is true for
simple disjunctions.
slide thanks to Tom Dietterich (modified)