PAC Learning and The VC Dimension Rectangle Game Fix a rectangle - - PowerPoint PPT Presentation

pac learning and the vc dimension rectangle game
SMART_READER_LITE
LIVE PREVIEW

PAC Learning and The VC Dimension Rectangle Game Fix a rectangle - - PowerPoint PPT Presentation

PAC Learning and The VC Dimension Rectangle Game Fix a rectangle (unknown to you): From An Introduction to Computational Learning Theory by Keanrs and Vazirani Rectangle Game Draw points from some fjxed unknown distribution: Rectangle


slide-1
SLIDE 1

PAC Learning and The VC Dimension

slide-2
SLIDE 2

 Fix a rectangle (unknown to you):

Rectangle Game

From An Introduction to Computational Learning Theory by Keanrs and Vazirani

slide-3
SLIDE 3

 Draw points from some fjxed unknown

distribution:

Rectangle Game

slide-4
SLIDE 4

 You are told the points and whether they are

in or out:

Rectangle Game

slide-5
SLIDE 5

 You propose a hypothesis:

Rectangle Game

slide-6
SLIDE 6

 Your hypothesis is tested on points drawn

from the same distribution:

Rectangle Game

slide-7
SLIDE 7

 We want an algorithm that:

  • With high probability will choose a hypothesis that is

approximately correct.

Goal

slide-8
SLIDE 8

 Choose the minimum area rectangle containing

all the positive points:

Minimum Rectangle Learner:

h

slide-9
SLIDE 9

 Derive a PAC bound:  For fjxed:

  • R : Rectangle
  • D : Data Distribution
  • ε : Test Error
  • δ : Probability of failing
  • m : Number of Samples

How Good is this?

h R

slide-10
SLIDE 10

 We want to show that with high probability the

area below measured with respect to D is bounded by ε :

Proof:

h R < ε

slide-11
SLIDE 11

 We want to show that with high probability the

area below measured with respect to D is bounded by ε :

Proof:

h R < ε/4

slide-12
SLIDE 12

 Defjne T to be the region that contains

exactly ε/4 of the mass in D sweeping down from the top of R.

 p(T’) > ε/4 = p(T) IFF

T’ contains T

 T’ contains T IFF

none of our m samples are from T

 What is the probability

that all samples miss T

Proof:

h R < ε/4 T’ T

slide-13
SLIDE 13

 What is the probability that all m samples

miss T:

 What is the probability that

we miss any of the rectangles?

  • Union Bound

Proof:

h R < ε/4 T’ T

slide-14
SLIDE 14

Union Bound

A B

slide-15
SLIDE 15

 What is the probability that all m samples

miss T:

 What is the probability that

we miss any of the rectangles:

  • Union Bound

Proof:

h R T = ε/4

slide-16
SLIDE 16

 Probability that any region has weight greater

than ε/4 after m samples is at most:

 If we fjx m such that:  Than with probability 1- δ

we achieve an error rate of at most ε

Proof:

h R T = ε/4

slide-17
SLIDE 17

 Common Inequality:  We can show:  Obtain a lower bound on the samples:

Extra Inequality

slide-18
SLIDE 18

 Provides a measure of the complexity of a

“hypothesis space” or the “power” of “learning machine”

 Higher VC dimension implies the ability to

represent more complex functions

 The VC dimension is the maximum number

  • f points that can be arranged so that f

shatters them.

 What does it mean to shatter?

VC – Dimension

slide-19
SLIDE 19

 A classifjer f can shatter a set of points if

and only if for all truth assignments to those points f gets zero training error

 Example: f(x,b) = sign(x.x-b)

Defjne: Shattering

slide-20
SLIDE 20

 What is the VC Dimension of the classifjer:

  • f(x,b) = sign(x.x-b)

Example Continued:

slide-21
SLIDE 21

 Conjecture:  Easy Proof (lower Bound):

VC Dimension of 2D Half-Space:

slide-22
SLIDE 22

 Harder Proof (Upper Bound):

VC Dimension of 2D Half-Space:

slide-23
SLIDE 23

 VC Dimension Conjecture:

VC-Dim: Axis Aligned Rectangles

slide-24
SLIDE 24

 VC Dimension Conjecture: 4  Upper bound (more Diffjcult):

VC-Dim: Axis Aligned Rectangles

slide-25
SLIDE 25

 What is the VC Dimension of:

  • f(x,{w,b})=sign( w . x + b )
  • X in R^d

 Proof (lower bound):

  • Pick {x_1, …, x_n} (point) locations:
  • Adversary gives assignments {y_1, …, y_n} and

you choose {w_1, …, w_n} and b:

General Half-Spaces in (d – dim)

slide-26
SLIDE 26

Extra Space:

slide-27
SLIDE 27

 Proof (upper bound): VC-Dim = d+1

  • Observe that the last d+1 points can always be

expressed as:

General Half-Spaces

slide-28
SLIDE 28

 Proof (upper bound):

VC-Dim = d+1

  • Observe that the last d+1 points

can always be expressed as:

slide-29
SLIDE 29

Extra Space