Pushpak Bhattacharyya CSE Dept., IIT B IIT Bombay b Lecture 38: - - PowerPoint PPT Presentation

pushpak bhattacharyya cse dept iit b iit bombay b lecture
SMART_READER_LITE
LIVE PREVIEW

Pushpak Bhattacharyya CSE Dept., IIT B IIT Bombay b Lecture 38: - - PowerPoint PPT Presentation

Pushpak Bhattacharyya CSE Dept., IIT B IIT Bombay b Lecture 38: PAC Learning, VC dimension; S lf O Self Organization i ti VC dimension VC-dimension Gives a necessary and sufficient condition for Gives a necessary and sufficient


slide-1
SLIDE 1

Pushpak Bhattacharyya CSE Dept., IIT B b IIT Bombay Lecture 38: PAC Learning, VC dimension; S lf O i ti Self Organization

slide-2
SLIDE 2

VC dimension VC-dimension

Gives a necessary and sufficient condition for Gives a necessary and sufficient condition for PAC learnability.

slide-3
SLIDE 3

Def:- Def: Let C be a concept class, i.e., it has members c1,c2,c3,…… as concepts in it. , , , p C1 C3 C C2 C3

slide-4
SLIDE 4

Let S be a subset of U (universe). Let S be a subset of U (universe). Now if all the subsets of S can be Now if all the subsets of S can be produced by intersecting with Ci

s, then we say

C shatters S.

slide-5
SLIDE 5

The highest cardinality set S that can be The highest cardinality set S that can be shattered gives the VC-dimension of C. VC-dim(C)= |S| VC-dim: Vapnik-Cherronenkis dimension.

slide-6
SLIDE 6

2 – Dim surface C = { half planes} y x

IIT Bombay 6

slide-7
SLIDE 7

S1= { a } y a

1

{ } {a}, Ø x |s| = 1 can be shattered

IIT Bombay 7

slide-8
SLIDE 8

b S2= { a,b } y a b {a,b}, {a} {a}, {b}, Ø x |s| = 2 can be shattered

IIT Bombay 8

slide-9
SLIDE 9

b y S3= { a b a,b,c } c x |s| = 3 can be shattered

IIT Bombay 9

slide-10
SLIDE 10

IIT Bombay 10

slide-11
SLIDE 11

y S4= { a,b,c,d } A B D C D C x |s| = 4 cannot be shattered

IIT Bombay 11

slide-12
SLIDE 12

A Concept Class C is learnable for all A Concept Class C is learnable for all

probability distributions and all concepts in C if and only if the VC dimension of C is finite

If the VC dimension of C is d, then…(next

page)

IIT Bombay 12

slide-13
SLIDE 13

(a) for 0<ε<1 and the sample size at least ( ) p

max[(4/ε)log(2/δ), (8d/ε)log(13/ε)]

any consistent function A:ScC is a learning function for C (b) for 0<ε<1/2 and sample size less than max[((1-ε)/ ε)ln(1/ δ), d(1-2(ε(1- δ)+ δ))] No function A:ScH, for any hypothesis l f f space is a learning function for C.

IIT Bombay 13

slide-14
SLIDE 14

Book

1.

Computational Learning Theory, M. H. G. Anthony, N. Biggs, Cambridge Tracts in h l C S 1997 Theoretical Computer Science, 1997.

Paper’s

1 A theory of the learnable Valiant LG (1984)

  • 1. A theory of the learnable, Valiant, LG (1984),

Communications of the ACM 27(11):1134 -1142.

  • 2. Learnability and the VC-dimension, A Blumer,

A Ehrenfeucht, D Haussler, M Warmuth - Journal f th ACM 1989

  • f the ACM, 1989.
slide-15
SLIDE 15
slide-16
SLIDE 16

Biological Motivation Biological Motivation Brain

slide-17
SLIDE 17

Higher brain Brain Cerebellum Cerebellum Cerebrum 3 Layers: Cerebrum Cerebrum 3- Layers: Cerebrum Cerebellum Higher brain

slide-18
SLIDE 18

Search Search for Meaning Contributing to humanity Achievement,recognition Food,rest survival

slide-19
SLIDE 19

Higher brain ( responsible for higher needs) C b 3 L C b Cerebrum (crucial for survival) 3- Layers: Cerebrum Cerebellum Higher brain

slide-20
SLIDE 20

Back of brain( vision)

Lot of resilience: Lot of resilience: Visual and auditory areas can do each th ’ j b

  • ther’s job

Side areas For auditory information processing For auditory information processing

slide-21
SLIDE 21

Left Brain and Right Brain Left Brain and Right Brain

Dichotomy Left Brain Right Brain

slide-22
SLIDE 22

Left Brain – Logic, Reasoning, Verbal ability Right Brain Emotion Creativity Right Brain – Emotion, Creativity Words left Brain M i Words – left Brain Music Tune – Right Brain g Maps in the brain. Limbs are mapped to brain

slide-23
SLIDE 23

Character Reognition , O/p grid O/p grid . . . . I/p neuron . . . . I/p neuron

slide-24
SLIDE 24

Self Organization or Kohonen network fires a

  • Self Organization or Kohonen network fires a

group of neurons instead of a single one.

  • The group “some how” produces a “picture” of

The group some how produces a picture of the cluster.

  • Fundamentally SOM is competitive learning.
  • But weight changes are incorporated on a

neighborhood. Fi d h i l i h h f

  • Find the winner neuron, apply weight change for

the winner and its “neighbors”.

slide-25
SLIDE 25

Wi Neurons on the contour are the Winner “neighborhood” neurons.

slide-26
SLIDE 26

Weight change rule for SOM Weight change rule for SOM

W( +1) W( ) + ( ) (I( ) W( )) W(n+1) = W(n) + η(n) (I(n) – W(n))

P+δ(n) P+δ(n) P+δ(n)

Neighborhood: function of n Learning rate: function of n

δ( ) i d i f ti f δ(n) is a decreasing function of n η(n) learning rate is also a decreasing function of n 0 < η(n) < η(n –1 ) <=1

slide-27
SLIDE 27

Pictorially Winner δ(n) δ(n)

Convergence for kohonen not proved except for uni- dimension

. . . .

slide-28
SLIDE 28

A … P neurons o/p layer Wp … … . n neurons Clusters: A A : A A : B : C : : : C :