Support Vector Machines and their Applications Purushottam Kar - - PowerPoint PPT Presentation

support vector machines and their applications
SMART_READER_LITE
LIVE PREVIEW

Support Vector Machines and their Applications Purushottam Kar - - PowerPoint PPT Presentation

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications Support Vector Machines and their Applications Purushottam Kar Department of Computer Science and Engineering, Indian Institute of Technology Kanpur.


slide-1
SLIDE 1

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Support Vector Machines and their Applications

Purushottam Kar

Department of Computer Science and Engineering, Indian Institute of Technology Kanpur.

Summer School on “Expert Systems And Their Applications”, Indian Institute of Information Technology Allahabad. June 14, 2009

1 / 24 Support Vector Machines and their Applications

slide-2
SLIDE 2

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Support Vector Machines

What is being “supported” ?

2 / 24 Support Vector Machines and their Applications

slide-3
SLIDE 3

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Support Vector Machines

What is being “supported” ? How can vectors support anything ?

2 / 24 Support Vector Machines and their Applications

slide-4
SLIDE 4

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Support Vector Machines

What is being “supported” ? How can vectors support anything ? Wait !! Machines ?? - Is this a Mechanical Engineering Lecture ?

2 / 24 Support Vector Machines and their Applications

slide-5
SLIDE 5

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The Learning Methodology

Is it possible to write an algorithm to distinguish between ...

3 / 24 Support Vector Machines and their Applications

slide-6
SLIDE 6

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The Learning Methodology

Is it possible to write an algorithm to distinguish between ...

a well-formed and an ill-formed C++ program ?

3 / 24 Support Vector Machines and their Applications

slide-7
SLIDE 7

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The Learning Methodology

Is it possible to write an algorithm to distinguish between ...

a well-formed and an ill-formed C++ program ? a palindrome and a non-palindrome ?

3 / 24 Support Vector Machines and their Applications

slide-8
SLIDE 8

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The Learning Methodology

Is it possible to write an algorithm to distinguish between ...

a well-formed and an ill-formed C++ program ? a palindrome and a non-palindrome ? a graph with and without cliques of size bigger than 1000 ?

3 / 24 Support Vector Machines and their Applications

slide-9
SLIDE 9

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The Learning Methodology

Is it possible to write an algorithm to distinguish between ...

a handwritten 4 and a handwritten 9 ?

4 / 24 Support Vector Machines and their Applications

slide-10
SLIDE 10

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The Learning Methodology

Is it possible to write an algorithm to distinguish between ...

a handwritten 4 and a handwritten 9 ? a spam and a non-spam e-mail ?

4 / 24 Support Vector Machines and their Applications

slide-11
SLIDE 11

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The Learning Methodology

Is it possible to write an algorithm to distinguish between ...

a handwritten 4 and a handwritten 9 ? a spam and a non-spam e-mail ? a positive movie review and a negative movie review ?

4 / 24 Support Vector Machines and their Applications

slide-12
SLIDE 12

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Statistical Machine Learning

“Synthesize” a program based on training data

5 / 24 Support Vector Machines and their Applications

slide-13
SLIDE 13

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Statistical Machine Learning

“Synthesize” a program based on training data Assume training data that is randomly generated from some unknown but fixed distribution and a target function

5 / 24 Support Vector Machines and their Applications

slide-14
SLIDE 14

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Statistical Machine Learning

“Synthesize” a program based on training data Assume training data that is randomly generated from some unknown but fixed distribution and a target function Give probabilistic error bounds

5 / 24 Support Vector Machines and their Applications

slide-15
SLIDE 15

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Statistical Machine Learning

“Synthesize” a program based on training data Assume training data that is randomly generated from some unknown but fixed distribution and a target function Give probabilistic error bounds In other words be probably-approximately-correct

5 / 24 Support Vector Machines and their Applications

slide-16
SLIDE 16

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Statistical Machine Learning

“Synthesize” a program based on training data Assume training data that is randomly generated from some unknown but fixed distribution and a target function Give probabilistic error bounds In other words be probably-approximately-correct The motto - Let the data decide the algorithm

5 / 24 Support Vector Machines and their Applications

slide-17
SLIDE 17

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Expert Systems

... a computing system capable of representing and reasoning about some knowledge rich domain, such as internal medicine or geology ...

Introduction to Expert Systems, Peter Jackson, Addison Wesley Publishing Company, 1986.

6 / 24 Support Vector Machines and their Applications

slide-18
SLIDE 18

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Linear Machines

7 / 24 Support Vector Machines and their Applications

slide-19
SLIDE 19

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Linear Machines

Arguably the simplest of classifiers acting on vectoral data

7 / 24 Support Vector Machines and their Applications

slide-20
SLIDE 20

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Linear Machines

Arguably the simplest of classifiers acting on vectoral data Numerous Learning Algorithms - Perceptron, SVM

7 / 24 Support Vector Machines and their Applications

slide-21
SLIDE 21

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Support Vector Machines

8 / 24 Support Vector Machines and their Applications

slide-22
SLIDE 22

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Support Vector Machines

A “special” hyperplane - with the maximum margin

8 / 24 Support Vector Machines and their Applications

slide-23
SLIDE 23

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Support Vector Machines

A “special” hyperplane - with the maximum margin Margin of a point measures how far is it from the hyperplane

8 / 24 Support Vector Machines and their Applications

slide-24
SLIDE 24

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Learning the Maximum Margin Classifier

minimize

w,b

  • w2

subject to yi( w · xi + b) ≥ 1, i = 1, . . . , l.

A Linearly-constrained Quadratic program

9 / 24 Support Vector Machines and their Applications

slide-25
SLIDE 25

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Learning the Maximum Margin Classifier

minimize

w,b

  • w2

subject to yi( w · xi + b) ≥ 1, i = 1, . . . , l.

A Linearly-constrained Quadratic program Solvable in polynomial time - several algorithms known

9 / 24 Support Vector Machines and their Applications

slide-26
SLIDE 26

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Learning the Maximum Margin Classifier

minimize

w,b

  • w2

subject to yi( w · xi + b) ≥ 1, i = 1, . . . , l.

A Linearly-constrained Quadratic program Solvable in polynomial time - several algorithms known Does not give us much insight into the nature of the hyperplane

9 / 24 Support Vector Machines and their Applications

slide-27
SLIDE 27

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Non-linearly Separable Data

10 / 24 Support Vector Machines and their Applications

slide-28
SLIDE 28

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Non-linearly Separable Data

Use slack variables to allow points to lie on the “wrong” side of the hyperplane

10 / 24 Support Vector Machines and their Applications

slide-29
SLIDE 29

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Non-linearly Separable Data

Use slack variables to allow points to lie on the “wrong” side of the hyperplane Can still be solved using a QCQP

10 / 24 Support Vector Machines and their Applications

slide-30
SLIDE 30

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Learning the Soft Margin Classifier

minimize

w,b

  • w2 + C

l

  • i=1

ξ2

i

subject to yi( w · xi + b) ≥ 1 − ξi, i = 1, . . . , l.

Again a Linearly-constrained Quadratic program

11 / 24 Support Vector Machines and their Applications

slide-31
SLIDE 31

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Learning the Soft Margin Classifier

minimize

w,b

  • w2 + C

l

  • i=1

ξ2

i

subject to yi( w · xi + b) ≥ 1 − ξi, i = 1, . . . , l.

Again a Linearly-constrained Quadratic program More insight gained by looking at the dual program

11 / 24 Support Vector Machines and their Applications

slide-32
SLIDE 32

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The Dual Program for the Hard Margin SVM

maximize

l

  • i=1

αi − 1 2

l

  • i,j=1

yiyjαiαj xi · xj, subject to

l

  • i=1

yiαi = 0, α1 ≥ 0, i = 1, . . . , l. Some properties of the optimum ( w∗, b∗, α∗):

12 / 24 Support Vector Machines and their Applications

slide-33
SLIDE 33

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The Dual Program for the Hard Margin SVM

maximize

l

  • i=1

αi − 1 2

l

  • i,j=1

yiyjαiαj xi · xj, subject to

l

  • i=1

yiαi = 0, α1 ≥ 0, i = 1, . . . , l. Some properties of the optimum ( w∗, b∗, α∗):

α∗

i [yi(

w ∗ · xi + b∗) − 1] = 0, i = 1, . . . , l

12 / 24 Support Vector Machines and their Applications

slide-34
SLIDE 34

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The Dual Program for the Hard Margin SVM

maximize

l

  • i=1

αi − 1 2

l

  • i,j=1

yiyjαiαj xi · xj, subject to

l

  • i=1

yiαi = 0, α1 ≥ 0, i = 1, . . . , l. Some properties of the optimum ( w∗, b∗, α∗):

α∗

i [yi(

w ∗ · xi + b∗) − 1] = 0, i = 1, . . . , l

  • w ∗ = l

i=1 yiα∗ i

xi

12 / 24 Support Vector Machines and their Applications

slide-35
SLIDE 35

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The Dual Program for the Hard Margin SVM

maximize

l

  • i=1

αi − 1 2

l

  • i,j=1

yiyjαiαj xi · xj, subject to

l

  • i=1

yiαi = 0, α1 ≥ 0, i = 1, . . . , l. Some properties of the optimum ( w∗, b∗, α∗):

α∗

i [yi(

w ∗ · xi + b∗) − 1] = 0, i = 1, . . . , l

  • w ∗ = l

i=1 yiα∗ i

xi f ( x, α∗, b∗) = l

i=1 yiα∗ i

xi · x + b∗

12 / 24 Support Vector Machines and their Applications

slide-36
SLIDE 36

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The Support

Consider each vector applying a force of αi on the hyperplane in the direction yi

  • w
  • w2 .

13 / 24 Support Vector Machines and their Applications

slide-37
SLIDE 37

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The Support

Consider each vector applying a force of αi on the hyperplane in the direction yi

  • w
  • w2 .

The conditions exactly correspond to the force and the torque on the hyperplane being zero

13 / 24 Support Vector Machines and their Applications

slide-38
SLIDE 38

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The Support

Consider each vector applying a force of αi on the hyperplane in the direction yi

  • w
  • w2 .

The conditions exactly correspond to the force and the torque on the hyperplane being zero Hence the vectors lying on the margin, for whom αi = 0, “support” the hyperplane.

13 / 24 Support Vector Machines and their Applications

slide-39
SLIDE 39

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Luckiness and Generalization

Essentially the idea is to show that the probability of the training sample misleading us into learning an erroneous classifier is small

14 / 24 Support Vector Machines and their Applications

slide-40
SLIDE 40

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Luckiness and Generalization

Essentially the idea is to show that the probability of the training sample misleading us into learning an erroneous classifier is small To do this model selection has to be done carefully

14 / 24 Support Vector Machines and their Applications

slide-41
SLIDE 41

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Luckiness and Generalization

Essentially the idea is to show that the probability of the training sample misleading us into learning an erroneous classifier is small To do this model selection has to be done carefully The classifier family should not be too powerful to prevent

  • verfitting

14 / 24 Support Vector Machines and their Applications

slide-42
SLIDE 42

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Bounds for Linear Classifiers

Theorem (Vapnik and Chervonenkis) For any probability distribution on the input domain Rd, and any target function g, with probability no less than 1 − δ, any linear hyperplane f classifying a randomly chosen training set of size l perfectly cannot disagree with the target function on more than ǫ fraction of the input domain (with respect to the underlying distribution) where ǫ = 2 l

  • (d + 1) log

2el d + 1 + log 2 δ

  • 15 / 24

Support Vector Machines and their Applications

slide-43
SLIDE 43

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Bounds for Large Margin Classifiers

Theorem (Vapnik) For any probability distribution on the input domain - a ball of radius R, and any target function g, with probability no less than 1 − δ, any linear hyperplane f classifying a randomly chosen training set of size l perfectly with margin ≥ γ cannot disagree with the target function on more than ǫ fraction of the input domain (with respect to the underlying distribution) where ǫ = ˜ O 1 l R2 γ2 + log 1 δ

  • NOTE : Error bound Independent of the dimension !

16 / 24 Support Vector Machines and their Applications

slide-44
SLIDE 44

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The XOR problem

17 / 24 Support Vector Machines and their Applications

slide-45
SLIDE 45

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The XOR problem

The feature map Φ : (x, y) − → (x2, y 2, √ 2xy) makes the problem linearly a separable one.

17 / 24 Support Vector Machines and their Applications

slide-46
SLIDE 46

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The XOR problem

The feature map Φ : (x, y) − → (x2, y 2, √ 2xy) makes the problem linearly a separable one. But do we need to perform the actual feature map ?

17 / 24 Support Vector Machines and their Applications

slide-47
SLIDE 47

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The XOR problem

The feature map Φ : (x, y) − → (x2, y 2, √ 2xy) makes the problem linearly a separable one. But do we need to perform the actual feature map ? No ! Since all that the SVM training algorithm requires are values of Φ( xi) · Φ( xj)

17 / 24 Support Vector Machines and their Applications

slide-48
SLIDE 48

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The XOR problem

The feature map Φ : (x, y) − → (x2, y 2, √ 2xy) makes the problem linearly a separable one. But do we need to perform the actual feature map ? No ! Since all that the SVM training algorithm requires are values of Φ( xi) · Φ( xj) But Φ( xi) · Φ( xj) = xi · xj2

17 / 24 Support Vector Machines and their Applications

slide-49
SLIDE 49

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The Kernel Trick

Many algorithms admit the Kernel trick - SVM, SVM-regression, Kernel-PCA, Kernel-clustering, Perceptron

18 / 24 Support Vector Machines and their Applications

slide-50
SLIDE 50

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The Kernel Trick

Many algorithms admit the Kernel trick - SVM, SVM-regression, Kernel-PCA, Kernel-clustering, Perceptron However kernel trick not used for the Perceptron algorithm - why ?

18 / 24 Support Vector Machines and their Applications

slide-51
SLIDE 51

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

The Kernel Trick

Many algorithms admit the Kernel trick - SVM, SVM-regression, Kernel-PCA, Kernel-clustering, Perceptron However kernel trick not used for the Perceptron algorithm - why ? Note that not all kernels correspond to feature maps

18 / 24 Support Vector Machines and their Applications

slide-52
SLIDE 52

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Implementations

Many techniques known - Ellipsoid method, Gradient descent, Sequential Minimal Optimization - the latter is tuned to the QP problem

19 / 24 Support Vector Machines and their Applications

slide-53
SLIDE 53

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Implementations

Many techniques known - Ellipsoid method, Gradient descent, Sequential Minimal Optimization - the latter is tuned to the QP problem LIBSVM http://www.csie.ntu.edu.tw/~cjlin/libsvm/ - supports multi-class classification and regression

19 / 24 Support Vector Machines and their Applications

slide-54
SLIDE 54

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Implementations

Many techniques known - Ellipsoid method, Gradient descent, Sequential Minimal Optimization - the latter is tuned to the QP problem LIBSVM http://www.csie.ntu.edu.tw/~cjlin/libsvm/ - supports multi-class classification and regression SVMlight http://svmlight.joachims.org/ - good scalability

19 / 24 Support Vector Machines and their Applications

slide-55
SLIDE 55

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Implementations

Many techniques known - Ellipsoid method, Gradient descent, Sequential Minimal Optimization - the latter is tuned to the QP problem LIBSVM http://www.csie.ntu.edu.tw/~cjlin/libsvm/ - supports multi-class classification and regression SVMlight http://svmlight.joachims.org/ - good scalability General Convex Optimization Solvers - CVX, SeDuMi - compatible with Matlab c

  • 19 / 24

Support Vector Machines and their Applications

slide-56
SLIDE 56

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Handwritten Digit Recognition

Boser-Guyon-Vapnik First real-world application of SVMS

20 / 24 Support Vector Machines and their Applications

slide-57
SLIDE 57

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Handwritten Digit Recognition

Boser-Guyon-Vapnik First real-world application of SVMS Two datasets USPS and NIST used for training and testing Performance stable even across a range of kernel choices

20 / 24 Support Vector Machines and their Applications

slide-58
SLIDE 58

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Handwritten Digit Recognition

Boser-Guyon-Vapnik First real-world application of SVMS Two datasets USPS and NIST used for training and testing Performance stable even across a range of kernel choices Even simple polynomial kernels gave improvements (3.2% error)

  • ver MSE techniques like backpropagation or ridge-regression

(12.7% error) on the USPS dataset.

20 / 24 Support Vector Machines and their Applications

slide-59
SLIDE 59

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Text Categorization

Joachims - used insight from information retrieval research

21 / 24 Support Vector Machines and their Applications

slide-60
SLIDE 60

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Text Categorization

Joachims - used insight from information retrieval research Reuters-21578 - a collection of labeled news stories used.

21 / 24 Support Vector Machines and their Applications

slide-61
SLIDE 61

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Text Categorization

Joachims - used insight from information retrieval research Reuters-21578 - a collection of labeled news stories used. Native kernel used.

21 / 24 Support Vector Machines and their Applications

slide-62
SLIDE 62

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Text Categorization

Joachims - used insight from information retrieval research Reuters-21578 - a collection of labeled news stories used. Native kernel used. Performance measured using Precision-Recall break-even point

21 / 24 Support Vector Machines and their Applications

slide-63
SLIDE 63

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Text Categorization

Joachims - used insight from information retrieval research Reuters-21578 - a collection of labeled news stories used. Native kernel used. Performance measured using Precision-Recall break-even point Aggregate performance of Bayes, k-NN, C4.5 around or less than 80% whereas performance of even native kernel > 84%

21 / 24 Support Vector Machines and their Applications

slide-64
SLIDE 64

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Text Categorization

Joachims - used insight from information retrieval research Reuters-21578 - a collection of labeled news stories used. Native kernel used. Performance measured using Precision-Recall break-even point Aggregate performance of Bayes, k-NN, C4.5 around or less than 80% whereas performance of even native kernel > 84% Use of RBF kernels improved performance to > 86%

21 / 24 Support Vector Machines and their Applications

slide-65
SLIDE 65

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Image based Gender Identification

Varma-Babu - application Kernel Learning

22 / 24 Support Vector Machines and their Applications

slide-66
SLIDE 66

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Image based Gender Identification

Varma-Babu - application Kernel Learning Learn the kernel as a linear combination of base kernels

22 / 24 Support Vector Machines and their Applications

slide-67
SLIDE 67

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Image based Gender Identification

Varma-Babu - application Kernel Learning Learn the kernel as a linear combination of base kernels Performance gains of 5-10% observed over other kernel based learning techniques

22 / 24 Support Vector Machines and their Applications

slide-68
SLIDE 68

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Topic Drift in Page-ranking Algorithms

Karnick-Saradhi - application of multi-class Suport vector data description

23 / 24 Support Vector Machines and their Applications

slide-69
SLIDE 69

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Topic Drift in Page-ranking Algorithms

Karnick-Saradhi - application of multi-class Suport vector data description Use SVDD to obtain representative pages for a topic and prune irrelevant pages

23 / 24 Support Vector Machines and their Applications

slide-70
SLIDE 70

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Topic Drift in Page-ranking Algorithms

Karnick-Saradhi - application of multi-class Suport vector data description Use SVDD to obtain representative pages for a topic and prune irrelevant pages Use a kernel based on both link and content information Performance gains in terms of precision and recall observed over

  • ther existing topic distillation techniques

23 / 24 Support Vector Machines and their Applications

slide-71
SLIDE 71

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Bibliography

An Introduction to Support Vector Machines, Nello Cristianini and John Shawe-Taylor, Cambridge University Press, 2000.

24 / 24 Support Vector Machines and their Applications

slide-72
SLIDE 72

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Bibliography

An Introduction to Support Vector Machines, Nello Cristianini and John Shawe-Taylor, Cambridge University Press, 2000. Learning with Kernels, Bernhard Schlkopf and Alexander J. Smola, The MIT Press, 2002.

24 / 24 Support Vector Machines and their Applications

slide-73
SLIDE 73

Introduction SVMs Generalization Bounds The Kernel Trick Implementations Applications

Bibliography

An Introduction to Support Vector Machines, Nello Cristianini and John Shawe-Taylor, Cambridge University Press, 2000. Learning with Kernels, Bernhard Schlkopf and Alexander J. Smola, The MIT Press, 2002. http://www.support-vector.net/

24 / 24 Support Vector Machines and their Applications