Scalable Machine Learning
- 5. (Generalized) Linear Models
Alex Smola Yahoo! Research and ANU
http://alex.smola.org/teaching/berkeley2012 Stat 260 SP 12
Scalable Machine Learning 5. (Generalized) Linear Models Alex Smola - - PowerPoint PPT Presentation
Scalable Machine Learning 5. (Generalized) Linear Models Alex Smola Yahoo! Research and ANU http://alex.smola.org/teaching/berkeley2012 Stat 260 SP 12 Administrative stuff Solutions will be posted by tomorrow New problem set will be
http://alex.smola.org/teaching/berkeley2012 Stat 260 SP 12
(x1, x2) (x1, x2, x1x2)
hx, x0i k(x, x0) := hφ(x), φ(x0)i
k(x, x0) := hx, x0i k(x, x0) := D (x2
1, x2 2,
p 2x1x2), (x0
1 2, x0 2 2,
p 2x0
1x0 2)
E = hx, x0i2 k(x, x0) := hx, x0ip = X
|α|=p
Y
i
αi!(xix0
i)αi with α 2 Nd
k(x, x0) := (hx, x0i + c)p =
p
X
i=0
✓p i ◆ hx, x0ii
k(x, x0) := exp ⇣ γ kx x0k2⌘ k(x, x0) := min(x, x0) for x, x0 ≥ 0 k(A, B) := |A ∩ B|
http://maktoons.blogspot.com/2009/03/support-vector-machine.html
,
w {x | <w x> + b = 0} ,
{x | <w x> + b = −1}
,
{x | <w x> + b = +1}
, x2 x1 Note: <w x1> + b = +1 <w x2> + b = −1 => <w (x1−x2)> = 2 => (x1−x2) = w ||w||
, , , , 2 ||w|| yi = −1 yi = +1
❍ ❍ ❍ ❍ ❍ ◆ ◆ ◆ ◆
minimize
w,b
1 2 kwk2 subject to yi [hw, xii + b] 1
hw, x1i + b = 1 hw, x2i + b = 1 hence hw, x1 x2i = 2 hence ⌧ w kwk, x1 x2
2 kwk
,
w {x | <w x> + b = 0} ,
{x | <w x> + b = −1}
,
{x | <w x> + b = +1}
, x2 x1 Note: <w x1> + b = +1 <w x2> + b = −1 => <w (x1−x2)> = 2 => (x1−x2) = w ||w||
, , , , 2 ||w|| yi = −1 yi = +1
❍ ❍ ❍ ❍ ❍ ◆ ◆ ◆ ◆
minimize
w,b
1 2 kwk2 subject to yi [hw, xii + b] 1
Kij = yiyj hxi, xji w = X
i
αiyixi minimize
α
1 2α>Kα − 1>α subject to X
i
αiyi = 0 αi ≥ 0
,
w {x | <w x> + b = 0} ,
{x | <w x> + b = −1}
,
{x | <w x> + b = +1}
, x2 x1 Note: <w x1> + b = +1 <w x2> + b = −1 => <w (x1−x2)> = 2 => (x1−x2) = w ||w||
, , , , 2 ||w|| yi = −1 yi = +1
❍ ❍ ❍ ❍ ❍ ◆ ◆ ◆ ◆
αi [yi(hxi, wi + b) 1] = 0 yi(hxi, wi + b) > 1 implies αi = 0 αi > 0 implies yi(hxi, wi + b) = 1
Java demo: http://svm.dcs.rhbnc.ac.uk/pagesnew/GPat.shtml
+ +
r ρ
minimize
w,b
1 2 kwk2 subject to yi [hw, xii + b] 1 minimize
w,b
1 2 kwk2 + C X
i
ξi subject to yi [hw, xii + b] 1 ξi and ξi 0 w = 0 and b = 0 and ξi = 1
,
w {x | <w x> + b = 0} ,
{x | <w x> + b = −1}
,
{x | <w x> + b = +1}
, x2 x1 Note: <w x1> + b = +1 <w x2> + b = −1 => <w (x1−x2)> = 2 => (x1−x2) = w ||w||
, , , , 2 ||w|| yi = −1 yi = +1
❍ ❍ ❍ ❍ ❍ ◆ ◆ ◆ ◆
Kij = yiyj hxi, xji w = X
i
αiyixi minimize
α
1 2α>Kα − 1>α subject to X
i
αiyi = 0 αi ∈ [0, C] minimize
w,b
1 2 kwk2 + C X
i
ξi subject to yi [hw, xii + b] 1 ξi and ξi 0
minimize
w,b
1 2 kwk2 + C X
i
ξi subject to yi [hw, xii + b] 1 ξi and ξi 0 minimize
w,b
1 2 kwk2 + C X
i
max [0, 1 yi [hw, xii + b]]
max(0, 1 − yf(x)) {yf(x) < 0}
if f(x) > 1
1 2(1 − f(x))2
if f(x) ∈ [0, 1]
1 2 − f(x)
if f(x) < 0
max(0, 1 − f(x))
(asymptotically) linear (asymptotically) 0
log h 1 + e−f(x)i
R[f] := Ex,y∼p(x,y) [{yf(x) > 0}] Remp[f] := 1 m
m
X
i=1
{yif(xi) > 0} Rreg[f] := 1 m
m
X
i=1
max(0, 1 − yif(xi)) + λΩ[f]
regularization
how to control ƛ
R[f] := Ex,y∼p(x,y) [l(y, f(x))] Remp[f] := 1 m
m
X
i=1
l(yi, f(xi)) Rreg[f] := 1 m
m
X
i=1
l(yi, f(xi)) + λΩ[f]
l(y, f(x)) = 1 2(y − f(x))2
l(y, f(x)) = |y − f(x)|
l(y, f(x)) = max(0, |y − f(x)| − ✏)
minimize
w
1 m
m
X
i=1
(yi hxi, wi)2 + λ 2 kwk2 ∂w [. . .] = 1 m
m
X
i=1
⇥ xix>
i w − xiyi
⇤ + λw = 1 mXX> + λ1
mXy = 0 hence w = ⇥ XX> + λm1 ⇤1 Xy
matrix inverse use CG or SMW
between X matters
x x x x x x x x x x x x x x
+ε −ε
x
ξ +ε −ε ξ
y x y − f(x) loss
minimize
w,b
1 2 kwk2 + C
m
X
i=1
[⇠i + ⇠∗
i ]
subject to hw, xii + b yi + ✏ + ⇠i and ⇠i 0 hw, xii + b yi ✏ ⇠∗
i and ⇠∗ i 0
L =1 2 kwk2 + C
m
X
i=1
[⇠i + ⇠∗
i ] m
X
i=1
[⌘i⇠i + ⌘∗
i ⇠∗ i ] + m
X
i=1
↵i [hw, xii + b yi ✏ ⇠i] +
m
X
i=1
↵∗
i [yi ✏ ⇠∗ i hw, xii b]
∂wL = 0 = w + X
i
[αi − α∗
i ] xi
∂bL = 0 = X
i
[αi − α∗
i ]
∂ξiL = 0 = C − ηi − αi ∂ξ∗
i L = 0 = C − η∗
i − α∗ i
minimize
α,α∗
1 2(↵ − ↵⇤)>K(↵ − ↵⇤) + ✏1>(↵ + ↵⇤) + y>(↵ − ↵⇤) subject to 1>(↵ − ↵⇤) = 0 and ↵i, ↵⇤
i ∈ [0, C]
sinc x + 0.1 sinc x - 0.1 approximation
sinc x + 0.2 sinc x - 0.2 approximation
sinc x + 0.5 sinc x - 0.5 approximation
l(y, f(x)) = (
1 2(y − f(x))2
if |y − f(x)| < 1 |y − f(x)| − 1
2
Data Observations (xi) generated from some P(x), e.g., network usage patterns handwritten digits alarm sensors factory status Task Find unusual events, clean database, dis- tinguish typical ex- amples.
Network Intrusion Detection Detect whether someone is trying to hack the network, downloading tons of MP3s, or doing anything else un- usual on the network. Jet Engine Failure Detection You can’t destroy jet engines just to see how they fail. Database Cleaning We want to find out whether someone stored bogus in- formation in a database (typos, etc.), mislabelled digits, ugly digits, bad photographs in an electronic album. Fraud Detection Credit Cards, Telephone Bills, Medical Records Self calibrating alarm devices Car alarms (adjusts itself to where the car is parked), home alarm (furniture, temperature, windows, etc.)
Key Idea Novel data is one that we don’t see frequently. It must lie in low density regions. Step 1: Estimate density Observations x1, . . . , xm Density estimate via Parzen windows Step 2: Thresholding the density Sort data according to density and use it for rejection Practical implementation: compute p(xi) = 1 m X
j
k(xi, xj) for all i and sort according to magnitude. Pick smallest p(xi) as novel points.
Problems We do not care about estimating the density properly in regions of high density (waste of capacity). We only care about the relative density for threshold- ing purposes. We want to eliminate a certain fraction of observations and tune our estimator specifically for this fraction. Solution Areas of low density can be approximated as the level set of an auxiliary function. No need to estimate p(x) directly — use proxy of p(x). Specifically: find f(x) such that x is novel if f(x) ≤ c where c is some constant, i.e. f(x) describes the amount of novelty.
Maximum a Posteriori minimize
θ m
X
i=1
g(θ) hφ(xi), θi + 1 2σ2kθk2 Advantages Convex optimization problem Concentration of measure Problems Normalization g(θ) may be painful to compute For density estimation we need no normalized p(x|θ) No need to perform particularly well in high density regions
Optimization Problem MAP
m
X
i=1
log p(xi|θ) + 1 2σ2kθk2 Novelty
m
X
i=1
max ✓ log p(xi|θ) exp(ρ g(θ)), 0 ◆ + 1 2kθk2
m
X
i=1
max(ρ hφ(xi), θi, 0) + 1 2kθk2 Advantages No normalization g(θ) needed No need to perform particularly well in high density regions (estimator focuses on low-density regions) Quadratic program
Idea Find hyperplane, given by f(x) = hw, xi + b = 0 that has maximum distance from origin yet is still closer to the origin than the observations. Hard Margin minimize 1 2kwk2 subject to hw, xii 1 Soft Margin minimize 1 2kwk2 + C
m
X
i=1
ξi subject to hw, xii 1 ξi ξi 0
Primal Problem minimize 1 2kwk2 + C
m
X
i=1
ξi subject to hw, xii 1 + ξi 0 and ξi 0 Lagrange Function L Subtract constraints, multiplied by Lagrange multipli- ers (αi and ηi), from Primal Objective Function. Lagrange function L has saddlepoint at optimum. L = 1 2kwk2 + C
m
X
i=1
ξi
m
X
i=1
αi (hw, xii 1 + ξi)
m
X
i=1
ηiξi subject to αi, ηi 0.
Optimality Conditions ∂wL = w
m
X
i=1
αixi = 0 = ) w =
m
X
i=1
αixi ∂ξiL = C αi ηi = 0 = ) αi 2 [0, C] Now substitute the optimality conditions back into L. Dual Problem minimize 1 2
m
X
i=1
αiαjhxi, xji
m
X
i=1
αi subject to αi 2 [0, C] All this is only possible due to the convexity of the primal problem.
x
x x x x x
Problem Depending on C, the number of novel points will vary. We would like to specify the fraction ν beforehand. Solution Use hyperplane separating data from the origin H := {x|hw, xi = ρ} where the threshold ρ is adaptive. Intuition Let the hyperplane shift by shifting ρ Adjust it such that the ’right’ number of observations is considered novel. Do this automatically
Primal Problem minimize 1 2kwk2 +
m
X
i=1
ξi mνρ where hw, xii ρ + ξi 0 ξi 0 Dual Problem minimize 1 2
m
X
i=1
αiαjhxi, xji where αi 2 [0, 1] and
m
X
i=1
αi = νm. Similar to SV classification problem, use standard
minimize
w
1 2 kwk2 +
m
X
i=1
ξi mνρ subject to hw, xii ρ ξi and ξi 0
δ(m− − νm) ≤ 0 δ(m+ − νm) ≥ 0 m− m ≤ ν ≤ m+ m
ν, width c 0.5, 0.5 0.5, 0.5 0.1, 0.5 0.5, 0.1
0.54, 0.43 0.59, 0.47 0.24, 0.03 0.65, 0.38 margin ρ/w 0.84 0.70 0.62 0.48
Better estimates since we only optimize in low density regions. Specifically tuned for small number of outliers. Only estimates of a level-set. For ν = 1 we get the Parzen-windows estimator back.
f(x, y) f(x, y0) 1 for all y0 6= y ∆(y, y0)
l(x, y, f) = sup
y02Y
[f(x, y0) − f(x, y) + ∆(y, y0)] l(x, y, f) = sup
y02Y
[f(x, y0) − f(x, y) + 1] ∆(y, y0) ∆ ✓ y, argmax
y0
f(x, y0) ◆
p(x; µ, Σ) = (2π)
d 2 |Σ|− 1 2 exp
✓ −1 2(x − µ)Σ−1(x − µ) ◆ ˆ Σ = 1 m
m
X
i=1
xix>
i − ˆ
µˆ µ> where ˆ µ = 1 m
m
X
i=1
xi
x = X
i
σiviαi where αi ∼ N(0, 1) gi(x) = hvi, xi
http://www.plantsciences.ucdavis.edu/gepts/pb143/LEC17/pq0921251003.gif
R2 linear PCA R2 H Φ kernel PCA k k(x,x’) = < x,x’> e.g. k(x,x’) = < x,x’>
d
x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x
k k k
Σv = λv 1 m X
i
¯ xi¯ x>
i v = λv for ¯
xi = xi − 1 m X
i
xi hence v = X
j
αj ¯ xj using ¯ x>
l
1 m X
i
¯ xi¯ x>
i v = λ¯
x>
l v
yields 1 m ¯ K ¯ Kα = λ ¯ Kα 1 m ¯ Kα = λα where ¯ Kij = h¯ xi, ¯ xji
−1 1 −0.5 0.5 1 Eigenvalue=0.000 −1 1 −0.5 0.5 1 Eigenvalue=0.291 −1 1 −0.5 0.5 1 Eigenvalue=0.709 −1 1 −0.5 0.5 1 Eigenvalue=0.034 −1 1 −0.5 0.5 1 Eigenvalue=0.345 −1 1 −0.5 0.5 1 Eigenvalue=0.621 −1 1 −0.5 0.5 1 Eigenvalue=0.026 −1 1 −0.5 0.5 1 Eigenvalue=0.395 −1 1 −0.5 0.5 1 Eigenvalue=0.570 −1 1 −0.5 0.5 1 Eigenvalue=0.021 −1 1 −0.5 0.5 1 Eigenvalue=0.418 −1 1 −0.5 0.5 1 Eigenvalue=0.552
Eigenvalue=0.251 Eigenvalue=0.233 Eigenvalue=0.052 Eigenvalue=0.044 Eigenvalue=0.037 Eigenvalue=0.033 Eigenvalue=0.031 Eigenvalue=0.025 Eigenvalue=0.014 Eigenvalue=0.008 Eigenvalue=0.007 Eigenvalue=0.006 Eigenvalue=0.005 Eigenvalue=0.004 Eigenvalue=0.003 Eigenvalue=0.002
+ + +
c- x-c w x c
.
+ + +
c- x-c w x c
.
µ+ = 1 m+ X
i:yi=1
φ(xi) and µ− = 1 m− X
i:yi=−1
φ(xi)
like Watson Nadaraya
f(x) = hµ+ µ−, φ(x)i = X
i
yi myi k(xi, x)
represent a distribution
function complexity
p(weight|height) = p(height, weight) p(height) ∝ p(height, weight)
p(x2|x1) ∝ exp " −1 2 x1 − µ1 x2 − µ2 > Σ11 Σ12 Σ12 Σ22 1 x1 − µ1 x2 − µ2 #
Correlated Observations Assume that the random variables t 2 Rn, t0 2 Rn0 are jointly normal with mean (µ, µ0) and covariance matrix K p(t, t0) / exp 1 2 t µ t0 µ0 > Ktt Ktt0 K>
tt0 Kt0t0
1 t µ t0 µ0 ! . Inference Given t, estimate t0 via p(t0|t). Translation into machine learning language: we learn t0 from t. Practical Solution Since t0|t ⇠ N(˜ µ, ˜ K), we only need to collect all terms in p(t, t0) depending on t0 by matrix inversion, hence ˜ K = Kt0t0 K>
tt0K1 tt Ktt0 and ˜
µ = µ0 + K>
tt0
⇥ K1
tt (t µ)
⇤ | {z }
independent of t0
Key Idea Instead of a fixed set of random variables t, t0 we assume a stochastic process t : X ! R, e.g. X = Rn. Previously we had X = {age, height, weight, . . .}. Definition of a Gaussian Process A stochastic process t : X ! R, where all (t(x1), . . . , t(xm)) are normally distributed. Parameters of a GP Mean µ(x) := E[t(x)] Covariance Function k(x, x0) := Cov(t(x), t(x0)) Simplifying Assumption We assume knowledge of k(x, x0) and set µ = 0.
Covariance Function Function of two arguments Leads to matrix with nonnegative eigenvalues Describes correlation between pairs of observations Kernel Function of two arguments Leads to matrix with nonnegative eigenvalues Similarity measure between pairs of observations Lucky Guess We suspect that kernels and covariance functions are the same . . .
Gaussian Process on Parameters t ⇠ N(µ, K) where Kij = k(xi, xj) Linear Model in Feature Space t(x) = hΦ(x), wi + µ(x) where w ⇠ N(0, 1) The covariance between t(x) and t(x0) is then given by Ew [hΦ(x), wihw, Φ(x0)i] = hΦ(x), Φ(x0)i = k(x, x0) Conclusion A small weight vector in “feature space”, as commonly used in SVM amounts to observing t with high p(t). Log prior log p(t) ( ) Margin kwk2 Will get back to this later again.
˜ K = Kt0t0 − K>
tt0K1 tt Ktt0 and ˜
µ = µ0 + K>
tt0
⇥ K1
tt (t − µ)
⇤
Observation Any function k leading to a symmetric matrix with non- negative eigenvalues is a valid covariance function. Necessary and sufficient condition (Mercer’s Theorem) k needs to be a nonnegative integral kernel. Examples of kernels k(x, x0) Linear hx, x0i Laplacian RBF exp (λkx x0k) Gaussian RBF exp
Polynomial (hx, x0i + ci)d , c 0, d 2 N B-Spline B2n+1(x x0)
Ec[p(x|c)p(x0|c)]
Linear kernel: k(x, x0) = hx, x0i Kernel matrix X>X Mean and covariance ˜ K = X0>X0 X0>X(X>X)1X>X0 = X0>(1 PX)X0. ˜ µ = X0>⇥ X(X>X)1t ⇤ ˜ µ is a linear function of X0. Problem The covariance matrix X>X has at most rank n. After n observations (x 2 Rn) the variance vanishes. This is not realistic. “Flat pancake” or “cigar” distribution.
Indirect Model Instead of observing t(x) we observe y = t(x) + ξ, where ξ is a nuisance term. This yields p(Y |X) = Z
m
Y
i=1
p(yi|ti)p(t|X)dt where we can now find a maximum a posteriori solution for t by maximizing the integrand (we will use this later). Additive Normal Noise If ξ ∼ N(0, σ2) then y is the sum of two Gaussian ran- dom variables. Means and variances add up. y ∼ N(µ, K + σ21).
Covariance Matrices Additive noise K = Kkernel + σ21 Predictive mean and variance ˜ K = Kt0t0 K>
tt0K1 tt Ktt0 and ˜
µ = K>
tt0K1 tt t
Pointwise prediction Ktt = K + σ21 Kt0t0 = k(x, x) + σ2 Ktt0 = (k(x1, x), . . . , k(xm, x)) Plug this into the mean and covariance equations.
p(x; θ) = exp (hφ(x), θi g(θ)) where g(θ) = log X
x0
exp (hφ(x0), θi)
p(x; θ) = exp (hφ(x), θi g(θ)) where g(θ) = log X
x0
exp (hφ(x0), θi) ∂θg(θ) = E [φ(x)] ∂2
θg(θ) = Var [φ(x)]
p(x; θ) = exp (hφ(x), θi g(θ)) where g(θ) = log X
x0
exp (hφ(x0), θi) ∂θg(θ) = E [φ(x)] ∂2
θg(θ) = Var [φ(x)]
p(y|x; θ) = exp (hφ(x, y), θi g(θ|x)) where g(θ|x) = log X
y0
exp (hφ(x, y0), θi) ∂θg(θ|x) = E [φ(x, y)|x] ∂2
θg(θ|x) = Var [φ(x, y)|x]
p(y|x; θ) = exp (hφ(x, y), θi g(θ|x)) where g(θ|x) = log X
y0
exp (hφ(x, y0), θi) ∂θg(θ|x) = E [φ(x, y)|x] ∂2
θg(θ|x) = Var [φ(x, y)|x]
p(y|x; θ) = exp (hφ(x, y), θi g(θ|x)) where g(θ|x) = log X
y0
exp (hφ(x, y0), θi) ∂θg(θ|x) = E [φ(x, y)|x] ∂2
θg(θ|x) = Var [φ(x, y)|x]
p(y|x; θ) = exp (hφ(x, y), θi g(θ|x)) where g(θ|x) = log X
y0
exp (hφ(x, y0), θi) ∂θg(θ|x) = E [φ(x, y)|x] ∂2
θg(θ|x) = Var [φ(x, y)|x]
(Regression is special case where we can integrate)
p(y|x, t(x)) := et(x,y)−g(t(x)) where g(t(x)) = X
y
et(x,y) t ∼ N(µ, K) p(t|X, Y ) ∝ exp X
i
t(xi, yi) − g(t(xi)) − 1 2t>K1t !
− log p(y|t) = log ⇥ et + e−t⇤ − yt = log ⇥ 1 + e−2yt⇤
minimize
t
1 2t>K1t +
m
X
i=1
log ⇥ 1 + eyiti⇤
if f(x) > 1
1 2(1 − f(x))2
if f(x) ∈ [0, 1]
1 2 − f(x)
if f(x) < 0
max(0, 1 − f(x))
(asymptotically) linear (asymptotically) 0
log h 1 + e−f(x)i
Le, Canu, Smola, 2005
Engineer’s favorite p(x) = 1 p 2πσ2 exp ✓ 1 2σ2(x µ)2 ◆ where x 2 R =: X Massaging the math p(x) = exp ⇣ h(x, 0.5x2) | {z }
φ(x)
, θi ⇣ µ2 2σ2 + 1 2 log(2πσ2) ⌘ | {z }
g(θ)
⌘ Using the substitution θ2 := σ2 and θ1 := µσ2 yields g(θ) = 1 2 ⇥ θ2
1θ1 2
+ log 2π log θ2 ⇤
Sufficient Statistic We pick φ(x, y) = (yφ1(x), y2φ2(x)), that is k((x, y), (x0, y0)) = k1(x, x0)yy0+k2(x, x0)y2y02 where y, y0 2 R Hence estimate mean and variance simultaneously. Optimization Problem
minimize
m
X
i=1
2 41 4 " m X
j=1
α1jk1(xi, xj) #> " m X
j=1
α2jk2(xi, xj) #1 " m X
j=1
α1jk1(xi, xj) # 1 2 log det 2 " m X
j=1
α2jk2(xi, xj) #
X
j=1
h y>
i α1jk1(xi, xj) + (y> j α2jyj)k2(xi, xj)
i# + 1 2σ2 X
i,j
α>
1iα1jk1(xi, xj) + tr
h α2iα>
2j
i k2(xi, xj). subject to 0
m
X
i=1
α2ik(xi, xj)
The problem is convex The log-determinant from the normalization of the Gaussian acts as a barrrier function, i.e. a nice SDP .
Newton Method with CG Solver Use Newton method to compute update direction, CG solver instead of inverting Hessian. Lazy Evaluation Never build explicit Hessian. Reduced Rank Use incomplete Cholesky factorization for low-rank ap- proximation. Result m 100 200 500 1k 2k 5k 10k 20k Direct Hessian 8 18 90 607 3551
9 15 38 115 752
7 7 12 30 54 179 368 727 This yields scaling of O(m2.1), O(m1.4), and O(m0.95).
http://books.nips.cc/papers/files/nips24/NIPS2011_1222.pdf
http://users.cecs.anu.edu.au/~chteo/pub/LeSmoChaTeo09.pdf
http://mitpress.mit.edu/catalog/item/default.asp?ttype=2&tid=11755
http://users.cecs.anu.edu.au/~chteo/pub/Chaetal09.pdf
http://ttic.uchicago.edu/~altun/pubs/AltHofTso06.pdf
http://www.seas.upenn.edu/~taskar/pubs/icml05.pdf
http://www.seas.upenn.edu/~taskar/nips07tut/nips07tut.ppt
http://alex.smola.org/papers/2003/SmoSch03b.pdf
http://www.umiacs.umd.edu/~joseph/support-vector- machines4.pdf
http://alex.smola.org/teaching/berkeley2012/slides/ lwk_chapter1.pdf
http://alex.smola.org/teaching/berkeley2012/slides/ se_chapter2.pdf
http://dl.acm.org/citation.cfm?id=295919.295960