Learning From Data Lecture 20 Multilayer Perceptron
Multiple layers Universal Approximation The Neural Network
- M. Magdon-Ismail
CSCI 4100/6100
Learning From Data Lecture 20 Multilayer Perceptron Multiple - - PowerPoint PPT Presentation
Learning From Data Lecture 20 Multilayer Perceptron Multiple layers Universal Approximation The Neural Network M. Magdon-Ismail CSCI 4100/6100 recap: Unsupervised Learning k -Means Clustering Gaussian Mixture Model P ( x ) x Hard
Multiple layers Universal Approximation The Neural Network
CSCI 4100/6100
recap:Unsupervised Learning
c A M L Creator: Malik Magdon-Ismail
Multilayer Perceptron: 2 /18
Bio-inspired Neural Network − →
c A M L Creator: Malik Magdon-Ismail
Multilayer Perceptron: 3 /18
Planes don’t flap wings − →
c A M L Creator: Malik Magdon-Ismail
Multilayer Perceptron: 4 /18
xor − →
c A M L Creator: Malik Magdon-Ismail
Multilayer Perceptron: 5 /18
Decomposing xor − →
+1 +1 −1 −1
x1 x2
1x)
2x)
c A M L Creator: Malik Magdon-Ismail
Multilayer Perceptron: 6 /18
Perceptrons for or and and − →
1 1 1.5
1 1 −1.5
c A M L Creator: Malik Magdon-Ismail
Multilayer Perceptron: 7 /18
Representing f using or and and − →
c A M L Creator: Malik Magdon-Ismail
Multilayer Perceptron: 8 /18
Expand ands − →
c A M L Creator: Malik Magdon-Ismail
Multilayer Perceptron: 9 /18
Expand h1, h2 − →
2x
1x
c A M L Creator: Malik Magdon-Ismail
Multilayer Perceptron: 10 /18
The Multilayer Perceptron − →
wt
2x
1 f
1
1.5 1
1 1
−1.5 1 1 −1 −1 −1.5
x1 x2
wt
1x
w1
sign(wtx) w0 w2
c A M L Creator: Malik Magdon-Ismail
Multilayer Perceptron: 11 /18
Universal approximation − →
c A M L Creator: Malik Magdon-Ismail
Multilayer Perceptron: 12 /18
Circle Example − →
c A M L Creator: Malik Magdon-Ismail
Multilayer Perceptron: 13 /18
Approximation versus generalization − →
More nodes per hidden layer = ⇒ approximation↑ and generalization↓
c A M L Creator: Malik Magdon-Ismail
Multilayer Perceptron: 14 /18
Minimizing Ein − →
c A M L Creator: Malik Magdon-Ismail
Multilayer Perceptron: 15 /18
Neural Network − →
input layer ℓ = 0
. . .
hidden layers 0 < ℓ < L
c A M L Creator: Malik Magdon-Ismail
Multilayer Perceptron: 16 /18
Zooming into a hidden node − →
input layer ℓ = 0
1 1 h(x)
. . .
s θ(s) θ θ θ θ θ θ 1 x1 x2 xd
hidden layers 0 < ℓ < L
+ θ θ
layer (ℓ − 1) layer ℓ layer (ℓ + 1)
s(ℓ) x(ℓ) W(ℓ) W(ℓ+1)
layer ℓ parameters signals in s(ℓ) d(ℓ) dimensional input vector
x(ℓ) d(ℓ) + 1 dimensional output vector weights in W(ℓ) (d(ℓ−1) + 1) × d(ℓ) dimensional matrix weights out W(ℓ+1) (d(ℓ) + 1) × d(ℓ+1) dimensional matrix
layers ℓ = 0, 1, 2, . . . , L layer ℓ has “dimension” d(ℓ) = ⇒ d(ℓ) + 1 nodes W(ℓ) = w(ℓ)
1
w(ℓ)
2
· · · w(ℓ)
d(ℓ)
. . .
c A M L Creator: Malik Magdon-Ismail
Multilayer Perceptron: 17 /18
Neural Network − →
input layer ℓ = 0
1 1 h(x)
. . .
s θ(s) θ θ θ θ θ θ 1 x1 x2 xd
hidden layers 0 < ℓ < L
c A M L Creator: Malik Magdon-Ismail
Multilayer Perceptron: 18 /18