Chapter 10: Artificial Neural Networks
- Dr. Xudong Liu
Assistant Professor School of Computing University of North Florida Monday, 9/30/2019
1 / 17
Chapter 10: Artificial Neural Networks Dr. Xudong Liu Assistant - - PowerPoint PPT Presentation
Chapter 10: Artificial Neural Networks Dr. Xudong Liu Assistant Professor School of Computing University of North Florida Monday, 9/30/2019 1 / 17 Overview 1 Artificial neuron: linear threshold unit (LTU) 2 Perceptron 3 Multi-Layer Perceptron
1 / 17
1 Artificial neuron: linear threshold unit (LTU) 2 Perceptron 3 Multi-Layer Perceptron (MLP), Deep Neural Networks (DNN) 4 Learning ANN’s: Error Backpropagation Overview 2 / 17
dx = 2x + 2.
∂x = 2x + y.
dx = dz dy · dy dx = f ′(y) · g ′(x).
Some Preliminaries 3 / 17
Artificial Neuron 4 / 17
Artificial Neuron 5 / 17
Perceptrons 6 / 17
Perceptrons 7 / 17
Perceptrons 8 / 17
Multi-Layer Perceptrons 9 / 17
Learning MLP 10 / 17
Learning MLP 11 / 17
Learning MLP 12 / 17
Learning MLP 13 / 17
Learning MLP 14 / 17
1 Initialize all the weights and biases in the MLP using random values
2 Repeat the following until some stopping criteron: 1
1 Calculate ˆ y; 2 Calculate gradient descents ∆whj and ∆θj for the neurons in the output layer; 3 Calculate gradient descents ∆vih and ∆γh for the neurons in the hidden layer; 4 Update weights whj and vih, and biases θj and γh; Learning MLP 15 / 17
1 ∆whj = −η ∂E
2
3 We know
4 We know ∂E
5 We also know ∂ˆ
6 Bullets 3, 4 and 5 together can solve ∆whj in bullet 1. 7 Computing ∆θj is similar. Learning MLP 16 / 17
1 ∆vih = −η ∂E
2
3 (Long story short) 4 ∆vih = ηxibh(1 − bh)
5 So we update ∆whj and ∆θj for the output layer first, then ∆vih and
6 This is why it is called backpropagation. Learning MLP 17 / 17