Deep Learning T HEORY , H ISTORY , S TATE OF THE A RT & P - - PowerPoint PPT Presentation

deep learning
SMART_READER_LITE
LIVE PREVIEW

Deep Learning T HEORY , H ISTORY , S TATE OF THE A RT & P - - PowerPoint PPT Presentation

Deep Learning T HEORY , H ISTORY , S TATE OF THE A RT & P RACTICAL T OOLS by Ilya Kuzovkin ilya.kuzovkin@gmail.com Machine Learning Estonia http://neuro.cs.ut.ee 2016 Where it has started How it learns How it evolved What is the state


slide-1
SLIDE 1

by Ilya Kuzovkin ilya.kuzovkin@gmail.com

Machine Learning Estonia 2016

Deep Learning

THEORY, HISTORY, STATE OF THE ART & PRACTICAL TOOLS

http://neuro.cs.ut.ee

slide-2
SLIDE 2
slide-3
SLIDE 3

Where it has started How it evolved What is the state now How can you use it How it learns

slide-4
SLIDE 4

Where it has started

slide-5
SLIDE 5

Where it has started

Artificial Neuron

slide-6
SLIDE 6

Where it has started

Artificial Neuron

McCulloch and Pitts

“A Logical Calculus of the Ideas Immanent in Nervous Activity”

1943

slide-7
SLIDE 7

Where it has started

Artificial Neuron

McCulloch and Pitts

“A Logical Calculus of the Ideas Immanent in Nervous Activity”

1943

slide-8
SLIDE 8

Where it has started

Perceptron

Frank Rosenblatt 1957

slide-9
SLIDE 9

Where it has started

Perceptron

Frank Rosenblatt 1957

slide-10
SLIDE 10

Where it has started

Perceptron

Frank Rosenblatt 1957

slide-11
SLIDE 11

Where it has started

Perceptron

Frank Rosenblatt 1957 “[The Perceptron is] the embryo of an electronic computer that [the Navy] expects will be able to walk, talk, see, write, reproduce itself and be conscious of its existence.” THE NEW YORK TIMES

slide-12
SLIDE 12

Where it has started

Sigmoid & Backpropagation

slide-13
SLIDE 13

Where it has started

Sigmoid & Backpropagation

slide-14
SLIDE 14

Where it has started

Sigmoid & Backpropagation

Measure how small changes in weights affect output Can apply NN to regression

slide-15
SLIDE 15

Where it has started

Sigmoid & Backpropagation

(Werbos) Rumelhart, Hinton, Williams (1974) 1986

“Learning representations by back-propagating errors” (Nature)

Measure how small changes in weights affect output Can apply NN to regression

slide-16
SLIDE 16

Where it has started

Sigmoid & Backpropagation

(Werbos) Rumelhart, Hinton, Williams (1974) 1986

“Learning representations by back-propagating errors” (Nature)

Measure how small changes in weights affect output Multilayer neural networks, etc. Can apply NN to regression

slide-17
SLIDE 17

Where it has started

Why DL revolution did not happen in 1986?

FROM A TALK BY GEOFFREY HINTON

slide-18
SLIDE 18

Where it has started

Why DL revolution did not happen in 1986?

  • Not enough data (datasets were 1000 times

too small)

  • Computers were too slow (1,000,000 times)

FROM A TALK BY GEOFFREY HINTON

slide-19
SLIDE 19

Where it has started

Why DL revolution did not happen in 1986?

  • Not enough data (datasets were 1000 times

too small)

  • Computers were too slow (1,000,000 times)
  • Not enough attention to network initialization
  • Wrong non-linearity

FROM A TALK BY GEOFFREY HINTON

slide-20
SLIDE 20

How it learns

slide-21
SLIDE 21

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

NET OUT

slide-22
SLIDE 22

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

NET OUT

slide-23
SLIDE 23

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Forward Pass — Calculating the total error

NET OUT

slide-24
SLIDE 24

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Forward Pass — Calculating the total error

NET OUT

slide-25
SLIDE 25

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Forward Pass — Calculating the total error

NET OUT

slide-26
SLIDE 26

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Forward Pass — Calculating the total error

NET OUT

slide-27
SLIDE 27

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Forward Pass — Calculating the total error

NET OUT

Repeat for h2 = 0.596, o1 = 0.751, o2 = 0.773

slide-28
SLIDE 28

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Forward Pass — Calculating the total error

NET OUT

We have o1, o2

slide-29
SLIDE 29

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Forward Pass — Calculating the total error

NET OUT

We have o1, o2

slide-30
SLIDE 30

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Forward Pass — Calculating the total error

NET OUT

We have o1, o2

slide-31
SLIDE 31

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Forward Pass — Calculating the total error

NET OUT

We have o1, o2

slide-32
SLIDE 32

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Forward Pass — Calculating the total error

NET OUT

We have o1, o2

slide-33
SLIDE 33

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

slide-34
SLIDE 34

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

slide-35
SLIDE 35

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

slide-36
SLIDE 36

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

slide-37
SLIDE 37

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

slide-38
SLIDE 38

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

slide-39
SLIDE 39

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

slide-40
SLIDE 40

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

slide-41
SLIDE 41

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

slide-42
SLIDE 42

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

slide-43
SLIDE 43

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

slide-44
SLIDE 44

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

slide-45
SLIDE 45

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

slide-46
SLIDE 46

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

slide-47
SLIDE 47

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

slide-48
SLIDE 48

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

slide-49
SLIDE 49

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

slide-50
SLIDE 50

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

Learning rate

slide-51
SLIDE 51

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

  • 1. The Backwards Pass — updating weights

NET OUT

Gradient descent update rule Learning rate

slide-52
SLIDE 52

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

NET OUT

slide-53
SLIDE 53

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

NET OUT

  • Repeat for w6, w7, w8
slide-54
SLIDE 54

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

NET OUT

  • Repeat for w6, w7, w8
  • In analogous way for w1, w2, w3, w4
slide-55
SLIDE 55

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

NET OUT

  • Repeat for w6, w7, w8
  • In analogous way for w1, w2, w3, w4
  • Calculate the total error again:

it was: 0.291027924 0.298371109

slide-56
SLIDE 56

How it learns

Backpropagation

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

Given inputs 0.05 and 0.10, we want the neural network to output 0.01 and 0.99

NET OUT

  • Repeat for w6, w7, w8
  • In analogous way for w1, w2, w3, w4
  • Calculate the total error again:

it was:

  • Repeat 10,000 times:

0.291027924 0.298371109 0.000035085

slide-57
SLIDE 57

How it learns

Optimization methods

Alec Radford “Introduction to Deep Learning with Python”

slide-58
SLIDE 58

How it learns

Optimization methods

Alec Radford “Introduction to Deep Learning with Python”

slide-59
SLIDE 59

How it evolved

slide-60
SLIDE 60

How it evolved

1-layer NN

INPUT OUTPUT

Alec Radford “Introduction to Deep Learning with Python”

slide-61
SLIDE 61

How it evolved

1-layer NN

92.5% on the MNIST test set

INPUT OUTPUT

Alec Radford “Introduction to Deep Learning with Python”

slide-62
SLIDE 62

How it evolved

1-layer NN

92.5% on the MNIST test set

INPUT OUTPUT

Alec Radford “Introduction to Deep Learning with Python”

slide-63
SLIDE 63

How it evolved

One hidden layer

Alec Radford “Introduction to Deep Learning with Python”

slide-64
SLIDE 64

How it evolved

One hidden layer

98.2% on the MNIST test set

Alec Radford “Introduction to Deep Learning with Python”

slide-65
SLIDE 65

How it evolved

One hidden layer

98.2% on the MNIST test set Activity of a 100 hidden neurons (out of 625)

Alec Radford “Introduction to Deep Learning with Python”

slide-66
SLIDE 66

How it evolved

Overfitting

slide-67
SLIDE 67

How it evolved

Dropout

Srivastava, Hinton, Krizhevsky, Sutskever, Salakhutdinov, “Dropout: A Simple Way to Prevent Neural Networks from Overfitting”, 2014

slide-68
SLIDE 68

How it evolved

Dropout

Srivastava, Hinton, Krizhevsky, Sutskever, Salakhutdinov, “Dropout: A Simple Way to Prevent Neural Networks from Overfitting”, 2014

slide-69
SLIDE 69

How it evolved

Dropout

Srivastava, Hinton, Krizhevsky, Sutskever, Salakhutdinov, “Dropout: A Simple Way to Prevent Neural Networks from Overfitting”, 2014

slide-70
SLIDE 70

How it evolved

Dropout

Srivastava, Hinton, Krizhevsky, Sutskever, Salakhutdinov, “Dropout: A Simple Way to Prevent Neural Networks from Overfitting”, 2014

slide-71
SLIDE 71

How it evolved

ReLU

  • X. Glorot, A. Bordes, Y. Bengio, “Deep Sparse Rectifier Neural Networks”, 2011
slide-72
SLIDE 72

How it evolved

ReLU

  • X. Glorot, A. Bordes, Y. Bengio, “Deep Sparse Rectifier Neural Networks”, 2011
slide-73
SLIDE 73

How it evolved

“Modern” ANN

  • Several hidden layers
  • ReLU activation units
  • Dropout
slide-74
SLIDE 74

How it evolved

“Modern” ANN

  • Several hidden layers
  • ReLU activation units
  • Dropout
slide-75
SLIDE 75

How it evolved

“Modern” ANN

  • Several hidden layers
  • ReLU activation units
  • Dropout

99.0% on the MNIST test set

slide-76
SLIDE 76

How it evolved

Convolution

slide-77
SLIDE 77

How it evolved

Convolution

+1 +1 +1

  • 1
  • 1
  • 1

Prewitt edge detector

slide-78
SLIDE 78

How it evolved

Convolution

+1 +1 +1

  • 1
  • 1
  • 1

Prewitt edge detector

slide-79
SLIDE 79

How it evolved

Convolution

+1 +1 +1

  • 1
  • 1
  • 1

Prewitt edge detector

slide-80
SLIDE 80

How it evolved

Convolution

+1 +1 +1

  • 1
  • 1
  • 1

Prewitt edge detector

40 40 40 40 40 40 40 40 40 10 10 10 10 10 10 10 10 10

slide-81
SLIDE 81

How it evolved

Convolution

+1 +1 +1

  • 1
  • 1
  • 1

Prewitt edge detector

+1 +1 +1

  • 1
  • 1
  • 1

40 40 40 40 40 40 40 40 40 10 10 10 10 10 10 10 10 10

slide-82
SLIDE 82

How it evolved

Convolution

+1 +1 +1

  • 1
  • 1
  • 1

Prewitt edge detector

+40 +40 +40

  • 40
  • 40
  • 40

10 10 10 10 10 10 10 10 10

slide-83
SLIDE 83

How it evolved

Convolution

+1 +1 +1

  • 1
  • 1
  • 1

Prewitt edge detector

+40 +40 +40

  • 40
  • 40
  • 40

10 10 10 10 10 10 10 10 10

slide-84
SLIDE 84

How it evolved

Convolution

+1 +1 +1

  • 1
  • 1
  • 1

Prewitt edge detector

90 90

40 40 40 40 40 40 40 40 40 10 10 10 10 10 10 10 10 10

slide-85
SLIDE 85

How it evolved

Convolution

+1 +1 +1

  • 1
  • 1
  • 1

Prewitt edge detector

90 90

40 40 40 40 40 40 40 40 40 10 10 10 10 10 10 10 10 10

slide-86
SLIDE 86

How it evolved

Convolution

+1 +1 +1

  • 1
  • 1
  • 1

Prewitt edge detector

90 90

40 40 40 40 40 40 40 40 40 10 10 10 10 10 10 10 10 10

slide-87
SLIDE 87

How it evolved

Convolution

+1 +1 +1

  • 1
  • 1
  • 1

Prewitt edge detector

90 90

40 40 40 40 40 40 40 40 40 10 10 10 10 10 10 10 10 10

Edge detector is a handcrafted feature detector.

slide-88
SLIDE 88

How it evolved

Convolution

The idea of a convolutional layer is to learn feature detectors instead of using handcrafted ones

slide-89
SLIDE 89

How it evolved

Convolution

The idea of a convolutional layer is to learn feature detectors instead of using handcrafted ones

http://yann.lecun.com/exdb/mnist/

slide-90
SLIDE 90

How it evolved

Convolution

The idea of a convolutional layer is to learn feature detectors instead of using handcrafted ones 99.50% on the MNIST test set

CURRENT BEST: 99.77% by committee of 35 conv. nets

http://yann.lecun.com/exdb/mnist/

slide-91
SLIDE 91

How it evolved

More layers

slide-92
SLIDE 92

How it evolved

More layers

  • C. Szegedy, et al., “Going Deeper with Convolutions”, 2014
slide-93
SLIDE 93

How it evolved

More layers

  • C. Szegedy, et al., “Going Deeper with Convolutions”, 2014

ILSVRC 2015 winner — 152 (!) layers

  • K. He et al., “Deep

Residual Learning for Image Recognition”, 2015

slide-94
SLIDE 94

How it evolved

Hyperparameters

  • Network:
  • architecture
  • number of layers
  • number of units (in each layer)
  • type of the activation function
  • weight initialization
  • Convolutional layers:
  • size
  • stride
  • number of filters
  • Optimization method:
  • learning rate
  • ther method-specific

constants

slide-95
SLIDE 95

How it evolved

Hyperparameters

  • Network:
  • architecture
  • number of layers
  • number of units (in each layer)
  • type of the activation function
  • weight initialization
  • Convolutional layers:
  • size
  • stride
  • number of filters
  • Optimization method:
  • learning rate
  • ther method-specific

constants

Grid search :(

slide-96
SLIDE 96

How it evolved

Hyperparameters

  • Network:
  • architecture
  • number of layers
  • number of units (in each layer)
  • type of the activation function
  • weight initialization
  • Convolutional layers:
  • size
  • stride
  • number of filters
  • Optimization method:
  • learning rate
  • ther method-specific

constants

Grid search :( Random search :/

slide-97
SLIDE 97

How it evolved

Hyperparameters

  • Network:
  • architecture
  • number of layers
  • number of units (in each layer)
  • type of the activation function
  • weight initialization
  • Convolutional layers:
  • size
  • stride
  • number of filters
  • Optimization method:
  • learning rate
  • ther method-specific

constants

Grid search :( Random search :/ Bayesian optimization :)

Snoek, Larochelle, Adams, “Practical Bayesian Optimization of Machine Learning Algorithms”

slide-98
SLIDE 98

How it evolved

Hyperparameters

  • Network:
  • architecture
  • number of layers
  • number of units (in each layer)
  • type of the activation function
  • weight initialization
  • Convolutional layers:
  • size
  • stride
  • number of filters
  • Optimization method:
  • learning rate
  • ther method-specific

constants

Grid search :( Random search :/ Bayesian optimization :) Informal parameter search :)

Snoek, Larochelle, Adams, “Practical Bayesian Optimization of Machine Learning Algorithms”

slide-99
SLIDE 99

How it evolved

Major Types of ANNs

feedforward convolutional

slide-100
SLIDE 100

How it evolved

Major Types of ANNs

recurrent feedforward convolutional

slide-101
SLIDE 101

How it evolved

Major Types of ANNs

recurrent autoencoder feedforward convolutional

slide-102
SLIDE 102

What is the state now

slide-103
SLIDE 103

What is the state now

Computer vision

http://cs.stanford.edu/people/karpathy/deepimagesent/ Kaiming He, et al. “Deep Residual Learning for Image Recognition” 2015

slide-104
SLIDE 104

What is the state now

Natural Language Processing

speech recognition + translation Facebook bAbi dataset: question answering

http://smerity.com/articles/2015/keras_qa.html http://karpathy.github.io/2015/05/21/rnn-effectiveness/

slide-105
SLIDE 105

What is the state now

AI

DeepMind’s DQN

slide-106
SLIDE 106

What is the state now

AI

Sukhbaatar et al. “MazeBase: A Sandbox for Learning from Games”, 2015

DeepMind’s DQN

slide-107
SLIDE 107

What is the state now

Neuroscience

Güclü and Gerven, “Deep Neural Networks Reveal a Gradient in the Complexity of Neural Representations across the Ventral Stream”, 2015

slide-108
SLIDE 108

What is the state now

Neuroscience

Güclü and Gerven, “Deep Neural Networks Reveal a Gradient in the Complexity of Neural Representations across the Ventral Stream”, 2015

slide-109
SLIDE 109

How can you use it

slide-110
SLIDE 110

How can you use it

Pre-trained models

slide-111
SLIDE 111

How can you use it

Pre-trained models

  • Go to https://github.com/BVLC/caffe/wiki/Model-Zoo, pick a model
  • … and use it in your application
slide-112
SLIDE 112

How can you use it

Pre-trained models

  • Go to https://github.com/BVLC/caffe/wiki/Model-Zoo, pick a model
  • … and use it in your application
  • Or …
slide-113
SLIDE 113

How can you use it

Pre-trained models

  • Go to https://github.com/BVLC/caffe/wiki/Model-Zoo, pick a model
  • … and use it in your application
  • Or …
  • … use part of it as the starting point for your model

. . .

slide-114
SLIDE 114

How can you use it

Zoo of Frameworks

Low-level High-level & Wrappers

slide-115
SLIDE 115

How can you use it

Keras

http://keras.io

slide-116
SLIDE 116

How can you use it

Keras

http://keras.io

slide-117
SLIDE 117

How can you use it

Keras

http://keras.io

slide-118
SLIDE 118

How can you use it

Keras

http://keras.io

slide-119
SLIDE 119

How can you use it

Keras

http://keras.io

slide-120
SLIDE 120

How can you use it

Keras

http://keras.io

slide-121
SLIDE 121

How can you use it

Keras

http://keras.io

slide-122
SLIDE 122
  • A Step by Step Backpropagation Example

http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example

  • Online book by By Michael Nielsen

http://neuralnetworksanddeeplearning.com

  • CS231n: Convolutional Neural Networks for Visual Recognition

http://cs231n.stanford.edu/