Le Lecture 10 recap ap
- Prof. Leal-Taixé and Prof. Niessner
1
Le Lecture 10 recap ap Prof. Leal-Taix and Prof. Niessner 1 Le - - PowerPoint PPT Presentation
Le Lecture 10 recap ap Prof. Leal-Taix and Prof. Niessner 1 Le LeNet 60k parameters Digit recognition: 10 classes Conv -> Pool -> Conv -> Pool -> Conv -> FC As we go deeper: Width, height Number of filters
1
2
60k parameters
3
[Krizhevsky et al. 2012]
4
[Simonyan and Zisserman 2014]
5
Conv=3x3,s=1,same Maxpool=2x2,s=2
6
138M parameters
harder
7
8
Input Linear Non-linearity
W LxL−1 + bL xL = f(W LxL−1 + bL) xL+1 = f(W L+1xL + bL+1) xL−1 xL+1 xL
9
Linear
xL−1 xL+1 xL
Linear Main path Input Skip connection
10
Linear
xL−1 xL+1 xL
Linear Input
xL+1 = f(W L+1xL + bL+1) xL+1 = f(W L+1xL + bL+1 + xL−1)
dimensions
matrix of learned weights or zero padding
11
xL+1 xL xL−1
improve
12
xL+1 xL xL−1
NN
3 2
3 4 3 2 1
1 3 3 5
1 4 4 5 6 7 9
2 Image 5x5 Kernel 1x1
13
What is the output size?
3 2
3 4 3 2 1
1 3 3 5
1 4 4 5 6 7 9
2 Image 5x5 Kernel 1x1 −5 ∗ 2 = −10
14
3 2
3 4 3 2 1
1 3 3 5
1 4 4 5 6 7 9
2 Image 5x5 Kernel 1x1 −1 ∗ 2 = −2
15
6 4
6 8 6 4 2
2 6 6 10
2 8 8 10 12 14 18
3 2
3 4 3 2 1
1 3 3 5
1 4 4 5 6 7 9
Image 5x5
16
6 4
6 8 6 4 2
2 6 6 10
2 8 8 10 12 14 18
scales the input with a number
17
complex functions
32 32 200 32 32 32 32 Conv 1x1x200 + ReLU
18
32 32 200 92 Conv 5x5 + ReLU Multiplications: 1x1x200x32x32x16 5x5x16x32x32x92 ~ 40 million 32 32 92 16 Conv 1x1 + ReLU 32 32 16
Reduction of multiplications by 1/10
19
20
[Long et al. 15] Fully Convolutional Networks for Semantic Segmetnation (FCN)
21
22
Trained on ImageNet New dataset with C classes TRAIN FROZEN
Donahue 2014, Razavian 2014
segmentation (Fully Convolutional Network)
23
24
Classic Neural Networks for Image Classification
25
Image captioning
26
Language recognition
27
Machine translation
28
Event classification
29
Outputs Inputs Hidden states
30
Outputs Inputs Hidden states The hidden state will have its own internal dynamics More expressive model!
31
[Christopher Olah] Understanding LSTMs
Hidden state input Previous hidden state
32
Hidden state Parameters to be learned
33
Hidden state Note: non-linearities ignored for now Output
34
Hidden state Same parameters for each time step = generalization! Output
35
[Christopher Olah] Understanding LSTMs
Hidden state is the same
36
[Christopher Olah] Understanding LSTMs
37
w1 w2 w3 w4 w1 w2 w3 w4 w1 w2 w3 w4 w1 w2 w3 w4 xt
1
xt
2
xt+2
1
xt+2
2
xt+1
2
xt+1
1
Weights are the same!
38
w1 w2 w3 w4 w1 w2 w3 w4 w1 w2 w3 w4 w1 w2 w3 w4 Chain rule All the way to t=0
Add the derivatives at different times for each weight
39
I mo moved to Germany any … so I speak German an fluently
40
At = θtA0
Same weights are multiplied over and over again
41
What happens to small weights? What happens to large weights? Vanishing gradient Exploding gradient
At = θtA0
42
At = θtA0
Diagonal of this matrix are the eigenvalues Matrix of eigenvectors
43
At = QΛtQ|A0 At = θtA0
44
At = QΛtQ|A0
What happens to eigenvalues with magnitude less than one? What happens to eigenvalues with magnitude larger than one? Vanishing gradient Exploding gradient Gradient clipping
45
At = θtA0
Let us just make a matrix with eigenvalues = 1 Allow the ce cell to maintain its “state”
46
At = θtA0
47
At = θtA0
1
48
Hochreiter and Schmidhuber 1997
49
50
51
52
Sigmoid
53
Decides when to erase the cell state Sigmoid = output between 0 (forget) and 1 (keep)
54
Decides which values will be updated New cell state,
tanh (-1,1)
55
56
Decides which values will be
Output from a tanh (-1,1)
57
gt = Tanh(θxgxt + θhght−1 + bg) ht = ot Tanh(Ct)
58
weights Identity function
1 for important information
59
60
Xu et al. 2015
62
Xu et al. 2015
63
Romera-Paredes et al. 2015
64
65
you did the exercises it will be easier for you to answer them
66
– Basics of ML à from linear classifier to NN – Optimization schemes (not necessary to know all the formulas, but to have a good understanding of the differences between them and their behavior – Backpropagation: concept, math, hint: be fluent at computing backprop by hand – Loss functions and activation functions – CNN: convolution, backprop – RNN, LSTMs
67
July 16 16th
h at 08:
08:00 00
68
69
for a series of Advanced DL lectures on different topics
students
70
Intro to Deep Learning DL for Physics
(Th Thuerey)
DL for Vision
(Ni Niessner, , Le Leal-Ta Taixe)
DL for Medical Applicat.
(Me Menze)
DL in Robotics
(Bä Bäuml)
Machine Learning
(Gü Günnemann)
71
– Advanced architectures, e.g. Siamese neural networks – Variational Autoencoders – Generative models, e.g. GAN, – Multi-dimensional CNN – Bayesian Deep Learning
72
– 2 V + 5 P – Mu Must have attended the Intro to DL – Practical part is a project that will last the whole semester – Please do not sign up unless you are willing to spend a lot of time on the project!
73
– Must have attended the Intro to DL – Common detection and segmentation frameworks (YOLO, Faster-RCNN, Mask-RCNN) – Extension to videos à tracking – One project that will last the whole semester
74
Visual Computing
Dynamic Vision and Learning
Ni Niessner Le Leal-Ta Taixé
75