AMMI Introduction to Deep Learning 8.2. Looking at activations - - PowerPoint PPT Presentation

ammi introduction to deep learning 8 2 looking at
SMART_READER_LITE
LIVE PREVIEW

AMMI Introduction to Deep Learning 8.2. Looking at activations - - PowerPoint PPT Presentation

AMMI Introduction to Deep Learning 8.2. Looking at activations Fran cois Fleuret https://fleuret.org/ammi-2018/ Fri Nov 9 22:39:02 UTC 2018 COLE POLYTECHNIQUE FDRALE DE LAUSANNE Convnet internal layer activations Fran cois


slide-1
SLIDE 1

AMMI – Introduction to Deep Learning 8.2. Looking at activations

Fran¸ cois Fleuret https://fleuret.org/ammi-2018/ Fri Nov 9 22:39:02 UTC 2018

ÉCOLE POLYTECHNIQUE FÉDÉRALE DE LAUSANNE

slide-2
SLIDE 2

Convnet internal layer activations

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 1 / 21

slide-3
SLIDE 3

An alternative approach is to look at the activations themselves. Since the convolutional layers maintain the 2d structure of the signal, the activations can be visualized as images, where the local coding at any location

  • f an activation map is associated to the original content at that same location.

Given the large number of channels, we have to pick a few at random.

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 2 / 21

slide-4
SLIDE 4

An alternative approach is to look at the activations themselves. Since the convolutional layers maintain the 2d structure of the signal, the activations can be visualized as images, where the local coding at any location

  • f an activation map is associated to the original content at that same location.

Given the large number of channels, we have to pick a few at random. Since the representation is distributed across multiple channels, individual channel have usually no clear semantic.

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 2 / 21

slide-5
SLIDE 5

A MNIST character with LeNet (leCun et al., 1998).

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 3 / 21

slide-6
SLIDE 6

An RGB image with AlexNet (Krizhevsky et al., 2012).

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 4 / 21

slide-7
SLIDE 7

An RGB image with AlexNet (Krizhevsky et al., 2012).

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 4 / 21

slide-8
SLIDE 8

An RGB image with AlexNet (Krizhevsky et al., 2012).

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 4 / 21

slide-9
SLIDE 9

An RGB image with AlexNet (Krizhevsky et al., 2012).

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 4 / 21

slide-10
SLIDE 10

An RGB image with AlexNet (Krizhevsky et al., 2012).

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 4 / 21

slide-11
SLIDE 11

ILSVRC12 with ResNet152 (He et al., 2015).

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 5 / 21

slide-12
SLIDE 12

ILSVRC12 with ResNet152 (He et al., 2015).

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 5 / 21

slide-13
SLIDE 13

ILSVRC12 with ResNet152 (He et al., 2015).

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 5 / 21

slide-14
SLIDE 14

Yosinski et al. (2015) developed analysis tools to visit a network and look at the internal activations for a given input signal. This allowed them in particular to find units with a clear semantic in an AlexNet-like network trained on ImageNet.

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 6 / 21

slide-15
SLIDE 15

Figure 2. A view of the 13×13 activations of the 151st channel on the conv5 layer of a deep neural network trained on ImageNet, a dataset that does not contain a face class, but does contain many images with faces. The channel responds to human and animal faces and is robust to changes in scale, pose, lighting, and context, which can be discerned by a user by actively changing the scene in front of a webcam or by loading static images (e.g. of the lions) and seeing the corresponding response of the unit. Photo of lions via Flickr user arnolouise, licensed under CC BY-NC-SA 2.0.

(Yosinski et al., 2015)

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 7 / 21

slide-16
SLIDE 16

Prediction of 2d dynamics with a 18 layer residual network. Gn Sn Rn (Fleuret, 2016)

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 8 / 21

slide-17
SLIDE 17

Sn Gn Rn Ψ(Sn, Gn) (Fleuret, 2016)

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 9 / 21

slide-18
SLIDE 18

1/1024 2/1024 3/1024 511/1024 512/1024 513/1024 514/1024

. . . (Fleuret, 2016)

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 10 / 21

slide-19
SLIDE 19

(Fleuret, 2016)

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 11 / 21

slide-20
SLIDE 20

(Fleuret, 2016)

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 11 / 21

slide-21
SLIDE 21

Layers as embeddings

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 12 / 21

slide-22
SLIDE 22

In the classification case, the network can be seen as a series of processings aiming as disentangling classes to make them easily separable for the final decision. In this perspective, it makes sense to look at how the samples are distributed spatially after each layer.

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 13 / 21

slide-23
SLIDE 23

The main issue to do so is the dimensionality of the signal. If we look at the total number of dimensions in each layer:

  • A MNIST sample in a LeNet goes from 784 to up to 18k dimensions,
  • A ILSVRC12 sample in Resnet152 goes from 150k to up to 800k

dimensions. This require a mean to project a [very] high dimension point cloud into a 2d or 3d “human-brain accessible” representation

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 14 / 21

slide-24
SLIDE 24

We have already seen PCA and k-means as two standard methods for dimension reduction, but they poorly convey the structure of a smooth low-dimension and non-flat manifold.

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 15 / 21

slide-25
SLIDE 25

We have already seen PCA and k-means as two standard methods for dimension reduction, but they poorly convey the structure of a smooth low-dimension and non-flat manifold. It exists a plethora of methods that aim at reflecting in low-dimension the structure of data points in high dimension. A popular one is t-SNE developed by van der Maaten and Hinton (2008).

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 15 / 21

slide-26
SLIDE 26

Given data-points in high dimension 풟 =

  • xn ∈ RD, n = 1, . . . , N
  • the objective of data-visualization is to find a set of corresponding

low-dimension points ℰ =

  • yn ∈ RC , n = 1, . . . , N
  • such that the positions of the ys “reflect” that of the xs.

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 16 / 21

slide-27
SLIDE 27

The t-Distributed Stochastic Neighbor Embedding (t-SNE) proposed by van der Maaten and Hinton (2008) optimizes with SGD the yis so that the distances to close neighbors of each point are preserved.

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 17 / 21

slide-28
SLIDE 28

The t-Distributed Stochastic Neighbor Embedding (t-SNE) proposed by van der Maaten and Hinton (2008) optimizes with SGD the yis so that the distances to close neighbors of each point are preserved. It actually matches for DKL two distance-dependent distributions: Gaussian in the original space, and Student t-distribution in the low-dimension one.

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 17 / 21

slide-29
SLIDE 29

The scikit-learn toolbox http://scikit-learn.org/ is built around SciPy, and provides many machine learning algorithms, in particular embeddings, among which an implementation of t-SNE. The only catch to use it in PyTorch is the conversions to and from numpy arrays.

from sklearn.manifold import TSNE # x is the array of the original high-dimension points x_np = x.numpy() y_np = TSNE(n_components = 2, perplexity = 50).fit_transform(x_np) # y is the array of corresponding low-dimension points y = torch.from_numpy(y_np)

n_components specifies the embedding dimension and perplexity states [crudely] how many points are considered neighbors of each point.

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 18 / 21

slide-30
SLIDE 30

t-SNE unrolling of the swiss roll (with one noise dimension)

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 19 / 21

slide-31
SLIDE 31

t-SNE unrolling of the swiss roll (with one noise dimension)

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 19 / 21

slide-32
SLIDE 32

Input

t-SNE for LeNet on MNIST

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 20 / 21

slide-33
SLIDE 33

Layer #1

t-SNE for LeNet on MNIST

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 20 / 21

slide-34
SLIDE 34

Layer #4

t-SNE for LeNet on MNIST

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 20 / 21

slide-35
SLIDE 35

Layer #7

t-SNE for LeNet on MNIST

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 20 / 21

slide-36
SLIDE 36

Input

t-SNE for an home-baked resnet (no pooling, 66 layers) CIFAR10

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 21 / 21

slide-37
SLIDE 37

Layer #4

t-SNE for an home-baked resnet (no pooling, 66 layers) CIFAR10

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 21 / 21

slide-38
SLIDE 38

Layer #14

t-SNE for an home-baked resnet (no pooling, 66 layers) CIFAR10

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 21 / 21

slide-39
SLIDE 39

Layer #24

t-SNE for an home-baked resnet (no pooling, 66 layers) CIFAR10

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 21 / 21

slide-40
SLIDE 40

Layer #34

t-SNE for an home-baked resnet (no pooling, 66 layers) CIFAR10

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 21 / 21

slide-41
SLIDE 41

Layer #44

t-SNE for an home-baked resnet (no pooling, 66 layers) CIFAR10

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 21 / 21

slide-42
SLIDE 42

Layer #54

t-SNE for an home-baked resnet (no pooling, 66 layers) CIFAR10

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 21 / 21

slide-43
SLIDE 43

Layer #56

t-SNE for an home-baked resnet (no pooling, 66 layers) CIFAR10

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 21 / 21

slide-44
SLIDE 44

Layer #58

t-SNE for an home-baked resnet (no pooling, 66 layers) CIFAR10

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 21 / 21

slide-45
SLIDE 45

Layer #60

t-SNE for an home-baked resnet (no pooling, 66 layers) CIFAR10

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 21 / 21

slide-46
SLIDE 46

Layer #62

t-SNE for an home-baked resnet (no pooling, 66 layers) CIFAR10

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 21 / 21

slide-47
SLIDE 47

Layer #64

t-SNE for an home-baked resnet (no pooling, 66 layers) CIFAR10

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 21 / 21

slide-48
SLIDE 48

Layer #65

t-SNE for an home-baked resnet (no pooling, 66 layers) CIFAR10

Fran¸ cois Fleuret AMMI – Introduction to Deep Learning / 8.2. Looking at activations 21 / 21

slide-49
SLIDE 49

The end

slide-50
SLIDE 50

References

  • F. Fleuret. Predicting the dynamics of 2d objects with a deep residual network. CoRR,

abs/1610.04032, 2016.

  • K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. CoRR,

abs/1512.03385, 2015.

  • A. Krizhevsky, I. Sutskever, and G. Hinton. Imagenet classification with deep convolutional

neural networks. In Neural Information Processing Systems (NIPS), 2012.

  • Y. leCun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to

document recognition. Proceedings of the IEEE, 86(11):2278–2324, 1998.

  • L. van der Maaten and G. Hinton. Visualizing high-dimensional data using t-sne. Journal of

Machine Learning Research (JMLR), 9:2579–2605, 2008.

  • J. Yosinski, J. Clune, A. Nguyen, T. Fuchs, and H. Lipson. Understanding neural networks

through deep visualization. In Deep Learning Workshop, International Conference on Machine Learning (WS/ICML), 2015.