LRP revisited General Images (Bach 15, Lapuschkin16) Text Analysis - - PowerPoint PPT Presentation

lrp revisited
SMART_READER_LITE
LIVE PREVIEW

LRP revisited General Images (Bach 15, Lapuschkin16) Text Analysis - - PowerPoint PPT Presentation

LRP revisited General Images (Bach 15, Lapuschkin16) Text Analysis (Arras16 &17) Speech (Becker18) Morphing (Seibold18) Games (Lapuschkin18) VQA (Arras18) Video (Anders18) Gait Patterns (Horst18) EEG


slide-1
SLIDE 1
slide-2
SLIDE 2

ICIP’18 Tutorial on Interpretable Deep Learning 2

General Images (Bach’ 15, Lapuschkin’16) Text Analysis (Arras’16 &17) Speech (Becker’18) Games (Lapuschkin’18) EEG (Sturm’16) fMRI (Thomas’18) Morphing (Seibold’18) Video (Anders’18) VQA (Arras’18) Histopathology (Binder’18) Faces (Lapuschkin’17) Gait Patterns (Horst’18) Digits (Bach’ 15)

LRP revisited

slide-3
SLIDE 3

ICIP’18 Tutorial on Interpretable Deep Learning 3

LRP revisited

LSTM (Arras’17, Thomas’18) Convolutional NNs (Bach’15, Arras’17 …) Bag-of-words / Fisher Vector models (Bach’15, Arras’16, Lapuschkin’17, Binder’18) One-class SVM (Kauffmann’18) Local Renormalization Layers (Binder’16)

slide-4
SLIDE 4

MICCAI’18 Tutorial on Interpretable Machine Learning

Application of LRP Compare models

slide-5
SLIDE 5

ICIP’18 Tutorial on Interpretable Deep Learning 5

Application: Compare Classifiers

(Arras et al. 2016 & 2017)

word2vec/CNN: Performance: 80.19% Strategy to solve the problem: identify semantically meaningful words related to the topic. BoW/SVM: Performance: 80.10% Strategy to solve the problem: identify statistical patterns, i.e., use word statistics

slide-6
SLIDE 6

ICIP’18 Tutorial on Interpretable Deep Learning 6

Application: Compare Classifiers

word2vec / CNN model BoW/SVM model

(Arras et al. 2016 & 2017)

Words with maximum relevance

slide-7
SLIDE 7

ICIP’18 Tutorial on Interpretable Deep Learning 7

Visual Object Classes Challenge: 2005 - 2012

LRP in Practice

slide-8
SLIDE 8

ICIP’18 Tutorial on Interpretable Deep Learning 8 (Lapuschkin et al. 2016)

same performance —> same strategy ?

Application: Compare Classifiers

slide-9
SLIDE 9

ICIP’18 Tutorial on Interpretable Deep Learning 9

‘horse’ images in PASCAL VOC 2007

Application: Compare Classifiers

slide-10
SLIDE 10

ICIP’18 Tutorial on Interpretable Deep Learning 10

Application: Compare Classifiers

GoogleNet:

  • 22 Layers
  • ILSRCV: 6.7%
  • Inception layers

BVLC:

  • 8 Layers
  • ILSRCV: 16.4%
slide-11
SLIDE 11

ICIP’18 Tutorial on Interpretable Deep Learning 11

GoogleNet focuses on faces of animal. —> suppresses background noise BVLC CaffeNet heatmaps are much more noisy.

(Binder et al. 2016)

Application: Compare Classifiers

Is it related to the architecture ? Is it related to the performance ?

performance heatmap structure

?

slide-12
SLIDE 12

Application of LRP Quantify Context Use

slide-13
SLIDE 13

ICIP’18 Tutorial on Interpretable Deep Learning 13

Application: Measure Context Use

classifier

how important is context ? how important is context ? relevance outside bbox relevance inside bbox importance

  • f context

=

LRP decomposition allows meaningful pooling over bbox !

slide-14
SLIDE 14

ICIP’18 Tutorial on Interpretable Deep Learning 14

Application: Measure Context Use

(Lapuschkin et al., 2016)

  • BVLC reference model + fine tuning
  • PASCAL VOC 2007
slide-15
SLIDE 15

ICIP’18 Tutorial on Interpretable Deep Learning 15

GoogleNet BVLC CaffeNet

(Lapuschkin et al. 2016)

Application: Measure Context Use

Context use

  • Differen models (BVLC CaffeNet,

GoogleNet, VGG CNN S)

  • ILSVCR 2012

VGG CNN S Context use anti-correlated with performance. BVLC CaffeNet GoogleNet VGG CNN S

slide-16
SLIDE 16

MICCAI’18 Tutorial on Interpretable Machine Learning

Application of LRP Detect Biases & Improve Models

slide-17
SLIDE 17

ICIP’18 Tutorial on Interpretable Deep Learning (Lapuschkin et al., 2017) 17

Application: Face analysis

  • Compare AdienceNet, CaffeNet,

GoogleNet, VGG-16

  • Adience dataset, 26,580 images

Age classification Gender classification

A = AdienceNet C = CaffeNet G = GoogleNet V = VGG-16 [i] = in-place face alignment [r] = rotation based alignment [m] = mixing aligned images for training [n] = initialization on Imagenet [w] = initialization on IMDB-WIKI

slide-18
SLIDE 18

ICIP’18 Tutorial on Interpretable Deep Learning (Lapuschkin et al., 2017) 18

with pretraining without pretraining

Strategy to solve the problem: Focus on chin / beard, eyes & hear, but without pretraining the model overfits

Gender classification

Application: Face analysis

slide-19
SLIDE 19

ICIP’18 Tutorial on Interpretable Deep Learning (Lapuschkin et al., 2017) 19

Application: Face analysis

Age classification

Predictions 25-32 years old 60+ years old pretraining on ImageNet pretraining on IMDB-WIKI

Strategy to solve the problem: Focus on the laughing … laughing speaks against 60+ (i.e., model learned that old people do not laugh)

slide-20
SLIDE 20

ICIP’18 Tutorial on Interpretable Deep Learning (Seibold et al., 2018) 20

Application: Face analysis

real person real person fake person Different training methods

  • 1,900 images of different individuals
  • pretrained VGG19 model
  • different ways to train the models

50% genuine images, 50% complete morphs 50% genuine images, 10% complete morphs and 4 × 10% one region morphed 50% genuine images, 10% complete morphs, partial morphs with 10%

  • ne, two, three and four

region morphed partial morphs with zero,

  • ne, two, three or four

morphed regions, for two class classification last layer reinitialized

slide-21
SLIDE 21

ICIP’18 Tutorial on Interpretable Deep Learning 21

Application: Face analysis

Semantic attack

  • n the model

Black box adversarial attack on the model

slide-22
SLIDE 22

ICIP’18 Tutorial on Interpretable Deep Learning (Seibold et al., 2018) 22

Application: Face analysis

slide-23
SLIDE 23

ICIP’18 Tutorial on Interpretable Deep Learning (Seibold et al., 2018) 23

Application: Face analysis

multiclass

network seems to compare different structures network seems to identify “original” parts

multiclass

Different models have different strategies !

slide-24
SLIDE 24

Application of LRP Learn new Representations

slide-25
SLIDE 25

ICIP’18 Tutorial on Interpretable Deep Learning 25 (Arras et al. 2016 & 2017)

Application: Learn new Representations

… …

document vector

+ + =

word2vec word2vec word2vec relevance relevance relevance

slide-26
SLIDE 26

ICIP’18 Tutorial on Interpretable Deep Learning 26 (Arras et al. 2016 & 2017)

Application: Learn new Representations

uniform TFIDF

2D PCA projection of document vectors Document vector computation is unsupervised (given we have a classifier).

slide-27
SLIDE 27

Application of LRP Interpreting Scientific Data

slide-28
SLIDE 28

ICIP’18 Tutorial on Interpretable Deep Learning 28

CNN DNN

explain

LRP

(Sturm et al. 2016)

Brain-Computer Interfacing

Application: EEG Analysis

Neural network learns that: Left hand movement imagination leads to desynchronization over right sensorimotor cortext (and vice versa).

slide-29
SLIDE 29

ICIP’18 Tutorial on Interpretable Deep Learning 29 (Sturm et al. 2016)

Application: EEG Analysis

Our neural networks are interpretable: We can see for every trial “why” it is classified the way it is.

slide-30
SLIDE 30

30

Difficulty to apply deep learning to fMRI :

  • high dimensional data (100 000 voxels), but only few subjects
  • results must be interpretable (key in neuroscience)

Our approach:

  • Recurrent neural networks

(CNN + LSTM) for whole- brain analysis

  • LRP allows to interpret the

results

Application: fMRI Analysis

(Thomas et al. 2018) Dataset:

  • 100 subjects from Human Connectome Project
  • N-back task (faces, places, tools and body parts)
slide-31
SLIDE 31

31

Application: fMRI Analysis

(Thomas et al. 2018)

slide-32
SLIDE 32

32

Our approach:

  • Classify & explain individual gait

patterns

  • Important for understanding

diseases such as Parkinson

Application: Gait Analysis

(Horst et al. 2018)

slide-33
SLIDE 33

Application of LRP Understand Model & Obtain new Insights

slide-34
SLIDE 34

ICIP’18 Tutorial on Interpretable Deep Learning 34 (Lapuschkin et al. 2016)

  • Fisher Vector / SVM classifier
  • PASCAL VOC 2007

Application: Understand the model

slide-35
SLIDE 35

ICIP’18 Tutorial on Interpretable Deep Learning 35 (Lapuschkin et al. 2016)

Application: Understand the model

slide-36
SLIDE 36

ICIP’18 Tutorial on Interpretable Deep Learning 36

Motion vectors can be extracted from the compressed video

  • > allows very efficient analysis

Application: Understand the model

(Srinivasan et al. 2017)

  • Fisher Vector / SVM classifier
  • Model of Kantorov & Laptev, (CVPR’14)
  • Histogram Of Flow, Motion Boundary Histogram
  • HMDB51 dataset
slide-37
SLIDE 37

ICIP’18 Tutorial on Interpretable Deep Learning (Arras et al., 2017 & 2018)

Negative sentiment movie review: ++, —

37

Application: Understand the model

  • bidirectional LSTM model (Li’16)
  • Stanford Sentiment Treebank dataset

Model understands negation ! How to handle multiplicative interactions ?

gate neuron indirectly affect relevance distribution in forward pass

slide-38
SLIDE 38

ICIP’18 Tutorial on Interpretable Deep Learning 38

Application: Understand the model

(Anders et al., 2018)

  • 3-dimensional CNN (C3D)
  • trained on Sports-1M
  • explain predictions for 1000 videos

from the test set

slide-39
SLIDE 39

ICIP’18 Tutorial on Interpretable Deep Learning 39

Application: Understand the model

(Anders et al., 2018)

slide-40
SLIDE 40

ICIP’18 Tutorial on Interpretable Deep Learning 40

Application: Understand the model

Observation: Explanations focus on the bordering

  • f the video, as if it wants to watch more of it.
slide-41
SLIDE 41

ICIP’18 Tutorial on Interpretable Deep Learning 41

Application: Understand the model

Idea: Play video in fast forward (without retraining) and then the classification accuracy improves.

slide-42
SLIDE 42

ICIP’18 Tutorial on Interpretable Deep Learning (Becker et al., 2018) 42

Application: Understand the model

female speaker male speaker

model classifies gender based on the fundamental frequency and its immediate harmonics (see also Traunmüller & Eriksson 1995)

  • AlexNet model
  • trained on spectrograms
  • spoken digits dataset (AudioMNIST)
slide-43
SLIDE 43

ICIP’18 Tutorial on Interpretable Deep Learning (Arras et al., 2018) 43

Application: Understand the model

  • reimplement model of (Santoro et al.,

2017)

  • test accuracy of 91,0%
  • CLEVR dataset

model understands the question and correctly identifies the object of interest

slide-44
SLIDE 44

ICIP’18 Tutorial on Interpretable Deep Learning 44

Application: Understand the model

(Lapuschkin et al., in prep.)

Sensitivity Analysis LRP does not focus on where the ball is, but on where the ball could be in the next frame LRP shows that that model tracks the ball

slide-45
SLIDE 45

ICIP’18 Tutorial on Interpretable Deep Learning 45

Application: Understand the model

(Lapuschkin et al., in prep.) After 0 epochs After 25 epochs After 195 epochs

slide-46
SLIDE 46

ICIP’18 Tutorial on Interpretable Deep Learning 46

Application: Understand the model

(Lapuschkin et al., in prep.)

slide-47
SLIDE 47

ICIP’18 Tutorial on Interpretable Deep Learning 47

Application: Understand the model

(Lapuschkin et al., in prep.)

model learns

  • 1. track the ball
  • 2. focus on paddle
  • 3. focus on the tunnel
slide-48
SLIDE 48
slide-49
SLIDE 49

ICIP’18 Tutorial on Interpretable Deep Learning 49

Take Home Messages

slide-50
SLIDE 50

ICIP’18 Tutorial on Interpretable Deep Learning 50

Take Home Messages

slide-51
SLIDE 51

ICIP’18 Tutorial on Interpretable Deep Learning 51

Take Home Messages

slide-52
SLIDE 52

ICIP’18 Tutorial on Interpretable Deep Learning 52

Take Home Messages

High flexibility: Different LRP variants, free parameters

slide-53
SLIDE 53

ICIP’18 Tutorial on Interpretable Deep Learning

53

Take Home Messages

slide-54
SLIDE 54

ICIP’18 Tutorial on Interpretable Deep Learning 54

Take Home Messages

slide-55
SLIDE 55

ICIP’18 Tutorial on Interpretable Deep Learning 55

Take Home Messages

slide-56
SLIDE 56

ICIP’18 Tutorial on Interpretable Deep Learning 56

Take Home Messages

slide-57
SLIDE 57

ICIP’18 Tutorial on Interpretable Deep Learning 57

Take Home Messages

slide-58
SLIDE 58

CVPR 2018 Tutorial — W. Samek, G. Montavon & K.-R. Müller

More information

Tutorial Paper Montavon et al., “Methods for interpreting and understanding deep neural networks”, Digital Signal Processing, 73:1-5, 2018 Keras Explanation Toolbox https://github.com/albermax/innvestigate

slide-59
SLIDE 59

ICIP’18 Tutorial on Interpretable Deep Learning

59

References

Tutorial / Overview Papers

G Montavon, W Samek, KR Müller. Methods for Interpreting and Understanding Deep Neural Networks. Digital Signal Processing, 73:1-15, 2018. W Samek, T Wiegand, and KR Müller, Explainable Artificial Intelligence: Understanding, Visualizing and Interpreting Deep Learning Models, ITU Journal: ICT Discoveries - Special Issue 1 - The Impact of Artificial Intelligence (AI) on Communication Networks and Services, 1(1):39-48, 2018.

Methods Papers

S Bach, A Binder, G Montavon, F Klauschen, KR Müller, W Samek. On Pixel-wise Explanations for Non-Linear Classifier Decisions by Layer-wise Relevance Propagation. PLOS ONE, 10(7):e0130140, 2015. G Montavon, S Bach, A Binder, W Samek, KR Müller. Explaining NonLinear Classification Decisions with Deep Taylor Decomposition. Pattern Recognition, 65:211–222, 2017 L Arras, G Montavon, K-R Müller, W Samek. Explaining Recurrent Neural Network Predictions in Sentiment

  • Analysis. EMNLP'17 Workshop on Computational Approaches to Subjectivity, Sentiment & Social Media Analysis

(WASSA), 159-168, 2017. A Binder, G Montavon, S Lapuschkin, KR Müller, W Samek. Layer-wise Relevance Propagation for Neural Networks with Local Renormalization Layers. Artificial Neural Networks and Machine Learning – ICANN 2016, Part II, Lecture Notes in Computer Science, Springer-Verlag, 9887:63-71, 2016. J Kauffmann, KR Müller, G Montavon. Towards Explaining Anomalies: A Deep Taylor Decomposition of One-Class

  • Models. arXiv:1805.06230, 2018.

Evaluation Explanations

W Samek, A Binder, G Montavon, S Lapuschkin, KR Müller. Evaluating the visualization of what a Deep Neural Network has learned. IEEE Transactions on Neural Networks and Learning Systems, 28(11):2660-2673, 2017.

slide-60
SLIDE 60

ICIP’18 Tutorial on Interpretable Deep Learning

60

References

Application to Text

L Arras, F Horn, G Montavon, KR Müller, W Samek. Explaining Predictions of Non-Linear Classifiers in NLP. Workshop on Representation Learning for NLP, Association for Computational Linguistics, 1-7, 2016. L Arras, F Horn, G Montavon, KR Müller, W Samek. "What is Relevant in a Text Document?": An Interpretable Machine Learning Approach. PLOS ONE, 12(8):e0181142, 2017. L Arras, G Montavon, K-R Müller, W Samek. Explaining Recurrent Neural Network Predictions in Sentiment

  • Analysis. EMNLP'17 Workshop on Computational Approaches to Subjectivity, Sentiment & Social Media Analysis

(WASSA), 159-168, 2017. L Arras, A Osman, G Montavon, KR Müller, W Samek. Evaluating and Comparing Recurrent Neural Network Explanation Methods in NLP. arXiv, 2018.

Application to Images & Faces

S Lapuschkin, A Binder, G Montavon, KR Müller, Wojciech Samek. Analyzing Classifiers: Fisher Vectors and Deep Neural Networks. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2912-20, 2016. S Bach, A Binder, KR Müller, W Samek. Controlling Explanatory Heatmap Resolution and Semantics via Decomposition Depth. IEEE International Conference on Image Processing (ICIP), 2271-75, 2016. F Arbabzadeh, G Montavon, KR Müller, W Samek. Identifying Individual Facial Expressions by Deconstructing a Neural Network. Pattern Recognition - 38th German Conference, GCPR 2016, Lecture Notes in Computer Science, 9796:344-54, Springer International Publishing, 2016. S Lapuschkin, A Binder, KR Müller, W Samek. Understanding and Comparing Deep Neural Networks for Age and Gender Classification. IIEEE International Conference on Computer Vision Workshops (ICCVW), 1629-38, 2017. C Seibold, W Samek, A Hilsmann, P Eisert. Accurate and Robust Neural Networks for Security Related Applications Exampled by Face Morphing Attacks. arXiv:1806.04265, 2018.

slide-61
SLIDE 61

ICIP’18 Tutorial on Interpretable Deep Learning

61

Application to Video

C Anders, G Montavon, W Samek, KR Müller. Understanding Patch-Based Learning by Explaining Predictions. arXiv:1806.06926, 2018. V Srinivasan, S Lapuschkin, C Hellge, KR Müller, W Samek. Interpretable Human Action Recognition in Compressed Domain. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 1692- 96, 2017.

Application to Speech

S Becker, M Ackermann, S Lapuschkin, KR Müller, W Samek. Interpreting and Explaining Deep Neural Networks for Classification of Audio Signals. arXiv:1807.03418, 2018.

Application to Sciences

I Sturm, S Lapuschkin, W Samek, KR Müller. Interpretable Deep Neural Networks for Single-Trial EEG

  • Classification. Journal of Neuroscience Methods, 274:141–145, 2016.

A Thomas, H Heekeren, KR Müller, W Samek. Interpretable LSTMs For Whole-Brain Neuroimaging Analyses. arXiv, 2018. KT Schütt, F. Arbabzadah, S Chmiela, KR Müller, A Tkatchenko. Quantum-chemical insights from deep tensor neural networks. Nature communications, 8, 13890, 2017. A Binder, M Bockmayr, M Hägele and others. Towards computational fluorescence microscopy: Machine learning- based integrated prediction of morphological and molecular tumor profiles. arXiv:1805.11178, 2018. F Horst, S Lapuschkin, W Samek, KR Müller, WI Schöllhorn. What is Unique in Individual Gait Patterns? Understanding and Interpreting Deep Learning in Gait Analysis. arXiv:1808.04308, 2018

References

slide-62
SLIDE 62

ICIP’18 Tutorial on Interpretable Deep Learning

62

References

Software

M Alber, S Lapuschkin, P Seegerer, M Hägele, KT Schütt, G Montavon, W Samek, KR Müller, S Dähne, PJ

  • Kindermans. iNNvestigate neural networks!. arXiv:1808.04260, 2018.

S Lapuschkin, A Binder, G Montavon, KR Müller, W Samek. The Layer-wise Relevance Propagation Toolbox for Artificial Neural Networks. Journal of Machine Learning Research, 17(114):1-5, 2016.