tutorial on methods for interpreting and understanding
play

Tutorial on Methods for Interpreting and Understanding Deep Neural - PowerPoint PPT Presentation

Tutorial on Methods for Interpreting and Understanding Deep Neural Networks Wojciech Samek Grgoire Montavon Klaus-Robert Mller (Fraunhofer HHI) (TU Berlin) (TU Berlin) 1:30 - 2:00 Part 1: Introduction 2:00 - 3:00 Part 2a: Making Deep


  1. Tutorial on Methods for Interpreting and Understanding Deep Neural Networks Wojciech Samek Grégoire Montavon Klaus-Robert Müller (Fraunhofer HHI) (TU Berlin) (TU Berlin) 1:30 - 2:00 Part 1: Introduction 2:00 - 3:00 Part 2a: Making Deep Neural Networks Transparent 3:00 - 3:30 Break 3:30 - 4:00 Part 2b: Making Deep Neural Networks Transparent 4:00 - 5:00 Part 3: Applications & Discussion ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller

  2. Before we start We thank our collaborators ! Sebastian Lapuschkin Alexander Binder (Fraunhofer HHI) (SUTD) Lecture notes will be online soon at: Please ask questions at any time ! /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller

  3. Tutorial on Methods for Interpreting and Understanding Deep Neural Networks W. Samek, G. Montavon, K.-R. Müller Part 1: Introduction ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller

  4. Recent ML Systems achieve superhuman Performance Deep Net outperforms humans in image classification AlphaGo beats Go DeepStack beats human champ professional poker players Autonomous search-and-rescue drones outperform humans Deep Net beats human at Computer out-plays recognizing traffic signs humans in "doom" IBM's Watson destroys humans in jeopardy /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 4

  5. From Data to Information Huge volumes of data Solve task Interpretable Information extract Computing power Deep Nets / Kernel Machines / … Information (implicit) /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 5

  6. From Data to Information Interpretability AlexNet Clarifai VGG GoogleNet ResNet (16.4%) (11.1%) (7.3%) (6.7%) (3.57%) Performance Interpretable Data Information for human Crucial in many applications (industry, sciences …) /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 6

  7. Interpretable vs. Powerful Models ? Non-linear model Linear model vs. Poor fit, but easily Can be very complex interpretable “global explanation” “individual explanation” /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 7

  8. Interpretable vs. Powerful Models ? Non-linear model Linear model vs. Poor fit, but easily Can be very complex interpretable “global explanation” “individual explanation” 60 million parameters We have techniques to interpret and 650,000 neurons explain such complex models ! /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 8

  9. Interpretable vs. Powerful Models ? train best train interpretable interpret it vs. model model suboptimal or biased due to assumptions (linearity, sparsity …) /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 9

  10. Dimensions of Interpretability Different dimensions prediction prediction of “interpretability” “Explain why a certain pattern x has “Explain why a certain pattern x has been classified in a certain way f(x).” been classified in a certain way f(x).” model model “What would a pattern belonging “What would a pattern belonging to a certain category typically look to a certain category typically look like according to the model.” like according to the model.” data “Which dimensions of the data are most relevant for the task.” /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 10

  11. Why Interpretability ? 1) Verify that classifier works as expected Wrong decisions can be costly and dangerous “Autonomous car crashes, “AI medical diagnosis system because it wrongly recognizes …” misclassifies patient’s disease …” /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 11

  12. Why Interpretability ? 2) Improve classifier Generalization error Generalization error + human experience /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 12

  13. Why Interpretability ? 3) Learn from the learning machine “It's not a human move. I've Old promise: never seen a human play this “Learn about the human brain.” move.” (Fan Hui) /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 13

  14. Why Interpretability ? 4) Interpretability in the sciences Stock market analysis: In medical diagnosis: “Model predicts share value “Model predicts that X will with __% accuracy.” survive with probability __” What to do with this Great !!! information ? /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 14

  15. Why Interpretability ? 4) Interpretability in the sciences Learn about the physical / biological / chemical mechanisms. (e.g. find genes linked to cancer, identify binding sites …) /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 15

  16. Why Interpretability ? 5) Compliance to legislation European Union’s new General “right to explanation” Data Protection Regulation Retain human decision in order to assign responsibility. “With interpretability we can ensure that ML models work in compliance to proposed legislation.” /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 16

  17. Why Interpretability ? Interpretability as a gateway between ML and society • Make complex models acceptable for certain applications. • Retain human decision in order to assign responsibility. • “Right to explanation” Interpretability as powerful engineering tool • Optimize models / architectures • Detect flaws / biases in the data • Gain new insights about the problem • Make sure that ML models behave “correctly” /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 17

  18. Techniques of Interpretation /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 18

  19. Techniques of Interpretation Interpreting models better understand (ensemble) internal representation - find prototypical example of a category - find pattern maximizing activity of a neuron Explaining decisions crucial for many (individual) practical applications - “why” does the model arrive at this particular prediction - verify that model behaves as expected /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 19

  20. Techniques of Interpretation In medical context • Population view (ensemble) • Which symptoms are most common for the disease • Which drugs are most helpful for patients • Patient’s view (individual) • Which particular symptoms does the patient have • Which drugs does he need to take in order to recover Both aspects can be important depending on who you are (FDA, doctor, patient). /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 20

  21. Techniques of Interpretation Interpreting models - find prototypical example of a category - find pattern maximizing activity of a neuron cheeseburger goose car /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 21

  22. Techniques of Interpretation Interpreting models - find prototypical example of a category - find pattern maximizing activity of a neuron cheeseburger goose car simple regularizer (Simonyan et al. 2013) /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 22

  23. Techniques of Interpretation Interpreting models - find prototypical example of a category - find pattern maximizing activity of a neuron cheeseburger goose car complex regularizer (Nguyen et al. 2016) /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 23

  24. Techniques of Interpretation Explaining decisions - “why” does the model arrive at a certain prediction - verify that model behaves as expected /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 24

  25. Techniques of Interpretation Explaining decisions - “why” does the model arrive at a certain prediction - verify that model behaves as expected - Sensitivity Analysis - Layer-wise Relevance Propagation (LRP) /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 25

  26. Techniques of Interpretation Sensitivity Analysis (Simonyan et al. 2014) /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 26

  27. Techniques of Interpretation Layer-wise Relevance Propagation (LRP) (Bach et al. 2015) “every neuron gets it’s share of relevance depending on activation and strength of connection.” Theoretical interpretation Deep Taylor Decomposition (Montavon et al., 2017) /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 27

  28. Techniques of Interpretation Techniques of Interpretation Sensitivity Analysis: LRP / Taylor Decomposition: “what makes this image “what makes this image less / more ‘scooter’ ?” ‘scooter’ at all ?” /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 28

  29. More to come Part 2 Part 2 Part 3 quality of explanations, applications, interpretability in the sciences, discussion /29 ICASSP 2017 Tutorial — W. Samek, G. Montavon & K.-R. Müller 29

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend