Part 3: Best Practices, Pitfalls & Tricks
xkcd
Best Practices, Pitfalls & Tricks An Inconvenient Truth Deep - - PowerPoint PPT Presentation
xkcd Part 3: Best Practices, Pitfalls & Tricks An Inconvenient Truth Deep neural networks comprise millions of parameters: we dont know what these parameters mean Most of the time, we dont know what a NN learns NN are not
xkcd
what these parameters mean
including observational/operator bias (NN are not unbiased!)
then all red objects must be Ferraris!
Valentine & Trampert (2012)
Valentine & Trampert (2012)
Prediction: noise Prediction: noise Prediction: noise Prediction: noise Prediction: noise Prediction: noise 99.9% accuracy!
Pressure Depth Pressure Depth Good generalisation Overfitting
monotonic response
network confidence increases, whereas it should decrease!
earthquakes based on small ones
https://openai.com/blog/adversarial-example-research/
Solut lutio ion: sample from random distribution with variance inversely proportional to layer input. This depends on the activation function!
Parameter value Loss Low learning rate Parameter value Loss High learning rate
for real-world application
components fall into place
fundamentally wrong (e.g. initialisation)
around 1 (see basic MNIST tutorial)
increases, the network is overfitting
comparison between different “experiments” (architectures, hyperparameters, etc.)
normalisation, dropout, noise layers (not covered today)
a decent GPU