lecture 14 lecture 14 29 feb 2016 29 feb 2016 1
play

Lecture 14 - Lecture 14 - 29 Feb 2016 29 Feb 2016 1 - PowerPoint PPT Presentation

Fei-Fei Li & Andrej Karpathy & Justin Johnson Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - Lecture 14 - 29 Feb 2016 29 Feb 2016 1 Administrative Everyone should be done with Assignment 3 now Milestone


  1. Infinite (in theory) RNN temporal extent (neurons that are function of all video frames in the past) Finite temporal 3D extent CONVNET (neurons that are only a function of finitely many video frames in the past) video Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 36

  2. i.e. we obtain: Infinite (in theory) temporal extent (neurons that are function RNN of all video frames in the past) CONVNET video Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 37

  3. Summary - You think you need a Spatio-Temporal Fancy Video ConvNet - STOP. Do you really? - Okay fine: do you want to model: - local motion? (use 3D CONV), or - global motion? (use LSTM). - Try out using Optical Flow in a second stream (can work better sometimes) - Try out GRU-RCN! (imo best model) Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 38

  4. Unsupervised Learning Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 39

  5. Unsupervised Learning Overview ● Definitions ● Autoencoders ○ Vanilla ○ Variational ● Adversarial Networks Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 40

  6. Supervised vs Unsupervised Supervised Learning Data : (x, y) x is data, y is label Goal : Learn a function to map x -> y Examples : Classification, regression, object detection, semantic segmentation, image captioning, etc Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 41

  7. Supervised vs Unsupervised Supervised Learning Unsupervised Learning Data : (x, y) Data : x x is data, y is label Just data, no labels! Goal : Learn a function to Goal : Learn some structure map x -> y of the data Examples : Classification, Examples : Clustering, regression, object detection, dimensionality reduction, feature semantic segmentation, image learning, generative models, etc. captioning, etc Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 42

  8. Unsupervised Learning ● Autoencoders ○ Traditional: feature learning ○ Variational: generate samples ● Generative Adversarial Networks: Generate samples Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 43

  9. Autoencoders z Features Encoder x Input data Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 44

  10. Autoencoders Originally : Linear + nonlinearity (sigmoid) Later : Deep, fully-connected Later : ReLU CNN z Features Encoder x Input data Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 45

  11. Autoencoders Originally : Linear + nonlinearity (sigmoid) z usually smaller than x Later : Deep, fully-connected (dimensionality reduction) Later : ReLU CNN z Features Encoder x Input data Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 46

  12. Autoencoders Reconstructed xx input data Decoder z Features Encoder x Input data Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 47

  13. Originally : Linear + Autoencoders nonlinearity (sigmoid) Later : Deep, fully-connected Later : ReLU CNN (upconv) Reconstructed xx input data Decoder Encoder : 4-layer conv Decoder : 4-layer upconv z Features Encoder x Input data Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 48

  14. Originally : Linear + Autoencoders nonlinearity (sigmoid) Later : Deep, fully-connected Later : ReLU CNN (upconv) Reconstructed xx input data Train for Decoder Encoder / decoder reconstruction sometimes share with no labels! weights z Features Example : Encoder dim( x ) = D dim( z ) = H x Input data w e : H x D T w d : D x H = w e Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 49

  15. Autoencoders Loss function (Often L2) Reconstructed xx input data Train for Decoder reconstruction with no labels! z Features Encoder x Input data Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 50

  16. Autoencoders Reconstructed xx input data Decoder After training, throw away decoder! z Features Encoder x Input data Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 51

  17. Autoencoders Loss function (Softmax, etc) bird plane Predicted yy y dog deer truck Label Use encoder to initialize a Classifier supervised Train for final task Fine-tune model (sometimes with z encoder Features small data) jointly with classifier Encoder x Input data Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 52

  18. Autoencoders: Greedy Training In mid 2000s layer-wise pretraining with Restricted Boltzmann Machines (RBM) was common Training deep nets was hard in 2006! Hinton and Salakhutdinov, “Reducing the Dimensionality of Data with Neural Networks”, Science 2006 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 53

  19. Autoencoders: Greedy Training In mid 2000s layer-wise pretraining with Restricted Not common anymore Boltzmann Machines (RBM) was common Training deep nets was hard in 2006! With ReLU, proper initialization, batchnorm, Adam, etc easily train from scratch Hinton and Salakhutdinov, “Reducing the Dimensionality of Data with Neural Networks”, Science 2006 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 54

  20. Autoencoders Autoencoders can reconstruct data, and Reconstructed can learn features to xx input data initialize a supervised Decoder model z Features Can we generate images from an Encoder autoencoder? x Input data Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 55

  21. Variational Autoencoder A Bayesian spin on an autoencoder - lets us generate data! Assume our data is generated like this: Sample from true conditional Sample from z x true prior Kingma and Welling, “Auto-Encoding Variational Bayes”, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 56

  22. Variational Autoencoder Intuition : x is an image, z gives A Bayesian spin on an autoencoder! class, orientation, attributes, etc Assume our data is generated like this: Sample from true conditional Sample from z x true prior Kingma and Welling, “Auto-Encoding Variational Bayes”, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 57

  23. Variational Autoencoder Intuition : x is an image, z gives A Bayesian spin on an autoencoder! class, orientation, attributes, etc Assume our data is generated like this: Sample from true Problem : Estimate conditional � without access to Sample from latent states ! z x true prior Kingma and Welling, “Auto-Encoding Variational Bayes”, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 58

  24. Variational Autoencoder Prior : Assume is a unit Gaussian Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 59

  25. Variational Autoencoder Prior : Assume is a unit Gaussian Conditional : Assume is a diagonal Gaussian, predict mean and variance with neural net Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 60

  26. Variational Autoencoder Mean and (diagonal) Prior : Assume covariance of is a unit Gaussian � x Σ x Conditional : Assume is a Decoder network diagonal Gaussian, with parameters � predict mean and z variance with neural net Latent state Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 61

  27. Variational Autoencoder Mean and (diagonal) Prior : Assume covariance of is a unit Gaussian � x Σ x Conditional : Assume is a Decoder network diagonal Gaussian, with parameters � predict mean and z variance with neural Fully-connected or net Latent state upconvolutional Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 62

  28. Variational Autoencoder: Encoder By Bayes Rule the posterior is: Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 63

  29. Variational Autoencoder: Encoder By Bayes Rule the posterior is: Use decoder network =) Gaussian =) Intractible integral =( Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 64

  30. Variational Autoencoder: Encoder Mean and (diagonal) By Bayes Rule the posterior is: covariance of � z Σ z Use decoder network =) Encoder network Gaussian =) with parameters � Intractible integral =( x Data point Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 65

  31. Variational Autoencoder: Encoder Mean and (diagonal) By Bayes Rule the posterior is: covariance of � z Σ z Use decoder network =) Encoder network Gaussian =) with parameters � Intractible integral =( x Approximate posterior with encoder network Data point Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 66

  32. Variational Autoencoder: Encoder Mean and (diagonal) By Bayes Rule the posterior is: covariance of Fully-connected or convolutional � z Σ z Use decoder network =) Encoder network Gaussian =) with parameters � Intractible integral =( x Approximate posterior with encoder network Data point Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 67

  33. Variational Autoencoder x Data point Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 68

  34. Variational Autoencoder Mean and (diagonal) � z Σ z covariance of Encoder network x Data point Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 69

  35. Variational Autoencoder z Sample from Mean and (diagonal) � z Σ z covariance of Encoder network x Data point Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 70

  36. Variational Autoencoder � x Σ x Mean and (diagonal) Decoder network covariance of z Sample from Mean and (diagonal) � z Σ z covariance of Encoder network x Data point Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 71

  37. Variational Autoencoder xx Reconstructed Sample from � x Σ x Mean and (diagonal) Decoder network covariance of z Sample from Mean and (diagonal) � z Σ z covariance of Encoder network x Data point Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 72

  38. Variational Autoencoder Training like a normal autoencoder: xx Reconstructed reconstruction loss at the end, Sample from regularization toward prior in middle � x Σ x Mean and (diagonal) Decoder network covariance of (should be close to data x) z Sample from Mean and (diagonal) � z Σ z covariance of Encoder network (should be close to prior ) x Data point Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 73

  39. Variational Autoencoder: Generate Data! After network is trained: z Sample from prior Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 74

  40. Variational Autoencoder: Generate Data! After network is trained: � x Σ x Decoder network z Sample from prior Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 75

  41. Variational Autoencoder: Generate Data! After network is trained: xx Generated Sample from � x Σ x Decoder network z Sample from prior Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 76

  42. Variational Autoencoder: Generate Data! After network is trained: xx Generated Sample from � x Σ x Decoder network z Sample from prior Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 77

  43. Variational Autoencoder: Generate Data! After network is trained: xx Generated Sample from � x Σ x Decoder network z Sample from prior Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 78

  44. Variational Autoencoder: Generate Data! Diagonal prior on z => After network is trained: independent latent variables xx Generated Sample from � x Σ x Decoder network z Sample from prior Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 79

  45. Variational Autoencoder: Math Maximum Likelihood? Maximize likelihood of dataset Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 80

  46. Variational Autoencoder: Math Maximum Likelihood? Maximize likelihood of dataset Maximize log-likelihood instead because sums are nicer Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 81

  47. Variational Autoencoder: Math Maximum Likelihood? Maximize likelihood of dataset Maximize log-likelihood instead because sums are nicer Marginalize joint distribution Kingma and Welling, ICLR 2014 Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 82

  48. Variational Autoencoder: Math Maximum Likelihood? Maximize likelihood of dataset Maximize log-likelihood instead because sums are nicer Intractible integral =( Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 83

  49. Variational Autoencoder: Math Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 84

  50. Variational Autoencoder: Math Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 85

  51. Variational Autoencoder: Math Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 86

  52. Variational Autoencoder: Math Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 87

  53. Variational Autoencoder: Math Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 88

  54. Variational Autoencoder: Math Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 89

  55. Variational Autoencoder: Math “Elbow” Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 90

  56. Variational Autoencoder: Math “Elbow” Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 91

  57. Variational Autoencoder: Math “Elbow” Variational lower bound (elbow) Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 92

  58. Variational Autoencoder: Math “Elbow” Variational lower bound (elbow) Training: Maximize lower bound Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 93

  59. Variational Autoencoder: Math Reconstruct the input data “Elbow” Variational lower bound (elbow) Training: Maximize lower bound Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 94

  60. Variational Autoencoder: Math Latent states should follow the prior Reconstruct the input data “Elbow” Variational lower bound (elbow) Training: Maximize lower bound Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 95

  61. Variational Autoencoder: Math Latent states should follow the prior Reconstruct the input data Sampling with reparam. trick (see paper) “Elbow” Variational lower bound (elbow) Training: Maximize lower bound Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 96

  62. Variational Autoencoder: Math Latent states should follow the prior Reconstruct Everything is the input Gaussian, data closed form Sampling solution! with reparam. trick (see paper) “Elbow” Variational lower bound (elbow) Training: Maximize lower bound Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 97

  63. Autoencoder Overview ● Traditional Autoencoders ○ Try to reconstruct input ○ Used to learn features, initialize supervised model ○ Not used much anymore ● Variational Autoencoders ○ Bayesian meets deep learning ○ Sample from model to generate images Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 98

  64. Goodfellow et al, “Generative Adversarial Nets”, NIPS 2014 Generative Adversarial Nets Can we generate images with less math? z Random noise Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 99

  65. Goodfellow et al, “Generative Adversarial Nets”, NIPS 2014 Generative Adversarial Nets Can we generate images with less math? x Fake image Generator z Random noise Fei-Fei Li & Andrej Karpathy & Justin Johnson Lecture 14 - 29 Feb 2016 100

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend