Recent Progress in Generative Modeling Ilya Sutskever Goal of - - PowerPoint PPT Presentation

recent progress in generative modeling
SMART_READER_LITE
LIVE PREVIEW

Recent Progress in Generative Modeling Ilya Sutskever Goal of - - PowerPoint PPT Presentation

Recent Progress in Generative Modeling Ilya Sutskever Goal of OpenAI Make sure that AI is actually good for humanity Goal of OpenAI Prevent concentration of AI power Build AI to benefit as many people as possible Build AI that


slide-1
SLIDE 1

Recent Progress in Generative Modeling

Ilya Sutskever

slide-2
SLIDE 2

Goal of OpenAI

  • Make sure that AI is actually good for humanity
slide-3
SLIDE 3

Goal of OpenAI

  • Prevent concentration of AI power
  • Build AI to benefit as many people as possible
  • Build AI that will do what we want it to do
slide-4
SLIDE 4

ML: what works?

  • Deep supervised learning
  • Vision, speech, translation, language, ads,

robotics

slide-5
SLIDE 5

ML: what works?

  • Deep supervised learning:
  • Get lots of input-output examples
  • Train a very large deep neural network
  • Convolutional or seq2seq with attention
  • Great results are likely
slide-6
SLIDE 6

What’s next?

  • Agents that achieve goals
  • Systems that build a holistic understanding of the

world

  • Creative problem solving
  • etc
slide-7
SLIDE 7

Generative models

  • Critical for many of the upcoming problems
slide-8
SLIDE 8

What is a generative model?

  • Learn your data distribution
  • Assign high probability to it
  • Learn to generate plausible structure
  • Discover the “true” structure of the data
slide-9
SLIDE 9

Generative models

  • What are they good for?
  • What can we do with them?
slide-10
SLIDE 10

Conventional applications

  • Good generative models will definitely enable the

following:

  • Structured prediction (e.g., output text)
  • Much more robust prediction
  • Anomaly detection
  • Model-based RL
slide-11
SLIDE 11

Speculative applications

  • Really good feature learning
  • Exploration in RL
  • Inverse RL
  • Good dialog that actually works
  • “Understanding the world”
  • Transfer learning
slide-12
SLIDE 12

Generative models

  • Three broad categories of generative models:
  • Variational Autoencoders
  • Generative adversarial networks
  • Autoregressive models
slide-13
SLIDE 13

Improved techniques for training GANs

  • Tim Salimans, Ian Goodfellow, Wojciech Zaremba,

Vicki Cheung, Alec Radford, Xi Chen

slide-14
SLIDE 14

Generative adversarial networks

  • A generator G(z) and a discriminator D(x)
  • Discriminator aims to separate real data from

generator samples

  • Generator tries to fool the discriminator
  • GANs often produce best samples so far
slide-15
SLIDE 15

Generative adversarial networks

  • Yann LeCun: The most important [recent

development], in my opinion, is adversarial training (also called GAN for Generative Adversarial Networks)
 — from Quora Q&A session

slide-16
SLIDE 16

Promising early results

  • Best high-resolution image samples of any model

so far:

  • Deep generative image models using a

Laplacian pyramid of adversarial networks. 
 — Denton et al.

  • DCGAN 


— Radford et al.

slide-17
SLIDE 17

Hard to train

  • The model is defined in terms of a minimax

problem

  • No cost function
  • Hard to tell if progress is being made
slide-18
SLIDE 18

Simple ideas for improving GAN training

  • GANs fail to learn due to the collapse problem:
  • The generator becomes degenerate and the

learning gets stuck

  • Solution: the discriminator should see the entire

mini batch

  • If all the cases are identical, it will be easier to

discern

slide-19
SLIDE 19

Results

slide-20
SLIDE 20
slide-21
SLIDE 21

–Johnny Appleseed

“Type a quote here.”

slide-22
SLIDE 22
slide-23
SLIDE 23
slide-24
SLIDE 24
slide-25
SLIDE 25

Semi supervised learning with GANs

  • Semi supervised learning is the problem of getting

better classification using unlabelled data

  • A good generic semi supervised learning algorithm

will improve all ML applications

slide-26
SLIDE 26

Semi supervised learning with GANs

  • Discriminator should both tell the class of the

training samples, and tell real samples from fake samples apart

  • The specific way in which it is done is important,

but it is technical, and I will not explain it

  • The GAN training algorithm is also different here.

Details are available offline.

slide-27
SLIDE 27

Results

  • MNIST: 50 supervised training cases + ensemble
  • f 10 models = 1.4% test error
  • CIFAR 10: 4000 supervised training cases = 18.5%

test error

  • Both results are new state of the art
slide-28
SLIDE 28

Conclusions

  • We have better methods for training GANs
  • New simple way of using GANs to improve

discriminative models

  • New level of sample quality and semi-supervised

learning accuracy

slide-29
SLIDE 29

InfoGAN

  • Xi Chen, Rein Houthooft, John Schulman, Ilya

Sutskever, Pieter Abbeel

slide-30
SLIDE 30

Disentangled representations

  • Holy grail of representation learning
slide-31
SLIDE 31

InfoGAN

  • Train a GAN
  • such that: a small subset of its variables is

accurately predictable from the generated sample

  • Straightforward to add this constraint
slide-32
SLIDE 32

Actually works!

slide-33
SLIDE 33
slide-34
SLIDE 34
slide-35
SLIDE 35
slide-36
SLIDE 36
slide-37
SLIDE 37
slide-38
SLIDE 38
slide-39
SLIDE 39
slide-40
SLIDE 40

Exploration with generative models

  • Rein Houthooft, Xi Chen, John Schulman, Filip De

Turck, Pieter Abbeel

slide-41
SLIDE 41

The problem

  • In reinforcement learning, we take random actions
  • Sometimes the actions do us good
  • Then we do more of these actions in the future
slide-42
SLIDE 42

Exploration

  • Are random actions the best we can do?
  • Surely not
slide-43
SLIDE 43

Curiosity

  • Key idea: take actions to maximize “information

gain”

slide-44
SLIDE 44

Formally

  • Learn a Bayesian generative model of the

environment

  • For the action taken, calculate the amount of

information gained about the environment by the generative model

  • Add the amount of information to the reward
slide-45
SLIDE 45

Actually works

  • Extremely well on low-D environments
  • Many unsolvable problems become solvable
  • Current work: scaling up to high-D environments
slide-46
SLIDE 46
slide-47
SLIDE 47
slide-48
SLIDE 48

Improving Variational Autoencoders with Inverse Autoregressive Flow

  • Durk Kingma, Tim Salimans, Max Welling
slide-49
SLIDE 49

The Helmholtz Machine

  • Latent variable model
  • Use an approximate posterior
  • Maximize a lower bound to the likelihood
  • Has been impossible to train
slide-50
SLIDE 50

Reparameterization Trick

  • The Helmholtz machine has been forever

impossible to train

  • The reparameterization trick of Kingma and Welling

fixes this problem, whenever the latent variables are continuous

slide-51
SLIDE 51

High-quality posterior

  • Approximate posteriors matter
  • Typical approximate posteriors are very simple
  • Normal way of doing powerful posteriors is very

expensive

  • IAF = a new cheap way of getting extremely

powerful posteriors

slide-52
SLIDE 52

Results

  • Best non-pixel-CNN log probabilities on CIFAR-10
  • Excellent samples
  • Currently training huge ImageNet models
slide-53
SLIDE 53
slide-54
SLIDE 54
slide-55
SLIDE 55

Questions?