Introduction to Generative Adversarial Networks Ian Goodfellow, - - PowerPoint PPT Presentation

introduction to generative adversarial networks
SMART_READER_LITE
LIVE PREVIEW

Introduction to Generative Adversarial Networks Ian Goodfellow, - - PowerPoint PPT Presentation

Introduction to Generative Adversarial Networks Ian Goodfellow, OpenAI Research Scientist NIPS 2016 Workshop on Adversarial Training Barcelona, 2016-12-9 Adversarial Training A phrase whose usage is in flux; a new term that applies to both


slide-1
SLIDE 1

Introduction to Generative Adversarial Networks

Ian Goodfellow, OpenAI Research Scientist NIPS 2016 Workshop on Adversarial Training Barcelona, 2016-12-9

slide-2
SLIDE 2

(Goodfellow 2016)

Adversarial Training

  • A phrase whose usage is in flux; a new term that applies to both new and
  • ld ideas
  • My current usage: “Training a model in a worst-case scenario, with inputs

chosen by an adversary”

  • Examples:
  • An agent playing against a copy of itself in a board game (Samuel, 1959)
  • Robust optimization / robust control (e.g. Rustem and Howe 2002)
  • Training neural networks on adversarial examples (Szegedy et al 2013,

Goodfellow et al 2014)

slide-3
SLIDE 3

(Goodfellow 2016)

Generative Adversarial Networks

  • Both players are neural networks
  • Worst case input for one network is produced by

another network

slide-4
SLIDE 4

(Goodfellow 2016)

Generative Modeling

  • Density estimation
  • Sample generation

Training examples Model samples

slide-5
SLIDE 5

(Goodfellow 2016)

Adversarial Nets Framework

x sampled from data Differentiable function D D(x) tries to be near 1 Input noise z Differentiable function G x sampled from model D D tries to make D(G(z)) near 0, G tries to make D(G(z)) near 1

slide-6
SLIDE 6

(Goodfellow 2016)

Minimax Game

  • Equilibrium is a saddle point of the discriminator loss
  • Resembles Jensen-Shannon divergence
  • Generator minimizes the log-probability of the discriminator

being correct

J(D) = 1 2Ex∼pdata log D(x) 1 2Ez log (1 D (G(z))) J(G) = J(D)

slide-7
SLIDE 7

(Goodfellow 2016)

Discriminator Strategy

D(x) = pdata(x) pdata(x) + pmodel(x)

Data Model distribution

Optimal D(x) for any pdata(x) and pmodel(x) is always

z x

Discriminator

Estimating this ratio using supervised learning is the key approximation mechanism used by GANs

slide-8
SLIDE 8

(Goodfellow 2016)

Non-Saturating Game

J(D) = 1 2Ex∼pdata log D(x) 1 2Ez log (1 D (G(z))) J(G) = 1 2Ez log D (G(z))

  • Equilibrium no longer describable with a single loss
  • Generator maximizes the log-probability of the discriminator

being mistaken

  • Heuristically motivated; generator can still learn even when

discriminator successfully rejects all generator samples

slide-9
SLIDE 9

(Goodfellow 2016)

Vector Space Arithmetic

  • +

=

Man with glasses Man Woman Woman with Glasses (Radford et al, 2015)

slide-10
SLIDE 10

(Goodfellow 2016)

Non-convergence

  • Optimization algorithms often approach a saddle

point or local minimum rather than a global minimum

  • Game solving algorithms may not approach an

equilibrium at all

slide-11
SLIDE 11

(Goodfellow 2016)

Mode Collapse

  • D in inner loop: convergence to correct distribution
  • G in inner loop: place all mass on most likely point

min

G max D V (G, D) 6= max D min G V (G, D)

(Metz et al 2016)

slide-12
SLIDE 12

(Goodfellow 2016)

Minibatch Features

  • Add minibatch features that classify each example

by comparing it to other members of the minibatch (Salimans et al 2016)

  • Nearest-neighbor style features detect if a minibatch

contains samples that are too similar to each other

slide-13
SLIDE 13

(Goodfellow 2016)

Minibatch GAN on CIFAR

Training Data Samples (Salimans et al 2016)

slide-14
SLIDE 14

(Goodfellow 2016)

Minibatch GAN on ImageNet

(Salimans et al 2016)

slide-15
SLIDE 15

(Goodfellow 2016)

Cherry-Picked Results

slide-16
SLIDE 16

(Goodfellow 2016)

Problems with Counting

slide-17
SLIDE 17

(Goodfellow 2016)

Problems with Perspective

slide-18
SLIDE 18

(Goodfellow 2016)

Problems with Global Structure

slide-19
SLIDE 19

(Goodfellow 2016)

Image to Image Translation

Input Ground truth Output

(Isola et al 2016)

Aerial to Map Labels to Street Scene

input

  • utput

input

  • utput
slide-20
SLIDE 20

(Goodfellow 2016)

Plug and Play Generative Models

  • New state of the art generative model (Nguyen et al

2016) released days before NIPS

  • Generates 227x227 realistic images from all

ImageNet classes

  • Combines adversarial training, moment matching,

denoising autoencoders, and Langevin sampling

slide-21
SLIDE 21

(Goodfellow 2016)

PPGN Samples

(Nguyen et al 2016)

slide-22
SLIDE 22

(Goodfellow 2016)

GANs allow many answers

Mean Squared Error GANs

slide-23
SLIDE 23

(Goodfellow 2016)

Next Video Frame Prediction

Ground Truth MSE Adversarial

(Lotter et al 2016)

slide-24
SLIDE 24

(Goodfellow 2016)

Adversarial training for people

  • Markets
  • Cycles due to non-convergence?
  • Auctions, taxation, etc.
  • Deliberate practice (Ericsson et al 1993)
slide-25
SLIDE 25

(Goodfellow 2016)

Conclusion

  • Adversarial training is a term encompassing old and

new work

  • GANs are a generative models that use supervised

learning to estimate a density ratio

  • GANs allow a model to learn that there are many

correct answers

  • Adversarial training can be useful for people as well as

machine learning models