Generative Adversarial Networks presented by Ian Goodfellow - - PowerPoint PPT Presentation

generative adversarial networks
SMART_READER_LITE
LIVE PREVIEW

Generative Adversarial Networks presented by Ian Goodfellow - - PowerPoint PPT Presentation

Generative Adversarial Networks presented by Ian Goodfellow presentation co-developed with Aaron Courville 1 In todays talk Generative Adversarial Networks Goodfellow et al., NIPS 2014 Conditional Generative


slide-1
SLIDE 1

Generative Adversarial Networks

1

presented by Ian Goodfellow presentation co-developed with Aaron Courville

slide-2
SLIDE 2

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

In today’s talk…

  • “Generative Adversarial Networks” Goodfellow et al., NIPS

2014

  • “Conditional Generative Adversarial Nets” Mirza and

Osindero, NIPS Deep Learning Workshop 2014

  • “On Distinguishability Criteria for Estimating Generative

Models” Goodfellow, ICLR Workshop 2015

  • “Deep Generative Image Models using a Laplacian Pyramid
  • f Adversarial Networks” Denton, Chintala, et al., ArXiv

2015

2

slide-3
SLIDE 3

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Generative modeling

  • Have training examples x ~ pdata(x )
  • Want a model that can draw samples: x ~ pmodel(x )
  • Where pmodel ≈ pdata

3

x ~ pdata(x ) x ~ pmodel(x )

slide-4
SLIDE 4

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Why generative models?

  • Conditional generative models
  • Speech synthesis: Text ⇒ Speech
  • Machine Translation: French ⇒ English
  • French: Si mon tonton tond ton tonton, ton tonton sera tondu.
  • English: If my uncle shaves your uncle, your uncle will be shaved
  • Image ⇒ Image segmentation
  • Environment simulator
  • Reinforcement learning
  • Planning
  • Leverage unlabeled data?

4

slide-5
SLIDE 5

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Maximum likelihood: the dominant approach

  • ML objective function

5

θ∗ = max

θ

1 m

m

X

i=1

log p ⇣ x(i); θ ⌘

slide-6
SLIDE 6

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Undirected graphical models

  • Flagship undirected graphical model: Deep

Boltzmann machines

  • Several “hidden layers” h

6

p(h, x) = 1 Z ˜ p(h, x) ˜ p(h, x) = exp(−E(h, x)) Z =

  • h,x

˜ p(h, x)

h(1) h(2) h(3) x

slide-7
SLIDE 7

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow 7

Boltzmann Machines: disadvantage

  • Model is badly parameterized for learning high

quality samples: peaked distributions -> slow mixing

  • Why poor mixing?

MNIST dataset 1st layer features (RBM)

Coordinated flipping of low- level features

slide-8
SLIDE 8

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Directed graphical models

  • Two problems:
  • 1. Summation over exponentially many states in h
  • 2. Posterior inference, i.e. calculating p(h | x), is intractable.

8

p(x, h) = p(x | h(1))p(h(1) | h(2)) . . . p(h(L−1) | h(L))p(h(L))

h(1) h(2) h(3) x

d dθi log p(x) = 1 p(x) d dθi p(x) p(x) =

  • h

p(x | h)p(h)

slide-9
SLIDE 9

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Variational Autoencoder

9

E[x|z] Differentiable decoder x sampled from data Differentiable encoder Sample from q(z) Noise

z x

Maximize log p(x) DKL (q(x)kp(z | x))

(Kingma and Welling, 2014, Rezende et al 2014)

slide-10
SLIDE 10

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Generative stochastic networks

  • General strategy: Do not write a formula for p(x),

just learn to sample incrementally.

  • Main issue: Subject to some of the same constraints
  • n mixing as undirected graphical models.

10

... (Bengio et al 2013)

slide-11
SLIDE 11

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Generative adversarial networks

  • Don’t write a formula for p(x), just learn to sample

directly.

  • No Markov Chain
  • No variational bound
  • How? By playing a game.

11

slide-12
SLIDE 12

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Game theory: the basics

  • N>1 players
  • Clearly defined set of actions each player can take
  • Clearly defined relationship between actions and
  • utcomes
  • Clearly defined value of each outcome
  • Can’t control the other player’s actions

12

slide-13
SLIDE 13

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Two-player zero-sum game

  • Your winnings + your opponent’s winnings = 0
  • Minimax theorem: a rational strategy exists for all

such finite games

13

slide-14
SLIDE 14

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

  • Strategy: specification of which moves you make in which

circumstances.

  • Equilibrium: each player’s strategy is the best possible for

their opponent’s strategy.

  • Example: Rock-paper-scissors:
  • Mixed strategy equilibrium
  • Choose your action at random

14

  • 1

1 1

  • 1
  • 1

1

You Your opponent Rock Paper Scissors Rock Paper Scissors

Two-player zero-sum game

slide-15
SLIDE 15

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Adversarial nets framework

15

  • A game between two players:
  • 1. Discriminator D
  • 2. Generator G
  • D tries to discriminate between:
  • A sample from the data distribution.
  • And a sample from the generator G.
  • G tries to “trick” D by generating samples that are

hard for D to distinguish from data.

slide-16
SLIDE 16

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Adversarial nets framework

16

Input noise Z Differentiable function G x sampled from model Differentiable function D D tries to

  • utput 0

x sampled from data Differentiable function D D tries to

  • utput 1

x x z

slide-17
SLIDE 17

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

  • Minimax value function:

Zero-sum game

17

min

G max D V (D, G) = Ex∼pdata(x)[log D(x)] + Ez∼pz(z)[log(1 − D(G(z)))].

Generator pushes down Discriminator pushes up Discriminator’s ability to recognize data as being real Discriminator’s ability to recognize generator samples as being fake

slide-18
SLIDE 18

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Discriminator strategy

  • Optimal strategy for any pmodel(x) is always

18

D(x) = pdata(x) pdata(x) + pmodel(x)

slide-19
SLIDE 19

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Learning process

19

...

Poorly fit model After updating D After updating G Mixed strategy equilibrium Data distribution Model distribution

D(x)

slide-20
SLIDE 20

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Learning process

20

...

Poorly fit model After updating D After updating G Mixed strategy equilibrium Data distribution Model distribution

D(x)

slide-21
SLIDE 21

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Learning process

21

...

Poorly fit model After updating D After updating G Mixed strategy equilibrium Data distribution Model distribution

D(x)

slide-22
SLIDE 22

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Learning process

22

...

Poorly fit model After updating D After updating G Mixed strategy equilibrium Data distribution Model distribution

D(x)

slide-23
SLIDE 23

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Theoretical properties

  • Theoretical properties (assuming infinite data, infinite

model capacity, direct updating of generator’s distribution):

  • Unique global optimum.
  • Optimum corresponds to data distribution.
  • Convergence to optimum guaranteed.

23

min

G max D V (D, G) = Ex∼pdata(x)[log D(x)] + Ez∼pz(z)[log(1 − D(G(z)))].

In practice: no proof that SGD converges

slide-24
SLIDE 24

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Oscillation

24

(Alec Radford)

slide-25
SLIDE 25

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Visualization of model samples

25

MNIST TFD CIFAR-10 (fully connected) CIFAR-10 (convolutional)

slide-26
SLIDE 26

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Learned 2-D manifold of MNIST

26

slide-27
SLIDE 27

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

  • 1. Draw sample (A)
  • 2. Draw sample (B)
  • 3. Simulate samples

along the path between A and B

  • 4. Repeat steps 1-3 as

desired.

Visualizing trajectories

27

A B

slide-28
SLIDE 28

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Visualization of model trajectories

28

MNIST digit dataset Toronto Face Dataset (TFD)

slide-29
SLIDE 29

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow 29

CIFAR-10 (convolutional)

Visualization of model trajectories

slide-30
SLIDE 30

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

GANs vs VAEs

  • Both use backprop through continuous random number generation
  • VAE:
  • generator gets direct output target
  • need REINFORCE to do discrete latent variables
  • possible underfitting due to variational approximation
  • gets global image composition right but blurs details
  • GAN:
  • generator never sees the data
  • need REINFORCE to do discrete visible variables
  • possible underfitting due to non-convergence
  • gets local image features right but not global structure

30

slide-31
SLIDE 31

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

VAE + GAN

31

(Alec Radford, 2015) VAE VAE+GAN

  • Reduce

VAE blurriness

  • Reduce GAN oscillation
slide-32
SLIDE 32

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

MMD-based generator nets

32

(Li et al 2015) (Dziugaite et al 2015)

slide-33
SLIDE 33

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Supervised Generator Nets

33

(Dosovitskiy et al 2014) Generator nets are powerful—it is our ability to infer a mapping from an unobserved space that is limited.

slide-34
SLIDE 34

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

General game

34

slide-35
SLIDE 35

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Extensions

35

  • Inference net:
  • Learn a network to model p(z | x)
  • Wake/Sleep style approach
  • Sample z from prior
  • Sample x from p(z|x)
  • Learn mapping from x to z
  • Infinite training set!
slide-36
SLIDE 36

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Extensions

36

  • Conditional model:
  • Learn p(x | y)
  • Discriminator is trained
  • n (x,y) pairs
  • Generator net gets y

and z as input

  • Useful for: Translation,

speech synth, image segmentation. (Mirza and Osindero, 2014)

slide-37
SLIDE 37

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Laplacian Pyramid

37

(Denton + Chintala, et al 2015)

slide-38
SLIDE 38

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

LAPGAN results

  • 40% of samples mistaken by humans for real photos

38

(Denton + Chintala, et al 2015)

slide-39
SLIDE 39

Deep Learning Workshop, ICML 2015 --- Ian Goodfellow

Open problems

  • Is non-convergence a serious problem in

practice?

  • If so, how can we prevent non-

convergence?

  • Is there a better loss function for the

generator?

39

slide-40
SLIDE 40

Thank You.

40

Questions?