2/11/2020 1
CS3750: ADVANCED MACHINE LEARNING GENERATIVE ADVERSARIAL NETWORKS
Adapted from Slides made by Khushboo Thaker Presented by Tristan Maidment
CS3750: ADVANCED MACHINE LEARNING GENERATIVE ADVERSARIAL NETWORKS - - PDF document
2/11/2020 CS3750: ADVANCED MACHINE LEARNING GENERATIVE ADVERSARIAL NETWORKS Adapted from Slides made by Khushboo Thaker Presented by Tristan Maidment GROWTH (AND DECLINE) IN GAN PAPERS 1 2/11/2020 Overview Why Existing MiniMax
Adapted from Slides made by Khushboo Thaker Presented by Tristan Maidment
Why Generative Modelling? Existing Generative Models Properties of GANs GAN Framework MiniMax game theory for GANs Why GAN training is HARD Tricks for GAN training Common extensions to GANS Conclusion
Inpu nput
Training Examples
Outp tput ut
Some representation of a probability distribution, which defines this example space.
Unsu super pervise ised
Data: X Goal: Learn hidden underlying structure of data
Super pervised ised
Data: X, y Goal: Learn hidden mapping from X -> y
Noisy Input Simulated Data Features Representative of Data Prediction of Future State Missing Data Semi-supervised Learning
π
π(π¦π|π¦1,π¦2,β¦ , π¦πβ1 )
= πΉπ¨~π log π(π¦, π¨) + πΌ(π)
| π
Can use latent information Asymptotically consistent No Markov Chain assumption Samples produced are high quality
https://www.slideshare.net/xavigiro/deep-learning-for-computer-vision-generative-models-and-adversarial-training-upc-2016
z G G(z) x D D(x) D(G(z))
βThe generative model can be thought of as analogous to a team
without detection, while the discriminative model is analogous to the police, trying to detect the counterfeit currency. Competition in this game drives both teams to improve their methods until the counterfeits are indistinguishable from the genuine articles.β β Goodfellow, et. Al. βGenerative Adversarial Netsβ (2014)
discriminator being correct
discriminator being correct
very confident the loss value will be zero
discriminatorβs mistake
GENERATOR KEEPS GENERATING SIMILAR IMAGES β SO NOTHING TO LEARN MAINTAIN TRADE-OFF OF GENERATING MORE ACCURATE VS HIGH COVERAGE SAMPLES THE TWO LEARNING TASKS NEED TO HAVE BALANCE TO ACHIEVE STABILITY IF DISCRIMINATOR IS NOT SUFFICIENTLY TRAINED β LEADS TO POOR GENERATOR PERFORMANCE IF DISCRIMINATOR IS OVER- TRAINED β VANISHING GRADIENT PROBLEM
images or all generated images.
peaks around real data
discriminator in region around real data
Layers
Layers
Layers
images
function is used to generate different scales
images
function is used to generate different scales
GAN is still an active area of research GAN framework is flexible to support variety of learning problems GAN does not guarantee to converge GAN can capture perceptual similarity and generates better images than VAE Needs a lot of work in theoretic foundation of Network Evaluation of GAN is still an open research (Theis et. al)
Deep Learning Book GAN paper: https://arxiv.org/abs/1701.00160 GAN slides: http://slazebni.cs.illinois.edu/spring17/lec11_gan.pd GAN Tutorial: https://www.youtube.com/watch?v=HGYYEUSm-0Q