flow improving flow based generative models with
play

Flow++: Improving Flow-Based Generative Models with Variational - PowerPoint PPT Presentation

Flow++: Improving Flow-Based Generative Models with Variational Dequantization and Architecture Design Jonathan Ho*, Xi Chen*, Aravind Srinivas, Yan Duan, Pieter Abbeel Overview - Goal: likelihood-based model with Fast sampling and training -


  1. Flow++: Improving Flow-Based Generative Models with Variational Dequantization and Architecture Design Jonathan Ho*, Xi Chen*, Aravind Srinivas, Yan Duan, Pieter Abbeel

  2. Overview - Goal: likelihood-based model with Fast sampling and training - Good samples and density estimation performance - - Our strategy: improve flow models Uniform dequantization -> variational dequantization - Affine coupling -> mixture of logistics coupling - Convolutions -> convolutions + self-attention - Flow++ : Improving Flow-Based Generative Models with Variational Dequantization and Architecture Design

  3. Continuous flows for discrete data ■ A problem arises when fitting continuous density models to discrete data: degeneracy When the data are 3-bit pixel values, ■ What density does a model assign to values between bins like 0.4, ■ 0.42 … ? ■ Correct semantics: we want the integral of probability density within a discrete interval to approximate discrete probability mass Flow++ : Improving Flow-Based Generative Models with Variational Dequantization and Architecture Design 3

  4. Continuous flows for discrete data ■ Solution: Dequantization . Add noise to data. ■ We draw noise u uniformly from ■ [Theis, Oord, Bethge, 2016] Flow++ : Improving Flow-Based Generative Models with Variational Dequantization and Architecture Design 4

  5. Variational Dequantization ■ Variational Dequantization . Add a learnable noise q to data. [Ho et al., 2019] Flow++ : Improving Flow-Based Generative Models with Variational Dequantization and Architecture Design 5

  6. Coupling layers RealNVP convolutions Ours: logistic mixture CDF convolutions & self-attention Flow++ : Improving Flow-Based Generative Models with Variational Dequantization and Architecture Design

  7. Ablation on CIFAR Flow++ : Improving Flow-Based Generative Models with Variational Dequantization and Architecture Design 7

  8. Results Flow++ : Improving Flow-Based Generative Models with Variational Dequantization and Architecture Design

  9. Samples (CIFAR10, ImageNet 64x64) Flow++ : Improving Flow-Based Generative Models with Variational Dequantization and Architecture Design

  10. Samples (CelebA 5-bit) Flow++ : Improving Flow-Based Generative Models with Variational Dequantization and Architecture Design

  11. ■ Slides adapted from Berkeley CS294-158 Deep Unsupervised Learning class: https://sites.google.com/view/berkeley-cs294-158-sp19/home Want to learn more about foundation of Deep Generative Models & ■ Self-Supervised learning methods? All lecture videos are available on youtube, featuring guest speakers: ■ Ilya Sutskever, Alyosha Efros, Alec Radford, Aaron van den Oord Flow++ : Improving Flow-Based Generative Models with Variational Dequantization and Architecture Design

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend