self supervised model training and selection for
play

Self-Supervised Model Training and Selection for Disentangling GANs - PowerPoint PPT Presentation

Paper ID: 4410 Self-Supervised Model Training and Selection for Disentangling GANs Previous title: InfoGAN-CR: Disentangling Generative Adversarial Networks with Contrastive Regularizers Zinan Lin Kiran Thekumparampil Giulia Fanti Sewoong


  1. Paper ID: 4410 Self-Supervised Model Training and Selection for Disentangling GANs Previous title: InfoGAN-CR: Disentangling Generative Adversarial Networks with Contrastive Regularizers Zinan Lin Kiran Thekumparampil Giulia Fanti Sewoong Oh CMU UIUC CMU UW

  2. Generative Adversarial Networks (GANs) 𝑙 factors 𝑒 input noise 𝑨 ! β€’ Hair color 𝑨 " β€’ Rotation Generator … β€’ Background 𝑨 # β€’ Bangs β€’ How do 𝑨 ! s control the factors? Vanilla GANs Disentangled GANs 𝑨 ! Factor ! 𝑑 ! Factor ! 𝑨 " Factor " 𝑑 " Factor " … … … … 𝑨 # Factor $ 𝑑 $ Factor $ 𝑨 % s ID 4410: Self-supervised model training and selection for disentangling GAN Code & Paper: https://github.com/fjxmlzn/InfoGAN-CR 2

  3. Examples of Disentanglement 𝑑 ! 𝑑 " 𝑑 & 𝑑 ' 𝑑 ( Changing only: Latent (CelebA dataset) codes 𝑑 ! 𝑑 " hair color rotation lighting background bangs Generator … 𝑑 # 𝑨 % s The remaining (dSprites dataset) noise dimensions y-position shape scale rotation x-position * CelebA example is generated by InfoGAN-CR. Dsprites example is synthetic for illustration. ID 4410: Self-supervised model training and selection for disentangling GAN Code & Paper: https://github.com/fjxmlzn/InfoGAN-CR 3

  4. Challenges in Learning Disentangled GANs 1. How to train disentangled GANs? GANs are good at generating high fidelity images, β€’ but are reported to be bad at disentanglement (e.g. compared with VAEs) 2. How to do unsupervised model selection? In practice, we don’t have ground-truth factor labels β€’ for selecting the best disentangled models. ID 4410: Self-supervised model training and selection for disentangling GAN Code & Paper: https://github.com/fjxmlzn/InfoGAN-CR 4

  5. Our Solution: Self-Supervision Assessing the Pairs of images disentanglement with 1. How to train disentangled GANs? quality self-supervised labels InfoGAN-CR β€’ Contrastive Generator Regularizer … (CR) 2. How to do unsupervised model selection? ModelCentrality β€’ Self-supervised All trained A pair of distance metric models: trained models: Pick the most central model ID 4410: Self-supervised model training and selection for disentangling GAN Code & Paper: https://github.com/fjxmlzn/InfoGAN-CR 5

  6. Benchmark dSprites Dataset Model FactorVAE Metric ↑ VAE 0.63 𝛾 -TCVAE 0.62 HFVAE 0.63 VAE Supervised 𝛾 -VAE 0.63 hyper-parameter CHyVAE 0.77 selection FactorVAE 0.82 InfoGAN 0.59 InfoGAN (modified) 0.83 GAN IB-GAN 0.80 InfoGAN-CR 0.90 ±𝟏. 𝟏𝟐 Unsupervised InfoGAN-CR model 0.92 ±𝟏. 𝟏𝟏 model selection selected with ModelCentrality Code & Paper & More results: https://github.com/fjxmlzn/InfoGAN-CR ID 4410: Self-supervised model training and selection for disentangling GAN Code & Paper: https://github.com/fjxmlzn/InfoGAN-CR 6

  7. Details 1. InfoGAN-CR (model training) 2. ModelCentrality (model selection) ID 4410: Self-supervised model training and selection for disentangling GAN Code & Paper: https://github.com/fjxmlzn/InfoGAN-CR 7

  8. Intuition of Contrastive Regularizer (CR) $ ) to generate a pair of images $ ,…, 𝑑 # $ ,…, 𝑑 ! β€’ Use two latent codes ( 𝑑 " ,…, 𝑑 ! ,…, 𝑑 # ), ( 𝑑 " Equal Same i-th latent code Same shape 𝑗 = 1 Generator Same x-position 𝑗 = 2 ( 𝐻 ) 𝑗 = 3 Same y-position ID 4410: Self-supervised model training and selection for disentangling GAN Code & Paper: https://github.com/fjxmlzn/InfoGAN-CR 8

  9. Intuition of Contrastive Regularizer (CR) $ ) to generate a pair of images $ ,…, 𝑑 # $ ,…, 𝑑 ! β€’ Use two latent codes ( 𝑑 " ,…, 𝑑 ! ,…, 𝑑 # ), ( 𝑑 " Equal Same i-th latent code 1 𝑗 = 1 Contrastive Generator 2 𝑗 = 2 Regularizer ( 𝐻 ) (CR) 𝑗 = 3 3 Classification task! ID 4410: Self-supervised model training and selection for disentangling GAN Code & Paper: https://github.com/fjxmlzn/InfoGAN-CR 9

  10. Μ‚ InfoGAN-CR GAN Discriminator ℝ % ∈ ℝ # # Input Noise " InfoGAN Encoder ! ∈ ℝ ! ! $ ∈ ℝ " %β€² ∈ ℝ # Latent Factors $ ∈ ℝ " %β€²β€² ∈ ℝ # CR 𝐼 ℝ " InfoGAN-CR loss: %,',( max min 𝑀 *+, 𝐻, 𝐸 βˆ’ πœ‡π½ 𝐻, 𝑅 βˆ’ 𝛽𝑀 - (𝐻, 𝐼) ) Mutual info GAN’s classification accuracy of CR loss adversarial loss InfoGAN [1] [1] InfoGAN. Chen, X., Duan, Y., Houthooft, R., Schulman, J., Sutskever, I., & Abbeel, P. (NeurIPS 2016) ID 4410: Self-supervised model training and selection for disentangling GAN Code & Paper: https://github.com/fjxmlzn/InfoGAN-CR 10

  11. CR Achieves a Further Gain that Cannot be Gotten with InfoGAN Alone Disentanglement Split training #training iterations (dSprites dataset) ID 4410: Self-supervised model training and selection for disentangling GAN Code & Paper: https://github.com/fjxmlzn/InfoGAN-CR 11

  12. Details 1. InfoGAN-CR (model training) 2. ModelCentrality (model selection) ID 4410: Self-supervised model training and selection for disentangling GAN Code & Paper: https://github.com/fjxmlzn/InfoGAN-CR 12

  13. Intuition of ModelCentrality β€’ Well-disentangled models are close to the (unknown) ground-truth model Model space Ground-truth model (unknown) Bad Good model model β€’ Idea: pick the model at the β€œcenter” ID 4410: Self-supervised model training and selection for disentangling GAN Code & Paper: https://github.com/fjxmlzn/InfoGAN-CR 13

  14. ModelCentrality β€’ How to find the model at the β€œcenter” ? Two models are close Γ³ Their latent codes control similar factors (1) Treat as the Model 𝑗 ground-truth factors Any supervised Asymmetric similarity 𝑑 !" disentanglement metric Similarity 𝑒 !" ? Similarity 𝑒 !" = 𝑒 "! = (𝑑 !" + 𝑑 "! )/2 (e.g. FactorVAE metric [1]) Model π‘˜ (The higher, the more similar) ! (2) ./! βˆ‘ 01% 𝑒 %0 ModelCentrality score 𝑁 % = High similarity Low similarity (3) Selection the model with the highest 𝑁 % [1] Disentangling by Factorising. Kim, H., & Mnih, A. (ICML 2018) ID 4410: Self-supervised model training and selection for disentangling GAN Code & Paper: https://github.com/fjxmlzn/InfoGAN-CR 14

  15. Results of ModelCentrality β€’ Works well for GANs. Take InfoGAN-CR for example: Method FactorVAE Metric ↑ Supervised hyper-parameter selection 0.90 Unsupervised model selection with UDR Lasso [1] 0.86 Unsupervised model selection with UDR Spearman [1] 0.84 Unsupervised model selection with ModelCentrality 0.92 β€’ Also works well for VAEs! Take FactorVAE for example: Method FactorVAE Metric ↑ Supervised hyper-parameter selection 0.83 Unsupervised model selection with UDR Lasso [1] 0.81 Unsupervised model selection with UDR Spearman [1] 0.79 Unsupervised model selection with ModelCentrality 0.84 [1] Unsupervised Model Selection for Variational Disentangled Representation Learning. Duan, S., Matthey, L., Saraiva, A., Watters, N., Burgess, C. P., Lerchner, A., & Higgins, I. (ICLR 2020). ID 4410: Self-supervised model training and selection for disentangling GAN Code & Paper: https://github.com/fjxmlzn/InfoGAN-CR 15

  16. Conclusion β€’ InfoGAN-CR β€’ New disentangled GANs, achieving state-of-the-art results β€’ ModelCentrality β€’ Unsupervised model selection approach for disentangled GANs and VAEs β€’ InfoGAN-CR + ModelCentrality β€’ Unsupervised disentangled generative model training and selection package, achieving state-of-the-art results Code & Paper & More results: https://github.com/fjxmlzn/InfoGAN-CR Including: More theoretical & empirical analysis of InfoGAN & InfoGAN-CR β€’ Analysis of total correlation for GANs and CR for VAEs β€’ New challenging disentanglement datasets β€’ ID 4410: Self-supervised model training and selection for disentangling GAN Code & Paper: https://github.com/fjxmlzn/InfoGAN-CR 16

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend