structure into machine learning trinity of ai algorithms
play

STRUCTURE INTO MACHINE LEARNING TRINITY OF AI ALGORITHMS COMPUTE - PowerPoint PPT Presentation

Anima Anandkumar BEYOND BLACK BOXES: INFUSING STRUCTURE INTO MACHINE LEARNING TRINITY OF AI ALGORITHMS COMPUTE DATA 2 DEEP LEARNING IS DATA-HUNGRY STRUCTURE-INFUSED LEARNING Learning Data Priors = + Learning Data Priors = + How


  1. Anima Anandkumar BEYOND BLACK BOXES: INFUSING STRUCTURE INTO MACHINE LEARNING

  2. TRINITY OF AI ALGORITHMS COMPUTE DATA 2

  3. DEEP LEARNING IS DATA-HUNGRY STRUCTURE-INFUSED LEARNING Learning Data Priors = +

  4. Learning Data Priors = + How to use structure and domain knowledge to design Priors? Examples of Priors Graphs/Tensors • • Symbolic rules Physical laws • Simulations • Generative models •

  5. Learning Data Priors = + How to use structure and domain knowledge to design Priors? Examples of Priors Graphs/Tensors • • Symbolic rules Physical laws • Simulations • Generative models •

  6. NEXT GENERATION AI FROM PREDICTION TO GENERATION DOG DOG y y x x 6

  7. Generative Adversarial Networks 7

  8. TURING TEST FOR FACE GENERATION http://www.whichfaceisreal.com/index.php 8

  9. WHAT IS THE SOLUTION OF A GAN? GAN objective for loss function 𝑀 min max 𝑀 𝒣, 𝒠 𝒣 𝒠 Real Images Discriminator Loss Latent Generator Code

  10. COMPETITION IN GANS Generator vs Discriminator optimization • Training GANs challenging : unstable and mode collapse • Standard optimization: alternating gradient descent • Fails even for simple case with bilinear objectives

  11. A VERY SIMPLE GAN Current optimization methods:

  12. COMPETITIVE GRADIENT DESCENT Florian Schäfer A NeurIPS 2019

  13. INTUITIONS Opponent awareness in optimization Components in decision making: Belief about loss function 1. Uncertainty of environment 2. Anticipation of action of adversary 3.

  14. COMPETITIVE GRADIENT DESCENT Linear for one player → Bilinear for two players 𝑧 𝑔 + 1 2 𝑔 𝑧 + 𝑧 𝑈 𝛼 𝑦 𝑙+1 − 𝑦 𝑙 = argmin 𝑦 𝑔 + 𝑦 𝑈 𝛼 𝑦 𝑔 + 𝑦 𝑈 𝐸 𝑦𝑧 2𝜃 𝑦 𝑈 𝑦 𝑧 𝑕 + 1 2 𝑕 𝑧 + 𝑧 𝑈 𝛼 𝑧 𝑙+1 − 𝑧 𝑙 = argmin 𝑧 𝑕 + 𝑦 𝑈 𝛼 𝑦 𝑕 + 𝑦 𝑈 𝐸 𝑦𝑧 2𝜃 𝑧 𝑈 𝑧 Local approximation is interactive! Nash equilibrium of local game

  15. A VERY SIMPLE GAN CGD converges for all step sizes:

  16. RESULTS ON W-GAN We use architecture intended for Best performance by WGAN WGAN-GP, no additional loss with Adaptive-CGD (no hyperparameter tuning regularization)

  17. TAKE-AWAYS Competitive optimization in GANs Competition between generation and discriminator leads to instability and mode collapse Stabilization in competitive gradient descent through opponent awareness Implicit competitive regularization through stabilization State of art performance with no tuning and explicit penalties

  18. DISENTANGLEMENT IN STYLEGANS Weili Nie Terro Karas Animesh Garg A Ankit Patel 18

  19. CONTROLLABLE STYLEGAN ● Multi-resolution generator and discriminator ● Generator conditions on factor code ● Mapping network – conditioned styles – modulate each block in the synthesis network ● Encoder shares all layers with discriminator except last ○ to predict factor code. 19

  20. SEMI-SUPERVISED LEARNING 20

  21. DISENTANGLED LEARNING • Loss on encoder encourages disentanglement • Loss incorporates code of real images when available (semi-supervised) 21

  22. 5% OF LABELLED DATA ON CELEB-A (256X256) 22

  23. 1% OF LABELLED DATA ON ISAAC SIM (512X512) 23

  24. TAKE-AWAYS Disentangled learning in StyleGAN Controllable photo-realistic generation in StyleGANs Disentanglement through reconstruction of style codes Semi-supervised learning with very little labeled data 24

  25. Flow-based Generative Models 25

  26. CONTINUOUS NORMALIZING FLOWS p(x) - Exact likelihood - Invertibility - Use ODE solvers p(z) 26

  27. CONTINUOUS NORMALIZING FLOWS z l z L = x z = z 0 p 0 ( x 0 ) p l ( x l ) p L ( x L ) ✓ @f ( z ( t ); ✓ ◆ @ l og p ( z ( t )) ) 0 ) = z, @z ( t ) = f ( z ( t ) ,t ; ✓ = Tr z ( t ) @t @z ( t ) @t 27 ⇒

  28. CONTINUOUS NORMALIZING FLOWS z l z L = x z = z 0 p 0 ( x 0 ) p l ( x l ) p L ( x L ) ✓ @f ( z ( t ); ✓ ◆ @ l og p ( z ( t )) ) 0 ) = z, @z ( t ) = f ( z ( t ) ,t ; ✓ = Tr z ( t ) @t @z ( t ) @t Ordinary Differential Equation 28 ⇒

  29. NEURAL ODE MODELS FOR TIME SERIES 29

  30. AI4PHYSICS: TURBULENCE FORECASTING VIA NEURAL ODE Gavin Portwood, Peetak Mitra, Mateus Dias Riberio, Tan Mihn Nguyen, Anima Anandkumar 30

  31. MOTIVATION Fluid Turbulence is difficult to model: • Multi-scale: Dynamics of different scales non-linear and coupled • Direct numerical simulation (DNS) resolves all scales and hence, is expensive • Current reduced order models are heuristic, not high fidelity • Can neural ODEs help? 31

  32. EXPERIMENTAL RESULTS Neural ODE predictions of evolution of dissipation rate are better • Neural ODE generalizes well on unseen test data • 32

  33. TAKE-AWAYS Flow-based generative models Good alternatives to GANs when likelihood estimates are needed Ideal for scientific applications with underlying differential equations and need for uncertainty estimates Challenges in scaling 33

  34. FEEDBACK GENERATIVE MODELS

  35. NEXT GENERATION AI FROM PREDICTION TO GENERATION DOG DOG y y One model to do both? x x 35

  36. Taking inspiration from Biological brains.. 36

  37. HUMAN VISION: FEEDFORWARD & FEEDBACK Second-order predictions Hierarchical predictive coding First-order predictions First-order streams Prediction error (superficial pyramidal cells) Expectations (deep pyramidal cells) Second-order streams Modulatory backward connections “ Prediction error Excitatory (forward) connections (precision) Inhibitory (backward) connections Expectations (precision) s)” Visual input a ‘ ’ w Representational sharpening : “ [ a] s a Interaction between the feedforward and feedback connections are crucial for core object recognition in human vision 37 y—t s—h ” n– e– s— s ex—i g—a — –

  38. DECONVOLUTIONAL GENERATIVE MODEL object latent category • Feedback network for variables deconvolution • Latent variables to overcome non-invertibility intermediate rendering Nhat Ankit Michael Tan Rich Ho Patel Jordan Nyugen Baranuik image

  39. CONVOLUTIONAL NEURAL NETWORK WITH FEEDBACK y . . . . Yujia Sihui . Pinglei . Huang Dai Bao . . . . . . Doris Tan Rich x Tsaos Nyugen Baranuik CNN-F performs approximate belief propagation through feedforward CNN and feedback generative model 39

  40. CNN-F CAN RECOVER CLEAN DATA Noise Blur Occlusion Input CNN-F Reconstruction

  41. CNN-F YIELDS ROBUST CLASSIFICATION

  42. TAKE-AWAYS Adding feedback to CNNs Biological inspiration can lead to more robust architectures Combining feedforward and feedback networks for iterative inference Robust prediction on degraded images

  43. CONCLUSION Generative models are important in many applications • Photorealistic generation now possible • • Competitive optimization to stabilize GAN training Controlled disentangled generation in Style GANs • Continuous flow-based models for physical applications • • Brain inspired CNN with feedback • Outstanding challenges: How to combine generative models with simulations and • downstream taskss

  44. Thank you 44

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend