structured probabilistic models for deep learning
play

Structured Probabilistic Models for Deep Learning Lecture slides for - PowerPoint PPT Presentation

Structured Probabilistic Models for Deep Learning Lecture slides for Chapter 16 of Deep Learning www.deeplearningbook.org Ian Goodfellow 2016-10-04 Roadmap Challenges of Unstructured Modeling Using Graphs to Describe Model Structure


  1. Structured Probabilistic Models for Deep Learning Lecture slides for Chapter 16 of Deep Learning www.deeplearningbook.org Ian Goodfellow 2016-10-04

  2. Roadmap • Challenges of Unstructured Modeling • Using Graphs to Describe Model Structure • Sampling from Graphical Models • Advantages of Structured Modeling • Structure Learning and Latent Variables • Inference and Approximate Inference • The Deep Learning Approach to Structured Probabilistic Modeling (Goodfellow 2017)

  3. Tasks for Generative Models • Density estimation • Denoising • Sample generation • Missing value imputation • Conditional sample generation • Conditional density estimation (Goodfellow 2017)

  4. Samples from a BEGAN (Berthelot et al, 2017) Images are 128 pixels wide, 128 pixels tall R, G, and B pixel at each location. (Goodfellow 2017)

  5. Cost of Tabular Approach res k n p Number of variables For BEGAN faces: 128 × 128 = 16384 Number of values per variable For BEGAN faces: 256 There are roughly ten to the power of forty thousand times more points in the discretized domain of the BEGAN face model than there are atoms in the universe. (Goodfellow 2017)

  6. Tabular Approach is Infeasible • Memory: cannot store that many parameters • Runtime: inference and sampling are both slow • Statistical e ffi ciency: extremely high number of parameters requires extremely high number of training examples (Goodfellow 2017)

  7. Roadmap • Challenges of Unstructured Modeling • Using Graphs to Describe Model Structure • Sampling from Graphical Models • Advantages of Structured Modeling • Structure Learning and Latent Variables • Inference and Approximate Inference • The Deep Learning Approach to Structured Probabilistic Modeling (Goodfellow 2017)

  8. Insight of Model Structure • Most variables influence each other • Most variables do not influence each other directly • Describe influence with a graph • Edges represent direct influence • Paths represent indirect influence • Computational and statistical savings come from omissions of edges (Goodfellow 2017)

  9. Directed Models Alice Bob Carol t 0 t 0 t 1 t 1 t 2 t 2 Figure 16.2 p ( x ) = Π i p ( x i | Pa G ( x i )) . (16.1) p ( t 0 , t 1 , t 2 ) = p ( t 0 ) p ( t 1 | t 0 ) p ( t 2 | t 1 ) . (16.2) Directed models work best when influence clearly flows in one direction (Goodfellow 2017)

  10. Undirected Models Undirected models work best when influence has no clear direction or is best modeled as flowing in both directions h y h y h r h r h c h c Does your Does your work roommate have a colleague have a Do you have a cold? cold? cold? (Goodfellow 2017)

  11. Undirected Models Unnormalized probability p ( x ) = Π C ∈ G φ ( C ) . ˜ (16.3) p ( x ) = 1 Z ˜ p ( x ) , (16.4) Partition function Z Z = p ( x ) d x . ˜ (16.5) (Goodfellow 2017)

  12. Separation a s b a s b (a) (b) When s is observed, When s is not observed, it blocks the flow of influence can flow from a influence between a to b and vice versa through s. and b: they are separated (Goodfellow 2017)

  13. Separation example a b c d The nodes a and c are separated One path between a and d is still active, though the other path is blocked, so these two nodes are not separated. (Goodfellow 2017)

  14. d-separation The flow of influence is more complicated for directed models The path between a and b is active for all of these graphs: a s b a s b (b) a s b (b) (a) (a) a s b a b c s (Goodfellow 2017) (c)

  15. d-separation example a b • a and b are d-separated given the empty set • a and e are d-separated given c c • d and e are d-separated given c d e Observing variables can activate paths! • a and b are not d-separated given c • a and b are not d-separated given d (Goodfellow 2017)

  16. A complete graph can represent any probability distribution The benefits of graphical models come from omitting edges (Goodfellow 2017)

  17. Converting between graphs • Any specific probability distribution can be represented by either an undirected or a directed graph • Some probability distributions have conditional independences that one kind of graph fails to imply (the distribution is simpler than the graph describes; need to know the conditional probability distributions to see the independences) (Goodfellow 2017)

  18. Converting directed to undirected a a b h 1 h 1 h 2 h 2 h 3 h 3 b c v 1 v 1 v 2 v 2 v 3 v 3 c Must add an edge between unconnected coparents a a b h 1 h 1 h 2 h 2 h 3 h 3 b c v 1 v 1 v 2 v 2 v 3 v 3 c (Goodfellow 2017)

  19. Converting undirected to directed a b a b a b d c d c d c Assign No loops of Add edges to directions to length triangulate edges. No greater than long loops directed cycles three allowed! allowed. (Goodfellow 2017)

  20. Factor graphs are less ambiguous a b a b f 2 f 2 f 1 f 1 a b f 1 f 1 f 3 f 3 c c c Undirected graph: is Factor graphs this three pairwise disambiguate by potentials or one placing each potential potential over three in the graph variables? (Goodfellow 2017)

  21. Roadmap • Challenges of Unstructured Modeling • Using Graphs to Describe Model Structure • Sampling from Graphical Models • Advantages of Structured Modeling • Structure Learning and Latent Variables • Inference and Approximate Inference • The Deep Learning Approach to Structured Probabilistic Modeling (Goodfellow 2017)

  22. Sampling from directed models • Easy and fast to draw fair samples from the whole model • Ancestral sampling : pass through the graph in topological order. Sample each node given its parents. • Harder to sample some nodes given other nodes, unless the observed nodes are at the start of the topology (Goodfellow 2017)

  23. Sampling from undirected models • Usually requires Markov chains • Usually cannot be done exactly • Usually requires multiple iterations even to approximate • Described in Chapter 17 (Goodfellow 2017)

  24. Roadmap • Challenges of Unstructured Modeling • Using Graphs to Describe Model Structure • Sampling from Graphical Models • Advantages of Structured Modeling • Structure Learning and Latent Variables • Inference and Approximate Inference • The Deep Learning Approach to Structured Probabilistic Modeling (Goodfellow 2017)

  25. Tabular Case • Assume each node has a tabular distribution given its parents • Memory, sampling, inference are now exponential in number of variables in factor with largest scope • For many interesting models, this is very small • e.g., RBMs: all factor scopes are size 2 or 1 • Previously, these costs were exponential in total number of nodes • Statistically, much easier to estimate this manageable number of parameters (Goodfellow 2017)

  26. Roadmap • Challenges of Unstructured Modeling • Using Graphs to Describe Model Structure • Sampling from Graphical Models • Advantages of Structured Modeling • Structure Learning and Latent Variables • Inference and Approximate Inference • The Deep Learning Approach to Structured Probabilistic Modeling (Goodfellow 2017)

  27. Learning about dependencies • Suppose we have thousands of variables • Maybe gene expression data • Some interact • Some do not • We do not know which ahead of time (Goodfellow 2017)

  28. Structure learning strategy • Try out several graphs • See which graph does best job of some criterion • Fitting training set with small model complexity • Fitting validation set • Iterative search, propose new graphs similar to best graph so far (remove edge / add edge / flip edge) (Goodfellow 2017)

  29. Latent variable strategy • Use one graph structure • Many latent variables • Dense connections of latent variables to observed variables • Parameters learn that each latent variable interacts strongly with only a small subset of observed variables • Trainable just with gradient descent; no discrete search over graphs (Goodfellow 2017)

  30. Roadmap • Challenges of Unstructured Modeling • Using Graphs to Describe Model Structure • Sampling from Graphical Models • Advantages of Structured Modeling • Structure Learning and Latent Variables • Inference and Approximate Inference • The Deep Learning Approach to Structured Probabilistic Modeling (Goodfellow 2017)

  31. Inference and Approximate Inference • Inferring marginal distribution over some nodes or conditional distribution of some nodes given other nodes is #P hard • NP-hardness describes decision problems. #P- hardness describes counting problems, e.g., how many solutions are there to a problem where finding one solution is NP-hard • We usually rely on approximate inference, described in chapter 19 (Goodfellow 2017)

  32. Roadmap • Challenges of Unstructured Modeling • Using Graphs to Describe Model Structure • Sampling from Graphical Models • Advantages of Structured Modeling • Structure Learning and Latent Variables • Inference and Approximate Inference • The Deep Learning Approach to Structured Probabilistic Modeling (Goodfellow 2017)

  33. Deep Learning Stylistic Tendencies • Nodes organized into layers • High amount of connectivity between layers • Examples: RBMs, DBMs, GANs, VAEs h 1 h 1 h 2 h 2 h 3 h 3 h 4 h 4 v 1 v 1 v 2 v 2 v 3 v 3 Figure 16.14: An RBM drawn as a Markov network. (Goodfellow 2017)

  34. For more information… (Goodfellow 2017)

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend