probabilistic unsupervised learning parametric
play

Probabilistic & Unsupervised Learning Parametric Variational - PowerPoint PPT Presentation

Probabilistic & Unsupervised Learning Parametric Variational Methods and Recognition Models Maneesh Sahani maneesh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit, and MSc ML/CSML, Dept Computer Science University College London


  1. Probabilistic & Unsupervised Learning Parametric Variational Methods and Recognition Models Maneesh Sahani maneesh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit, and MSc ML/CSML, Dept Computer Science University College London Term 1, Autumn 2017

  2. Variational methods ◮ Our treatment of variational methods has (except EP) emphasised ‘natural’ choices of variational family – most often factorised using the same functional (ExpFam) form as joint. ◮ mostly restricted to joint exponential families – facilitates hierarchical and distributed models, but not non-linear/non-conjugate. ◮ Parametric variational methods might extend our reach. Define a parametric family of posterior approximations q ( Y ; ρ ) . The constrained (approximate) variational E-step becomes: � q ( Y ) , θ ( k − 1 ) � ρ ( k ) := argmax � q ( Y ; ρ ) , θ ( k − 1 ) � q ( Y ) := argmax F ⇒ F q ∈{ q ( Y ; ρ ) } ρ and so we can replace constrained optimisation of F ( q , θ ) with unconstrained optimisation of a constrained F ( ρ, θ ) : � � log P ( X , Y| θ ( k − 1 ) ) F ( ρ, θ ) = q ( Y ; ρ ) + H [ ρ ] It might still be valuable to use coordinate ascent in ρ and θ , although this is no longer necessary.

  3. Optimising the variational parameters � � log P ( X , Y| θ ( k − 1 ) ) F ( ρ, θ ) = q ( Y ; ρ ) + H [ ρ ] ◮ In some special cases, the expectations of the log-joint under q ( Y ; ρ ) can be expressed in closed form, but these are rare.

  4. Optimising the variational parameters � � log P ( X , Y| θ ( k − 1 ) ) F ( ρ, θ ) = q ( Y ; ρ ) + H [ ρ ] ◮ In some special cases, the expectations of the log-joint under q ( Y ; ρ ) can be expressed in closed form, but these are rare. ◮ Otherwise we might seek to follow ∇ ρ F .

  5. Optimising the variational parameters � � log P ( X , Y| θ ( k − 1 ) ) F ( ρ, θ ) = q ( Y ; ρ ) + H [ ρ ] ◮ In some special cases, the expectations of the log-joint under q ( Y ; ρ ) can be expressed in closed form, but these are rare. ◮ Otherwise we might seek to follow ∇ ρ F . ◮ Naively, this requires evaluting a high-dimensional expectation wrt q ( Y , ρ ) as a function of ρ – not simple.

  6. Optimising the variational parameters � � log P ( X , Y| θ ( k − 1 ) ) F ( ρ, θ ) = q ( Y ; ρ ) + H [ ρ ] ◮ In some special cases, the expectations of the log-joint under q ( Y ; ρ ) can be expressed in closed form, but these are rare. ◮ Otherwise we might seek to follow ∇ ρ F . ◮ Naively, this requires evaluting a high-dimensional expectation wrt q ( Y , ρ ) as a function of ρ – not simple. ◮ At least three solutions:

  7. Optimising the variational parameters � � log P ( X , Y| θ ( k − 1 ) ) F ( ρ, θ ) = q ( Y ; ρ ) + H [ ρ ] ◮ In some special cases, the expectations of the log-joint under q ( Y ; ρ ) can be expressed in closed form, but these are rare. ◮ Otherwise we might seek to follow ∇ ρ F . ◮ Naively, this requires evaluting a high-dimensional expectation wrt q ( Y , ρ ) as a function of ρ – not simple. ◮ At least three solutions: ◮ “Score-based” gradient estimate, and Monte-Carlo (Ranganath et al. 2014).

  8. Optimising the variational parameters � � log P ( X , Y| θ ( k − 1 ) ) F ( ρ, θ ) = q ( Y ; ρ ) + H [ ρ ] ◮ In some special cases, the expectations of the log-joint under q ( Y ; ρ ) can be expressed in closed form, but these are rare. ◮ Otherwise we might seek to follow ∇ ρ F . ◮ Naively, this requires evaluting a high-dimensional expectation wrt q ( Y , ρ ) as a function of ρ – not simple. ◮ At least three solutions: ◮ “Score-based” gradient estimate, and Monte-Carlo (Ranganath et al. 2014). ◮ Recognition network trained in separate phase – not strictly variational (Dayan et al. 1995).

  9. Optimising the variational parameters � � log P ( X , Y| θ ( k − 1 ) ) F ( ρ, θ ) = q ( Y ; ρ ) + H [ ρ ] ◮ In some special cases, the expectations of the log-joint under q ( Y ; ρ ) can be expressed in closed form, but these are rare. ◮ Otherwise we might seek to follow ∇ ρ F . ◮ Naively, this requires evaluting a high-dimensional expectation wrt q ( Y , ρ ) as a function of ρ – not simple. ◮ At least three solutions: ◮ “Score-based” gradient estimate, and Monte-Carlo (Ranganath et al. 2014). ◮ Recognition network trained in separate phase – not strictly variational (Dayan et al. 1995). ◮ Recognition network trained simultaneously with generative model using “frozen” samples (Kingma and Welling 2014; Rezende et al. 2014).

  10. Score-based gradient estimate We have: � ∇ ρ F ( ρ, θ ) = ∇ ρ d Y q ( Y ; ρ )( log P ( X , Y| θ ) − log q ( Y ; ρ )) � = d Y [ ∇ ρ q ( Y ; ρ )]( log P ( X , Y| θ ) − log q ( Y ; ρ )) + q ( Y ; ρ ) ∇ ρ [ log P ( X , Y| θ ) − log q ( Y ; ρ )]

  11. Score-based gradient estimate We have: � ∇ ρ F ( ρ, θ ) = ∇ ρ d Y q ( Y ; ρ )( log P ( X , Y| θ ) − log q ( Y ; ρ )) � = d Y [ ∇ ρ q ( Y ; ρ )]( log P ( X , Y| θ ) − log q ( Y ; ρ )) + q ( Y ; ρ ) ∇ ρ [ log P ( X , Y| θ ) − log q ( Y ; ρ )] Now, ∇ ρ log P ( X , Y| θ ) = 0 (no direct dependence) � � d Y q ( Y ; ρ ) ∇ ρ log q ( Y ; ρ ) = ∇ ρ d Y q ( Y ; ρ ) = 0 (always normalised) ∇ ρ q ( Y ; ρ ) = q ( Y ; ρ ) ∇ ρ log q ( Y ; ρ ) So, � � ∇ ρ F ( ρ, θ ) = [ ∇ ρ log q ( Y ; ρ )]( log P ( X , Y| θ ) − log q ( Y ; ρ )) q ( Y ; ρ )

  12. Score-based gradient estimate We have: � ∇ ρ F ( ρ, θ ) = ∇ ρ d Y q ( Y ; ρ )( log P ( X , Y| θ ) − log q ( Y ; ρ )) � = d Y [ ∇ ρ q ( Y ; ρ )]( log P ( X , Y| θ ) − log q ( Y ; ρ )) + q ( Y ; ρ ) ∇ ρ [ log P ( X , Y| θ ) − log q ( Y ; ρ )] Now, ∇ ρ log P ( X , Y| θ ) = 0 (no direct dependence) � � d Y q ( Y ; ρ ) ∇ ρ log q ( Y ; ρ ) = ∇ ρ d Y q ( Y ; ρ ) = 0 (always normalised) ∇ ρ q ( Y ; ρ ) = q ( Y ; ρ ) ∇ ρ log q ( Y ; ρ ) So, � � ∇ ρ F ( ρ, θ ) = [ ∇ ρ log q ( Y ; ρ )]( log P ( X , Y| θ ) − log q ( Y ; ρ )) q ( Y ; ρ ) Reduced gradient of expectation to expectation of gradient – easier to compute.

  13. Factorisation � � ∇ ρ F ( ρ, θ ) = [ ∇ ρ log q ( Y ; ρ )]( log P ( X , Y| θ ) − log q ( Y ; ρ )) q ( Y ; ρ ) ◮ Still requires a high-dimensional expectation, but can now be evaluated by Monte-Carlo. ◮ Dimensionality reduced by factorisation (particularly where P ( X , Y ) is factorised). Let q ( Y ) = � i q ( Y i | ρ i ) factor over disjoint cliques; let ¯ Y i be the minimal Markov blanket of Y i in the joint; P ¯ Y i be the product of joint factors that include any element of Y i (so the union of their arguments is ¯ Y i ); and P ¬ ¯ Y i the remaining factors. Then, � � � j log q ( Y j ; ρ j )]( log P ( X , Y| θ ) − � ∇ ρ i F ( { ρ j } , θ ) = [ ∇ ρ i j log q ( Y j ; ρ j )) q ( Y ) � � Y i ( X , ¯ = [ ∇ ρ i log q ( Y i ; ρ i )]( log P ¯ Y i ) − log q ( Y i ; ρ i ) q ( ¯ Y i ) � � � + [ ∇ ρ i log q ( Y i ; ρ i )] ( log P ¬ ¯ Y i ( X , Y ¬ i ) − log q ( Y j ; ρ j ) q ( Y ) j � = i � �� � constant wrt Y i So the second term is proportional to �∇ ρ i log q ( Y i ; ρ i ) � q ( Y i ) , which = 0 as before. So expectations are only needed wrt q ( ¯ Y i ) → Message passing!

  14. Sampling So the “black-box” variational approach is as follows: ◮ Choose a parametric (factored) variational family q ( Y ) = � i q ( Y i ; ρ i ) . ◮ Initialise factors. ◮ Repeat to convergence: ◮ Stochastic VE-step . For each i : ◮ Sample from q ( ¯ Y i ) and estimate expected gradient ∇ ρ i F . ◮ Update ρ i along gradient. ◮ Stochastic M-step . For each i : ◮ Sample from each q ( ¯ Y i ) . ◮ Update corresponding parameters. ◮ Stochastic gradient steps may employ a Robbins-Munro step-size sequence to promote convergence. ◮ Variance of the gradient estimators can also be controlled by clever Monte-Carlo techniques (orginal authors used a “control variate” method that we have not studied).

  15. Recognition Models We have not generally distinguished between multivariate models and iid data instances. However, even for large models (such as HMMs), we often work with multiple data draws (e.g. multiple strings) and each instance requires its own variational optimisation. Suppose we have fixed length vectors { ( x i , y i ) } ( y is still latent). ◮ Optimal variational distribution q ∗ ( y i ) depends on x i . � � ◮ Learn this mapping (in parametric form): q y i ; f ( x i ; ρ ) . ◮ f is a general function approximator (a GP , neural network or similar) parametrised by ρ , trained to map x i to the variational parameters of q ( y i ) . ◮ The mapping function f is called a recognition model. ◮ This is approach is now sometimes called amortised inference. How to learn f ?

  16. The Helmholtz Machine Dayan et al. (1995) originally studied binary sigmoid belief net, with parallel recognition model: • • • • • • • • • • • • • • • • • • Two phase learning: ◮ Wake phase: given current f , estimate mean-field representation from data (mean sufficient stats for Bernouilli are just probabilities): ˆ y i = f ( x i ; ρ ) Update generative parameters θ according to ∇ θ F ( { ˆ y i } , θ ) . ◮ Sleep phase: sample { y s , x s } S s = 1 from current generative model. Update recognition parameters ρ to direct f ( x s ) towards y s (simple gradient learning). � ∆ ρ ∝ ( y s − f ( x s ; ρ )) ∇ ρ f ( x s ; ρ ) s

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend