Approximate Posterior Sampling via Stochastic Optimisation
Connie Trojan Supervisor: Srshti Putcha 6th September 2019
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Approximate Posterior Sampling via Stochastic Optimisation Connie - - PowerPoint PPT Presentation
Approximate Posterior Sampling via Stochastic Optimisation Connie Trojan Supervisor: Srshti Putcha 6 th September 2019 Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation Background Large scale
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
N
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Notation
N
N
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
1 Take a subsample St of size n from the data
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
1 Take a subsample St of size n from the data 2 Estimate the gradient at θt by :
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
1 Take a subsample St of size n from the data 2 Estimate the gradient at θt by :
3 Set θt+1 = θt + ǫt∇ˆ
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
1 Take a subsample St of size n from the data 2 Estimate the gradient at θt by :
3 Set θt+1 = θt + ǫt∇ˆ
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
t=1 ǫt = ∞ and ∞ t=1 ǫ2 t < ∞, then θt will converge to a
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
t=1 ǫt = ∞ and ∞ t=1 ǫ2 t < ∞, then θt will converge to a
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
t=1 ǫt = ∞ and ∞ t=1 ǫ2 t < ∞, then θt will converge to a
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
1 Set θ∗ = θt + σ2 2 ∇f (θt) + σηt , where ηt ∼ N(0, I)
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
1 Set θ∗ = θt + σ2 2 ∇f (θt) + σηt , where ηt ∼ N(0, I) 2 Accept and set θt+1 = θ∗ with probability
π(θt)q(θ∗|θt)
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
1 Set θ∗ = θt + σ2 2 ∇f (θt) + σηt , where ηt ∼ N(0, I) 2 Accept and set θt+1 = θ∗ with probability
π(θt)q(θ∗|θt)
3 If rejected, set θt+1 = θt
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
MALA
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
1 Take a subsample St of size n from the data
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
1 Take a subsample St of size n from the data 2 Estimate the gradient at θt by
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
1 Take a subsample St of size n from the data 2 Estimate the gradient at θt by
3 Set θt+1 = θt + ǫt 2 ∇ˆ
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
1 Take a subsample St of size n from the data 2 Estimate the gradient at θt by
3 Set θt+1 = θt + ǫt 2 ∇ˆ
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
SGLD
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
SGLD-CV
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
SGLD-CV
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
SGLD-CV
Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
1 Take a subsample St of size n from the data
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
1 Take a subsample St of size n from the data 2 Estimate the gradient at θt by ∇˜
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
1 Take a subsample St of size n from the data 2 Estimate the gradient at θt by ∇˜
3 Set θt+1 = θt + ǫt 2 ∇˜
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Comparison
MALA SGLD (n=10) SGLD−CV 2.5 5.0 7.5 25 50 75 100
Passes through data Kernel Stein Discrepency
MALA SGLD (n=100) SGLD (n=50) SGLD (n=10) SGLD−CV
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Comparison
MALA −2 −1 1 2 −2 −1 1 2 SGLD −2 −1 1 2 SGLD−CV −2 −1 1 2
2N(µ1, σ) + 1 2N(µ2, σ)
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
The Covertype Dataset
Elevation (m) Aspect (degrees azimuth) Slope (degrees) Horizontal distance to nearest surface water (m) Vertical distance to nearest surface water (m) Horizontal distance to nearest roadway (m) Hillshade 9am (0-255) Hillshade Noon (0-255) Hillshade 3pm (0-255) Horizontal distance to wildfire ignition points (m) Wilderness area designation x4 (binary) Soil type x40 (binary) Class (1-7)
1: Spruce/Fir 2: Lodgepole Pine 3: Ponderosa Pine 4: Willow/ Cottonwood 5: Aspen 6: Douglas Fir 7: Krummholz Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
The Covertype Dataset
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
The Covertype Dataset
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
The Covertype Dataset
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
The Covertype Dataset
0.52 0.53 0.54 0.55 1 2 3
Passes through training data Log loss of test set
SGLD SGLD−CV
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
It is more practical to use numerical differentiation for this (e.g. sgmcmc for R)
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Gareth O. Roberts and Richard L. Tweedie. Exponential Convergence of Langevin Distributions and Their Discrete Approximations. https://www.jstor.org/stable/3318418 R´ emi Bardenet, Arnaud Doucet, and Chris Holmes. On Markov chain Monte Carlo methods for tall data. http://jmlr.org/papers/v18/15-205.html Max Welling and Yee W. Teh. Bayesian Learning via Stochastic Gradient Langevin Dynamics. https://www.ics.uci.edu/~welling/publications/papers/ stoclangevin_v6.pdf Jack Baker, Paul Fearnhead, Emily B. Fox, and Christopher Nemeth. Control Variates for Stochastic Gradient MCMC. https://arxiv.org/abs/1706.05439
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation
Connie TrojanSupervisor: Srshti Putcha Approximate Posterior Sampling via Stochastic Optimisation