COMS 4721: Machine Learning for Data Science Lecture 22, 4/18/2017
- Prof. John Paisley
COMS 4721: Machine Learning for Data Science Lecture 22, 4/18/2017 - - PowerPoint PPT Presentation
COMS 4721: Machine Learning for Data Science Lecture 22, 4/18/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University M ARKOV MODELS s 1 s 2 s 3 s 4 The sequence ( s 1 , s 2 , s 3 , . . . )
◮ For the Markov chain, we gave an example
◮ A continuous hidden Markov model might
◮ For example, each si can be modified by
◮ But s1:T is still a discrete Markov chain.
A12 A23 A31 A21 A32 A13 A11 A22 A33 k = 1 k = 2 k = 3
k = 1 k = 2 k = 3 0.5 1 0.5 1
◮ The state could be a data point xi (Markov Chain classifier) ◮ The state could be an object (object ranking) ◮ The state could be the destination of a link (internet search engines)
◮ Sequences of discrete data may come from a few discrete distributions. ◮ Sequences of continuous data may come from a few distributions.
◮ Discrete-state Markov models: The states live in a discrete space. ◮ Continuous-state Markov models: The states live in a continuous space.
◮ st ∈ Rp is a continuous-state latent (unobserved) Markov process ◮ xt ∈ Rd is a continuous-valued observation ◮ The process noise ǫt ∼ N(0, Q) ◮ The measurement noise εt ∼ N(0, V)
◮ Tracking moving objects ◮ Automatic control systems ◮ Economics and finance (e.g., stock modeling) ◮ etc.
1 2(∆t)2
1 2(∆t)2
◮ π is the initial state distribution ◮ A is the transition matrix among the discrete set of states ◮ B contains the state-dependent distributions on discrete-valued data
◮ #1: Kalman filtering problem. We’ll focus on this one today. ◮ #2: Kalman smoothing problem. Requires extra step (not discussed).
N(Dst,V)
N(Dst,V)
N(xt|Dst,V)
◮ We therefore only need to define a Gaussian prior on the first state to
t , Σx t )
t, Σs t)
◮ In the prior distribution notice that we add Q to the covariance,
◮ In the posterior p(st|x1, . . . , xt), xt “pulls” the distribution away.
sn−1 sn sn+1 xn−1 xn xn+1 s1 s2 x1 x2
◮ st ∼ Discrete(π) ◮ xt|st ∼ N(µst, Σst)
sn−1 sn sn+1 xn−1 xn xn+1 s1 s2 x1 x2
◮ st|st−1 ∼ Discrete(Ast−1) ◮ xt|st ∼ N(µst, Σst)
sn−1 sn sn+1 xn−1 xn xn+1 s1 s2 x1 x2
◮ st ∼ N(0, Q) ◮ xt|st ∼ N(Dst, V)
sn−1 sn sn+1 xn−1 xn xn+1 s1 s2 x1 x2
◮ st|st−1 ∼ N(Cst−1, Q) ◮ xt|st ∼ N(Dst, V)