markov models and hidden markov models
play

Markov Models and Hidden Markov Models Robert Platt Northeastern - PowerPoint PPT Presentation

Markov Models and Hidden Markov Models Robert Platt Northeastern University Some images and slides are used from: 1. CS188 UC Berkeley 2. RN, AIMA Markov Models We have already seen that an MDP provides a useful framework for modeling


  1. Markov Models and Hidden Markov Models Robert Platt Northeastern University Some images and slides are used from: 1. CS188 UC Berkeley 2. RN, AIMA

  2. Markov Models We have already seen that an MDP provides a useful framework for modeling stochastic control problems. Markov Models: model any kind of temporally dynamic system.

  3. Probability again: Independence Two random variables, x and y , are independent when: The outcomes of two different coin flips are usually independent of each other Image: Berkeley CS188 course notes (downloaded Summer 2015)

  4. Probability again: Independence If: Then: Why?

  5. Are T and W independent? T W P hot sun 0.4 hot rain 0.1 cold sun 0.2 cold rain 0.3 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  6. Are T and W independent? T P hot 0.5 cold 0.5 T W P hot sun 0.4 hot rain 0.1 cold sun 0.2 cold rain 0.3 W P sun 0.6 rain 0.4 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  7. Are T and W independent? T P hot 0.5 cold 0.5 T W P T W P hot sun 0.4 hot sun 0.3 hot rain 0.1 hot rain 0.2 cold sun 0.2 cold sun 0.3 cold rain 0.3 cold rain 0.2 W P sun 0.6 rain 0.4 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  8. Conditional independence Independence: Conditional independence: Equivalent statements of conditional independence: Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  9. Conditional independence: example cavity toothache catch P(toothache, catch | cavity) = P(toothache | cavity) = P(catch | cavity) or... P(toothache | cavity) = P(toothache | cavity, catch) P(catch | cavity) = P(catch | cavity, toothache)

  10. Conditional independence: example  What about this domain:  T raffjc  Umbrella  Raining Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  11. Conditional independence: example  What about this domain:  Fire  Smoke  Alarm Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  12. Markov Processes State at time=1 State at time=2 transitions

  13. Markov Processes State at time=1 State at time=2 transitions Since this is a Markov process, we assume transitions are Markov: Process model: Markov assumption:

  14. Markov Processes How do we calculate:

  15. Markov Processes How do we calculate:

  16. Markov Processes How do we calculate:

  17. Markov Processes How do we calculate: Can we simplify this expression?

  18. Markov Processes How do we calculate:

  19. Markov Processes How do we calculate:

  20. Markov Processes How do we calculate: In general:

  21. Markov Processes How do we calculate: Process model In general:

  22. Markov Processes: example  States: X = {rain, sun}  Initial distribution: 1.0 sun  Process model: P(X t | X t-1 ): T wo new ways of representing the same CPT X t-1 X t P(X t |X t-1 ) sun sun 0.9 0.9 sun sun 0.3 0.1 sun rain 0.1 0.9 0.3 rain sun rain rain rain sun 0.3 0.7 0.7 rain rain 0.7 0.1 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  23. Simulating dynamics forward Joint distribution: But, suppose we want to predict the state at time T, given a prior distribution at time 1? ...

  24. Markov Processes: example 0.3 0.9  Initial distribution: 1.0 sun rain sun 0.7 0.1  What is the probability distribution after one step? Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  25. Simulating dynamics forward  From initial observation of sun P( X 1 ) P( X 2 ) P( X 3 ) P( X 4 ) P( X ∞ )  From initial observation of rain P( X 1 ) P( X 2 ) P( X 3 ) P( X 4 ) P( X ∞ )  From yet another initial distribution P(X 1 ): … P( X 1 ) P( X ∞ ) Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  26. Simulating dynamics forward  From initial observation of sun P( X 1 ) P( X 2 ) P( X 3 ) P( X 4 ) P( X ∞ )  From initial observation of rain This is called the stationary distribution P( X 1 ) P( X 2 ) P( X 3 ) P( X 4 ) P( X ∞ )  From yet another initial distribution P(X 1 ): … P( X 1 ) P( X ∞ ) Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  27. Hidden Markov Models (HMMs) Hidden Markov Models: markov models applied to estimation problems – speech to text – tracking in computer vision – robot localization

  28. Hidden Markov Models (HMMs) Called an “emission” State, , is assumed to be unobserved However, you get to make one observation, , on each timestep.

  29. Hidden Markov Models (HMMs) Sensor Markov Assumption: the current observation depends only on current state:

  30. HMM example Rain t-1 Rain t Rain t+1 Umbrella Umbrella Umbrella R t R t+1 P(R t+1 |R t ) t-1 t t+1 +r +r 0.7 +r -r 0.3 -r +r 0.3 -r -r 0.7  An HMM is defjned by: R t U t P(U t |R t )  Initial distribution: +r +u 0.9  T ransitions: +r -u 0.1  Emissions: -r +u 0.2 -r -u 0.8 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  31. Real world HMM applications  Speech recognition HMMs:  Observations are acoustic signals (continuous valued)  States are specifjc positions in specifjc words (so, tens of thousands)  Machine translation HMMs:  Observations are words (tens of thousands)  States are translation options  Robot tracking:  Observations are range readings (continuous)  States are positions on a map (continuous) Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  32. HMM Filtering  Filtering, or monitoring, is the task of tracking the distribution B t (X) = P t (X t | e 1 , …, e t ) (the belief state) over time  We start with B 1 (X) in an initial setting, usually uniform  As time passes, or we get observations, we update B(X)  The Kalman fjlter was invented in the 60’s and fjrst implemented as a method of trajectory estimation for the Apollo program Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  33. HMM Filtering Given a prior distribution, , and a series of observations, , calculate the posterior distribution: Two steps: Process update Observation update

  34. HMM Filtering Given a prior distribution, , and a series of observations, , calculate the posterior distribution: Two steps: Process update Observation update

  35. HMM Filtering Given a prior distribution, , and a series of observations, , calculate the posterior distribution: “Beliefs” Two steps: Process update Observation update

  36. Process update This is just forward simulation of the Markov Model

  37. Process update: example  As time passes, uncertainty “accumulates” (T ransition model: ghosts usually go clockwise) T = 1 T = 2 T = 5 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  38. Observation update Where is a normalization factor

  39. Observation update  As we get observations, beliefs get reweighted, uncertainty “decreases” Before observation After observation Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  40. Robot localization example Prob 0 1 Observation model: can read in which directions there is a wall, never more than 1 mistake Process model: may not execute action with small prob. Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  41. Robot localization example Prob 0 1 Lighter grey: was possible to get the reading, but less likely b/c required 1 mistake Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  42. Robot localization example Prob 0 1 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  43. Robot localization example Prob 0 1 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  44. Robot localization example Prob 0 1 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  45. Robot localization example Prob 0 1 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  46. Weather HMM example R t R t+1 P(R t+1 |R t ) +r +r 0.7 +r -r 0.3 -r +r 0.3 B(+r) = 0.5 -r -r 0.7 B(-r) = 0.5 Rain 0 Rain 1 Rain 2 R t U t P(U t |R t ) +r +u 0.9 +r -u 0.1 -r +u 0.2 Umbrella 1 Umbrella 2 -r -u 0.8 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  47. Weather HMM example R t R t+1 P(R t+1 |R t ) +r +r 0.7 B’(+r) = 0.5 B’(-r) = 0.5 +r -r 0.3 -r +r 0.3 -r -r 0.7 B(+r) = 0.5 B(-r) = 0.5 Rain 0 Rain 1 Rain 2 R t U t P(U t |R t ) +r +u 0.9 +r -u 0.1 -r +u 0.2 Umbrella 1 Umbrella 2 -r -u 0.8 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

  48. Weather HMM example R t R t+1 P(R t+1 |R t ) +r +r 0.7 B’(+r) = 0.5 B’(-r) = 0.5 +r -r 0.3 -r +r 0.3 -r -r 0.7 B(+r) = 0.818 B(+r) = 0.5 B(-r) = 0.182 B(-r) = 0.5 Rain 0 Rain 1 Rain 2 R t U t P(U t |R t ) +r +u 0.9 +r -u 0.1 -r +u 0.2 Umbrella 1 Umbrella 2 -r -u 0.8 Slide: Berkeley CS188 course notes (downloaded Summer 2015)

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend