probability recap cs 4100 artificial intelligence
play

Probability Recap CS 4100: Artificial Intelligence Hidden Markov - PDF document

Probability Recap CS 4100: Artificial Intelligence Hidden Markov Models Co Conditional probability Pr Product rule Ch Chain rule X, , Y in independent if if and only ly if if: X an and Y ar are co e conditional ally


  1. Probability Recap CS 4100: Artificial Intelligence Hidden Markov Models • Co Conditional probability • Pr Product rule • Ch Chain rule • X, , Y in independent if if and only ly if if: • X an and Y ar are co e conditional ally i indep epen enden ent g given en Z Z i if an and o only i if: Jan-Willem van de Meent, Northeastern University [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.] Reasoning over Time or Space Markov Models • Va Value of X at at a a given en time e is cal called ed the e st state • Oft Often, we we wa want to re reason on abou bout a sequ quence of of obs observ rvation ons • Speech recognition X 1 X 2 X 3 X 4 • Robot localization • User attention • Medical monitoring • Tw Two o Distr tributi tion ons (CPTs Ts): • In Initial s state p probabilities , specify probabilities for first state X 1 • Ne Need to introduce time me (or sp space) into our mo models • Tra ransiti tion pr proba babi bilities or dynamics, specify how the state evolves • St Stationarity as assumption: transition probabilities the same at all times • Same as MDP transition model, but no no cho hoice of action Example Markov Chain: Weather Conditional Independence • St States: X X = {ra {rain, sun} • In Initial di distri ribu bution on: 1. 1.0 0 sun • Ba Basic ic condit itio ional l in independence: • CP CPT: P( P(X t | | X t-1 ) • Past and future independent given the present Two ways of representing the same CPT X t-1 X t P(X t |X t-1 ) • Each time step only depends on the previous sun sun 0.9 0.9 • This is called the (first order) Markov property 0.3 0.9 sun rain 0.1 sun sun rain sun 0.3 rain sun 0.1 • No Note that the chain is s just st a (grow growabl ble) ) BN 0.3 rain rain 0.7 rain rain 0.7 • We can always use generic BN reasoning on it if we 0.7 0.1 truncate the chain at a fixed length Example Markov Chain: Weather Mini-Forward Algorithm • In Initi tial al di distr tribu buti tion: 1. 1.0 0 sun 0.9 0.3 • Qu Question: What’s s P( P(X) on on some ome da day t? rain sun X 1 X 2 X 3 X 4 0.7 0.1 • Wh What i is t s the p probability d dist stribution a after o one st step? P ( x t ) = X Forward simulation P ( x t − 1 , x t ) x t − 1 X = P ( x t | x t − 1 ) P ( x t − 1 ) x t − 1 • Sp Special case of f va variable elimination wi with ordering: X 1 , X , X 2 , ..., X , ..., X t-1

  2. Hidden Markov Models Pacman – Sonar (P4) Hidden Markov Models Example: Weather HMM P ( X t | X t − 1 ) • Marko kov chains not so useful for most agents • Need observations to update your beliefs Rain t-1 Rain t Rain t+1 • Hidden Marko kov models (HMMs) P ( E t | X t ) • Underlying Markov chain over states X Umbrella t-1 • You observe outputs (effects) at each time step Umbrella t Umbrella t+1 X 1 X 2 X 3 X 4 X 5 • An An HMM is defined by: y: R t-1 R t P(R t |R t-1 ) R t U t P(U t |R t ) • In Initial al distribution: +r +r 0.7 +r +u 0.9 • Tr Transiti tion ons: P ( X t | X t − 1 ) +r -r 0.3 +r -u 0.1 E 1 E 2 E 3 E 4 E 5 • Emi Emissions: P ( E t | X t ) -r +r 0.3 -r +u 0.2 -r -r 0.7 -r -u 0.8 Example: Ghostbusters HMM Ghostbusters – Circular Dynamics -- HMM P( P(X 1 ) : uniform • 1/9 1/9 1/9 P( P(X X | X ’ ) : usually move clockwise, • 1/9 1/9 1/9 but sometimes move in a random direction or stay in place 1/9 1/9 1/9 P(X 1 ) P(R ij P( ij | X | X) : = same sensor model as before; • red means close, green means far away. 1/6 1/6 1/2 X 1 X 2 X 3 X 4 0 1/6 0 0 0 0 R i,j R i,j R i,j R i,j X 5 P(X|X ’ =<1,2>) [Demo: Ghostbusters – Circular Dynamics – HMM (L14D2)] Conditional Independence Real HMM Examples • HM HMMs ha have two wo impo portant nt inde ndepe pende ndenc nce pr prope perties: • Sp Speech recognition HMMs: • Marko kov hidden process: future depends on past via the present • Observations are acoustic signals (continuous valued) • States are specific positions in specific words (so, tens of thousands) • Current observation in independent of all else given current state • Mach Machine e tran anslat ation HMMs MMs: X 1 X 2 X 3 X 4 X 5 • Observations are words (tens of thousands) • States are translation options E 1 E 2 E 3 E 4 E 5 • Robot tracki king: • Observations are range readings (continuous) • States are positions on a map (continuous) • Qu Quiz: does this mean that evidence variables are guaranteed to be independent? • [No, they tend to correlated by the hidden state]

  3. Filtering / Monitoring Example: Robot Localization Example from • Fi Filtering ng , or monitoring, is the task of tracking the distribution Michael Pfeiffer B t (X) (X) = = P t (X t | | e 1 , …, …, e t ) (the belief state) over time • We start with B 1 (X (X) in an initial setting, usually uniform • Given B t (X) ) and new evidence e t+ t+1 evidence we compute B t+ 1 (X) X) t+1 • Ba Basic idea: Use interleaved jo join in , su sum , and no normalize that we have also seen in va variable elimi minati tion : • Jo Join P( P(X t+ t+1 , X , X t-1 1 | e | e 1 , …, …, e t ) Prob 0 1 • Su Sum P( P(X t+1 1 | e | e 1 , …, …, e t ) • Jo Join P(e (e t+ t+1 , X , X t+1 +1 | e | e 1 , …, …, e t , e t+ t+1 ) t=0 • No Normaliz lize P( P(X t+ 1 | e | e 1 , …, …, e t , e t+ t+1 ) • Se Sensor mo model: can read in which directions t+1 there is a wall, never more than 1 mistake • The Ka Kalman f filter r was invented in the 60’s and first implemented • Mo Motion model el: may not execute action with small prob. as a method of trajectory estimation for the Apollo program Example: Robot Localization Example: Robot Localization Prob 0 1 Prob 0 1 t=1 hter grey: was possible to get the reading, Li Light t=2 but less likely b/c required 1 mistake Example: Robot Localization Example: Robot Localization Prob 0 1 Prob 0 1 t=3 t=4 Inference: Base Cases Example: Robot Localization X 1 X 1 X 2 E 1 Prob 0 1 t=5

  4. Filtering (join + sum): Elapse Time Example: Passage of Time • As Assume me we have ve current belief P( P(X | evi vidence to date) • As time passes, uncertainty “ accumulates ” X 1 X 2 (Transition model: ghosts usually go clockwise) • Th Then, afte fter on one ti time ste tep passes: P ( X t +1 | e 1: t ) X = P ( X t +1 , x t | e 1: t ) • Or compactly: x t T = 1 T = 2 T = 5 X B 0 ( X t +1 ) = P ( X 0 | x t ) B ( x t ) X = P ( X t +1 | x t , e 1: t ) P ( x t | e 1: t ) x t x t X = P ( X t +1 | x t ) P ( x t | e 1: t ) x t Basic idea: beliefs get “pushed” through the transitions • Ba • With the “B” notation, we have to be careful about what time step t the belief is about, and what evidence it includes Filtering (join + normalize): Observation Example: Observation • As Assume me we have ve current belief P( P(X | previ vious evi vidence): X 1 • As we get observations, beliefs get reweighted, uncertainty “ decreases ” B 0 ( X t +1 ) = P ( X t +1 | e 1: t ) • Th Then, afte fter evidence com omes in: E 1 P ( X t +1 | e 1: t +1 ) = P ( X t +1 , e t +1 | e 1: t ) /P ( e t +1 | e 1: t ) ∝ X t +1 P ( X t +1 , e t +1 | e 1: t ) = P ( e t +1 | e 1: t , X t +1 ) P ( X t +1 | e 1: t ) Before observation After observation = P ( e t +1 | X t +1 ) P ( X t +1 | e 1: t ) B ( X t +1 ) ∝ X t +1 P ( e t +1 | X t +1 ) B 0 ( X t +1 ) Basic idea: beliefs “reweighted” Ba • • Or Or, compa pactly: by likelihood of evidence First compute pr produ duct (jo (join in) then no normalize • Unlike passage of time, we have B ( X t +1 ) ∝ X t +1 P ( e t +1 | X t +1 ) B 0 ( X t +1 ) to renormalize Pacman – Sonar (P4) Pacman – Sonar (with beliefs) [Demo: Pacman – Sonar – No Beliefs(L14D1)] Particle Filtering Particle Filtering • Fi Filtering ng: approximate solut ution 0.0 0.1 0.0 • So Some metime mes |X |X| is is too big ig to use exact in infere rence 0.0 0.0 0.2 • |X |X| may be too big to even store B( B(X) E.g. X is continuous • 0.0 0.2 0.5 • So Solution: approxima mate inference Track samples of X , not all values • • Assign weights w w to all samples (as in likelihood weighting) • Weighted samples are called pa particles In In me memo mory: list of particles, not states • • This is how robot localization works ks in practice

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend