Reasoning Under Uncertainty Over Time Alice Gao Lecture 16 - - PowerPoint PPT Presentation

reasoning under uncertainty over time
SMART_READER_LITE
LIVE PREVIEW

Reasoning Under Uncertainty Over Time Alice Gao Lecture 16 - - PowerPoint PPT Presentation

1/25 Reasoning Under Uncertainty Over Time Alice Gao Lecture 16 Readings: R & N 15.1 to 15.3 Based on work by K. Leyton-Brown, K. Larson, and P. van Beek 2/25 Outline Learning Goals Revisiting the Learning goals 3/25 Learning Goals


slide-1
SLIDE 1

1/25

Reasoning Under Uncertainty Over Time

Alice Gao

Lecture 16 Readings: R & N 15.1 to 15.3 Based on work by K. Leyton-Brown, K. Larson, and P. van Beek

slide-2
SLIDE 2

2/25

Outline

Learning Goals Revisiting the Learning goals

slide-3
SLIDE 3

3/25

Learning Goals

By the end of the lecture, you should be able to

▶ Construct a hidden Markov model given a real-world scenario. ▶ Perform fjltering, prediction, smoothing and derive the most

likely explanation given a hidden Markov model.

slide-4
SLIDE 4

4/25

Inference in a Changing World

So far, we can reason probabilistically in a static world. However, the world evolves over time. Applications:

▶ weather predictions ▶ stock market predictions ▶ patient monitoring ▶ robot localization ▶ speech and handwriting recognition

slide-5
SLIDE 5

5/25

The Umbrella World

You are the security guard stationed at a secret underground

  • installation. You want to know whether it’s raining today, but your
  • nly access to the outside world occurs each morning when you see

the director coming in with, or without, an umbrella.

slide-6
SLIDE 6

6/25

States and Observations

▶ The world contains a series of time slices. ▶ Each time slice contains a set of random variables,

some observable, some not. Xt the un-observable variables at time t Et the observable variables at time t What are the observable and unobservable random variables in the umbrella world?

slide-7
SLIDE 7

7/25

The transition model

How does the current state depend on the previous states? P(Xt|Xt−1 ∧ Xt−2 ∧ Xt−3 ∧ · · · ∧ X1) Problem: As t increases, the number of previous states is

  • unbounded. The conditional probability distribution can be

unboundedly large. Solutions: Make the Markov assumption — the current state depends on only a fjnite fjxed number of previous states.

slide-8
SLIDE 8

8/25

K-order Markov processes

First-order Markov process: Xt−2 Xt−1 Xt Xt+1 The transition model: P(Xt|Xt−1 ∧ Xt−2 ∧ Xt−3 ∧ · · · ∧ X1) = P(Xt|Xt−1) Second-order Markov process: Xt−2 Xt−1 Xt Xt+1 The transition model: P(Xt|Xt−1 ∧ Xt−2 ∧ Xt−3 ∧ · · · ∧ X1) = P(Xt|Xt−1 ∧ Xt−2)

slide-9
SLIDE 9

9/25

Transition model for the umbrella world

The future is independent of the past given the present. Rt−2 Rt−1 Rt Rt+1 The transition model: P(Rt|Rt−1 ∧ Rt−2 ∧ Rt−3 ∧ · · · ∧ R1) = P(Rt|Rt−1)

slide-10
SLIDE 10

10/25

Stationary Process

Is there a difgerent conditional probability distribution for each time step? Stationary process:

▶ The dynamics does not change over time. ▶ The conditional probability distribution for each time step

remains the same.

slide-11
SLIDE 11

11/25

Transition model for the umbrella world

P(Rt|Rt−1) = 0.7 P(Rt|¬Rt−1) = 0.3 Rt−2 Rt−1 Rt Rt+1

slide-12
SLIDE 12

12/25

Sensor model

How does the evidence variable Et for each time step t depend on the previous and current state variables? Sensor Markov assumption: Any state is suffjcient to generate the current sensor values. P(Et|Xt ∧ Xt−1 ∧ · · · ∧ X1 ∧ Et−1 ∧ Et−2 ∧ · · · ∧ E1) = P(Et|Xt)

slide-13
SLIDE 13

13/25

Complete model for the umbrella world

P(R1) = 0.5 P(Rt|Rt−1) = 0.7 P(Rt|¬Rt−1) = 0.3 P(Ut|Rt) = 0.9 P(Ut|¬Rt) = 0.2 Rt−2 Rt−1 Rt Rt+1 Ut−2 Ut−1 Ut Ut+1

slide-14
SLIDE 14

14/25

Hidden Markov Model

▶ A Markov process ▶ The state variables are unobservable. ▶ The evidence variables, which depend on the states,

are observable.

slide-15
SLIDE 15

15/25

Common Inference Tasks

▶ Filtering: the posterior distribution over the most recent state

given all evidence to date.

▶ Prediction: the posterior distribution over the future state

given all evidence to date.

▶ Smoothing: the posterior distribution over a past state, given

all evidence to date.

▶ Most likely explanation: fjnd the sequence of states that is

most likely to have generated all the evidence to date.

slide-16
SLIDE 16

16/25

Filtering

Given xt−1 = P(Rt−1|U1 ∧ · · · ∧ Ut−1), how do we compute xt = P(Rt|U1 ∧ · · · ∧ Ut)? Examples: P(R1 = r1|U1) and P(R2 = r2|U1 ∧ U2)

slide-17
SLIDE 17

17/25

CQ: Filtered Estimate

CQ: What is P(R1 = t|U1 = t)? (A) 0.518 (B) 0.618 (C) 0.718 (D) 0.818 (E) 0.918

slide-18
SLIDE 18

18/25

Filtered Estimate P(R2 = r2|U1 = u1 ∧ U2 = u2)

P(R2 = r2|U1 = u1 ∧ U2 = u2) =αP(U2 = u2|R2 = r2) ∑

r1

P(R2 = r2|R1 = r1)P(R1 = r1|U1 = u1) Forward recursion:

▶ From P(R1 = r1|U1 = u1) to P(R2 = r2|U1 = u1 ∧ U2 = u2) ▶ From P(R2 = r2|U1 = u1 ∧ U2 = u2) to

P(R3 = r3|U1 = u1 ∧ U2 = u2 ∧ U3 = u3)

▶ ...

slide-19
SLIDE 19

19/25

CQ: Filtering

CQ: Consider P(U2|R2 ∧ U1). Which one of the following simplifjcations is valid? (A) P(U2|R2 ∧ U1) = P(U2|R2) (B) P(U2|R2 ∧ U1) = P(U2|U1) (C) P(U2|R2 ∧ U1) = P(U2) (D) None of (A), (B), and (C) is a valid simplifjcation.

slide-20
SLIDE 20

20/25

CQ: Filtering

CQ: Consider P(R2|R1 ∧ U1). Which one of the following simplifjcations is valid? (A) P(R2|R1 ∧ U1) = P(R2|R1) (B) P(R2|R1 ∧ U1) = P(R2|U1) (C) P(R2|R1 ∧ U1) = P(R2) (D) None of (A), (B), and (C) is a valid simplifjcation.

slide-21
SLIDE 21

21/25

Forward Recursion for Filtering

P(Rt|U1 ∧ · · · ∧ Ut) =αP(Ut|Rt) ∑

Rt−1

P(Rt|Rt−1)P(Rt−1|U1 ∧ · · · ∧ Ut−1)

slide-22
SLIDE 22

22/25

Prediction

Given P(Rt+k|U1 ∧ · · · ∧ Ut−1), how do we compute P(Rt+k+1|U1 ∧ · · · ∧ Ut−1)? Forward Recursion P(Rt+k+1|U1 ∧ · · · ∧ Ut−1) = ∑

Rt+k

P(Rt+k+1|Rt+k)P(Rt+k|U1 ∧ · · · ∧ Ut−1)

slide-23
SLIDE 23

23/25

Smoothing

For 1 ≤ k < t, P(Rk|U1 ∧ . . . Ut) =α P(Rk|U1 ∧ · · · ∧ Uk) P(Uk+1 ∧ · · · ∧ Ut|Rk) Forward Recursion P(Rt|U1 ∧ · · · ∧ Ut) =αP(Ut|Rt) ∑

Rt−1

P(Rt|Rt−1)P(Rt−1|U1 ∧ · · · ∧ Ut−1) Backward Recursion P(Uk+1 ∧ · · · ∧ Ut|Rk) = ∑

Rk+1

P(Uk+1|Rk+1) P(Uk+2 ∧ · · · ∧ Ut|Rk+1) P(Rk+1|Rk)

slide-24
SLIDE 24

24/25

Most likely explanation

slide-25
SLIDE 25

25/25

Revisiting the Learning Goals

By the end of the lecture, you should be able to

▶ Construct a hidden Markov model given a real-world scenario. ▶ Perform fjltering, prediction, smoothing and derive the most

likely explanation given a hidden Markov model.