CS354 Nathan Sprague October 13, 2020 Probabilistic State - - PowerPoint PPT Presentation

cs354
SMART_READER_LITE
LIVE PREVIEW

CS354 Nathan Sprague October 13, 2020 Probabilistic State - - PowerPoint PPT Presentation

CS354 Nathan Sprague October 13, 2020 Probabilistic State Representations: Continuous Probabilistic Robotics. Thrun, Burgard, Fox, 2005 Probability Density Functions Represent probability distributions over random variables: Properties: f (


slide-1
SLIDE 1

CS354

Nathan Sprague October 13, 2020

slide-2
SLIDE 2

Probabilistic State Representations: Continuous

Probabilistic Robotics. Thrun, Burgard, Fox, 2005

slide-3
SLIDE 3

Probability Density Functions

Represent probability distributions over random variables: Properties:

f (x) ≥ 0 ∞

−∞

f (x)dx = 1

Interpretation:

P(a ≤ x ≤ b) = b

a

f (x)dx

slide-4
SLIDE 4

Expectation, Variance

Expectation (continuous) (also referred to as the ”mean” or ”first moment”) µ = E[x] =

  • xf (x)dx

Expectation (discrete) E[X] =

n

  • 1

P(xi)xi Variance (also referred to as the ”second moment”) σ2 = E[(x − E[x])2]

slide-5
SLIDE 5

Quiz 1

What is the expectation of this pdf?

slide-6
SLIDE 6

Quiz 2

E[X] =

n

  • 1

P(xi)xi σ2 = E[(x − E[x])2]

Imagine we are rolling a four-sided die. The following probability distribution describes the probability for each number that we could roll: P(X = 1) = .7 P(X = 2) = .1 P(X = 3) = .1 P(X = 4) = .1 What is the expected value of this distribution? What is the variance?

slide-7
SLIDE 7

Sample Mean and Variance

Expectation and variance are properties of distributions. We can also calculate the sample mean and the sample variance for a given data set: {x1, x2, ..., xn}. Sample mean m = 1 n

n

  • i=1

xi Sample variance s2 = 1 n

n

  • i=1

(xi − m)2

slide-8
SLIDE 8

Normal Distribution

f (x, µ, σ) = 1 σ √ 2π e− (x−µ)2

2σ2

(Normal because of the central limit theorem.) All distributions

slide-9
SLIDE 9

Vector-Valued State

We’ll need to generalize all of this to the case where the state

  • f the system can’t be represented as a single number.

Use a vector x to represent the state.

slide-10
SLIDE 10

Covariance

cov(x, y) = E[(x − µx)(y − µy)] Properties:

cov(x, y) = cov(y, x) If x and y are independent, cov(x, y) = 0 If cov(x, y) > 0, y tends to increase when x increases. If cov(x, y) < 0, y tends to decrease when x increases.

slide-11
SLIDE 11

Covariance Matrix

Covariance matrix: cov(x) = Σx = E[(x − ˆ x)(x − ˆ x)T] Where x is a random vector and ˆ x is the vector mean. The entry at row i, column j in the matrix is cov(xi,xj) Multivariate normal distribution is parameterized by the mean vector and covariance matrix.

slide-12
SLIDE 12

Multivariate PDF Example

x = 3 3

  • , Σ =

1 1

slide-13
SLIDE 13

Multivariate PDF Example

x = 3 3

  • , Σ =

1 1

slide-14
SLIDE 14

Multivariate PDF Example

x = 3 3

  • , Σ =

.2 1

slide-15
SLIDE 15

Multivariate PDF Example

x = 3 3

  • , Σ =

.2 1

slide-16
SLIDE 16

Multivariate PDF Example

x = 3 3

  • , Σ =

1 .9 .9 1

slide-17
SLIDE 17

Multivariate PDF Example

x = 3 3

  • , Σ =

1 .9 .9 1

slide-18
SLIDE 18

Multivariate PDF Example

x = 3 3

  • , Σ =

.5 −.3 −.3 2

slide-19
SLIDE 19

Multivariate PDF Example

x = 3 3

  • , Σ =

.5 −.3 −.3 2

slide-20
SLIDE 20

Why is this Useful For Localization?

Memory and computation requirements grow exponentially for grid-based disributions. E.g. if we want 100 cells per dimension, we need 100d cells. To approximate with a normal distribution we need d2 + d to store.

slide-21
SLIDE 21

Can We Do Recursive State Estimation?

Two Steps:

Prediction based on system dynamics: Bel−(xt) =

  • p(xt | xt−1)Bel(xt−1)dxt−1

Correction based on sensor reading: Bel(xt) = ηp(zt | xt)Bel−(xt)

  • YES. The Kalman filter. Next time.