15-780: Grad AI Lecture 16: Probability Geoff Gordon (this lecture) - - PowerPoint PPT Presentation

15 780 grad ai lecture 16 probability
SMART_READER_LITE
LIVE PREVIEW

15-780: Grad AI Lecture 16: Probability Geoff Gordon (this lecture) - - PowerPoint PPT Presentation

15-780: Grad AI Lecture 16: Probability Geoff Gordon (this lecture) Tuomas Sandholm TAs Erik Zawadzki, Abe Othman Randomness in search Rapidly-exploring Random Trees Break up C-space into Voronoi regions around random landmarks Invariant:


slide-1
SLIDE 1

15-780: Grad AI Lecture 16: Probability

Geoff Gordon (this lecture) Tuomas Sandholm TAs Erik Zawadzki, Abe Othman

slide-2
SLIDE 2

Randomness in search

slide-3
SLIDE 3

Rapidly-exploring Random Trees

Break up C-space into Voronoi regions around random landmarks Invariant: landmarks always form a tree

  • known path to root

Subject to this requirement, placed in a way that tends to split large Voronoi regions

  • coarse-to-fine search

Goal: feasibility not optimality (*)

slide-4
SLIDE 4

RRT: required subroutines

RANDOM_CONFIG

  • samples from C-space

EXTEND(q, q’)

  • local controller, heads toward q’ from q
  • stops before hitting obstacle (and perhaps also

after bound on time or distance) FIND_NEAREST(q, Q)

  • searches current tree Q for point near q
slide-5
SLIDE 5

Path Planning with RRTs

[ Kuffner & LaValle , ICRA’00]

RRT = Rapidly-Exploring Random Tree

BUILT_RRT(qinit) { T = qinit for k = 1 to K { qrand = RANDOM_CONFIG() EXTEND(T, qrand); } } EXTEND(T, q) { qnear = FIND_NEAREST(q, T) qnew = EXTEND(qnear, q) T = T + (qnear, qnew) }

slide-6
SLIDE 6

Path Planning with RRTs

qinit

[ Kuffner & LaValle , ICRA’00]

RRT = Rapidly-Exploring Random Tree

BUILT_RRT(qinit) { T = qinit for k = 1 to K { qrand = RANDOM_CONFIG() EXTEND(T, qrand); } } EXTEND(T, q) { qnear = FIND_NEAREST(q, T) qnew = EXTEND(qnear, q) T = T + (qnear, qnew) }

slide-7
SLIDE 7

Path Planning with RRTs

qinit qrand

[ Kuffner & LaValle , ICRA’00]

RRT = Rapidly-Exploring Random Tree

BUILT_RRT(qinit) { T = qinit for k = 1 to K { qrand = RANDOM_CONFIG() EXTEND(T, qrand); } } EXTEND(T, q) { qnear = FIND_NEAREST(q, T) qnew = EXTEND(qnear, q) T = T + (qnear, qnew) }

slide-8
SLIDE 8

Path Planning with RRTs

qnear qinit qrand

[ Kuffner & LaValle , ICRA’00]

RRT = Rapidly-Exploring Random Tree

BUILT_RRT(qinit) { T = qinit for k = 1 to K { qrand = RANDOM_CONFIG() EXTEND(T, qrand); } } EXTEND(T, q) { qnear = FIND_NEAREST(q, T) qnew = EXTEND(qnear, q) T = T + (qnear, qnew) }

slide-9
SLIDE 9

Path Planning with RRTs

qnear

qnew

qinit qrand

[ Kuffner & LaValle , ICRA’00]

RRT = Rapidly-Exploring Random Tree

BUILT_RRT(qinit) { T = qinit for k = 1 to K { qrand = RANDOM_CONFIG() EXTEND(T, qrand); } } EXTEND(T, q) { qnear = FIND_NEAREST(q, T) qnew = EXTEND(qnear, q) T = T + (qnear, qnew) }

slide-10
SLIDE 10

RRT example

Planar holonomic robot

slide-11
SLIDE 11

RRTs explore coarse to fine

Tend to break up large Voronoi regions

  • higher probability of qrand being in them

Limiting distribution of vertices given by RANDOM_CONFIG

  • as RRT grows, probability that qrand is

reachable with local controller (and so immediately becomes a new vertex) approaches 1

slide-12
SLIDE 12
slide-13
SLIDE 13
slide-14
SLIDE 14
slide-15
SLIDE 15
slide-16
SLIDE 16
slide-17
SLIDE 17

RRT example

slide-18
SLIDE 18

RRT for a car (3 dof)

slide-19
SLIDE 19

Planning with RRTs

Build RRT from start until we add a node that can reach goal using local controller (Unique) path: root → last node → goal Optional: “rewire” tree during growth by testing connectivity to more than just closest node Optional: grow forward and backward

slide-20
SLIDE 20
slide-21
SLIDE 21
slide-22
SLIDE 22
slide-23
SLIDE 23
slide-24
SLIDE 24

Probability

slide-25
SLIDE 25

Probability

Random variables Atomic events Sample space

slide-26
SLIDE 26

Probability

Events Combining events

slide-27
SLIDE 27

Probability

Measure:

  • disjoint union:
  • e.g.:
  • interpretation:

Distribution:

  • interpretation:
  • e.g.:
slide-28
SLIDE 28

Example

Weather AAPL price

up same down sun rain

0.09 0.15 0.06 0.21 0.35 0.14

slide-29
SLIDE 29

Bigger example

Weather AAPL price

up same dow n sun rain

0.03 0.05 0.02 0.07 0.12 0.05

Weather

up same dow n sun rain

0.14 0.23 0.09 0.06 0.10 0.04

LAX PIT

slide-30
SLIDE 30

Notation

X=x: event that r.v. X is realized as value x P(X=x) means probability of event X=x

  • if clear from context, may omit “X=”
  • instead of P(Weather=rain), just P(rain)
  • complex events too: e.g., P(X=x,

Y≠y) P(X) means a function: x → P(X=x)

slide-31
SLIDE 31

Functions of RVs

Extend definition: any deterministic function of RVs is also an RV E.g., “profit”: Weather AAPL price

up same down sun rain

–11 11 –11 11

slide-32
SLIDE 32

Sample v. population

Suppose we watch for 100 days and count up our

  • bservations

Weather AAPL price

up same dow n sun rain

0.09 0.15 0.06 0.21 0.35 0.14

Weather AAPL price

up same dow n sun rain

7 12 3 22 41 15

slide-33
SLIDE 33

Law of large numbers

If we take a sample of size N from distribution P , count up frequencies of atomic events, and normalize (divide by N) to get a distribution P Then P → P as N → ∞ ~ ~ (simple version)

slide-34
SLIDE 34

Working w/ distributions

Marginals (eliminate an irrelevant RV) Conditionals (incorporate an observation) Joint (before marginalizing or conditioning)

slide-35
SLIDE 35

Marginals

Weather AAPL price

up same down sun rain

0.09 0.15 0.06 0.21 0.35 0.14

slide-36
SLIDE 36

Law of total probability

Two RVs, X and Y Y has values y1, y2, …, yk P(X) = P(X, Y=y1) + P(X, Y=y2) + … also called “sum rule”

slide-37
SLIDE 37

Conditioning on an observation

Two steps:

  • enforce consistency
  • renormalize

Notation: Weather Coin

H T sun rain

0.15 0.15 0.35 0.35

slide-38
SLIDE 38

Conditionals

Weather AAPL price

up same down sun rain

0.03 0.05 0.02 0.07 0.12 0.05

Weather

up same down sun rain

0.14 0.23 0.09 0.06 0.10 0.04

LAX PIT

slide-39
SLIDE 39

Conditionals

Thought experiment: what happens if we condition on an event of zero probability?

slide-40
SLIDE 40

P(X | Y) is a function: x, y → P(X=x | Y=y) So:

  • P(X |

Y) P(Y) means the function x, y →

Notation

slide-41
SLIDE 41

Conditionals in literature

When you have eliminated the impossible, whatever remains, however improbable, must be the truth.

—Sir Arthur Conan Doyle, as Sherlock Holmes

slide-42
SLIDE 42

$$$

Exercise

slide-43
SLIDE 43

$$$

Exercise